Search (138 results, page 1 of 7)

  • × theme_ss:"Semantic Web"
  1. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.26
    0.26491702 = product of:
      0.46360478 = sum of:
        0.118422605 = product of:
          0.197371 = sum of:
            0.115060724 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.115060724 = score(doc=701,freq=2.0), product of:
                0.3070917 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03622214 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
            0.054806177 = weight(_text_:retrieval in 701) [ClassicSimilarity], result of:
              0.054806177 = score(doc=701,freq=28.0), product of:
                0.109568894 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03622214 = queryNorm
                0.5001983 = fieldWeight in 701, product of:
                  5.2915025 = tf(freq=28.0), with freq of:
                    28.0 = termFreq=28.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
            0.027504109 = weight(_text_:system in 701) [ClassicSimilarity], result of:
              0.027504109 = score(doc=701,freq=6.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.24108742 = fieldWeight in 701, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.6 = coord(3/5)
        0.115060724 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.115060724 = score(doc=701,freq=2.0), product of:
            0.3070917 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03622214 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.115060724 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.115060724 = score(doc=701,freq=2.0), product of:
            0.3070917 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03622214 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.115060724 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.115060724 = score(doc=701,freq=2.0), product of:
            0.3070917 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03622214 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
      0.5714286 = coord(4/7)
    
    Abstract
    By the explosion of possibilities for a ubiquitous content production, the information overload problem reaches the level of complexity which cannot be managed by traditional modelling approaches anymore. Due to their pure syntactical nature traditional information retrieval approaches did not succeed in treating content itself (i.e. its meaning, and not its representation). This leads to a very low usefulness of the results of a retrieval process for a user's task at hand. In the last ten years ontologies have been emerged from an interesting conceptualisation paradigm to a very promising (semantic) modelling technology, especially in the context of the Semantic Web. From the information retrieval point of view, ontologies enable a machine-understandable form of content description, such that the retrieval process can be driven by the meaning of the content. However, the very ambiguous nature of the retrieval process in which a user, due to the unfamiliarity with the underlying repository and/or query syntax, just approximates his information need in a query, implies a necessity to include the user in the retrieval process more actively in order to close the gap between the meaning of the content and the meaning of a user's query (i.e. his information need). This thesis lays foundation for such an ontology-based interactive retrieval process, in which the retrieval system interacts with a user in order to conceptually interpret the meaning of his query, whereas the underlying domain ontology drives the conceptualisation process. In that way the retrieval process evolves from a query evaluation process into a highly interactive cooperation between a user and the retrieval system, in which the system tries to anticipate the user's information need and to deliver the relevant content proactively. Moreover, the notion of content relevance for a user's query evolves from a content dependent artefact to the multidimensional context-dependent structure, strongly influenced by the user's preferences. This cooperation process is realized as the so-called Librarian Agent Query Refinement Process. In order to clarify the impact of an ontology on the retrieval process (regarding its complexity and quality), a set of methods and tools for different levels of content and query formalisation is developed, ranging from pure ontology-based inferencing to keyword-based querying in which semantics automatically emerges from the results. Our evaluation studies have shown that the possibilities to conceptualize a user's information need in the right manner and to interpret the retrieval results accordingly are key issues for realizing much more meaningful information retrieval systems.
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  2. Eckert, K.: SKOS: eine Sprache für die Übertragung von Thesauri ins Semantic Web (2011) 0.05
    0.046315853 = product of:
      0.16210547 = sum of:
        0.008982805 = product of:
          0.044914022 = sum of:
            0.044914022 = weight(_text_:system in 4331) [ClassicSimilarity], result of:
              0.044914022 = score(doc=4331,freq=4.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.3936941 = fieldWeight in 4331, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4331)
          0.2 = coord(1/5)
        0.15312266 = sum of:
          0.11386186 = weight(_text_:zugriff in 4331) [ClassicSimilarity], result of:
            0.11386186 = score(doc=4331,freq=2.0), product of:
              0.2160124 = queryWeight, product of:
                5.963546 = idf(docFreq=308, maxDocs=44218)
                0.03622214 = queryNorm
              0.52710795 = fieldWeight in 4331, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.963546 = idf(docFreq=308, maxDocs=44218)
                0.0625 = fieldNorm(doc=4331)
          0.0392608 = weight(_text_:22 in 4331) [ClassicSimilarity], result of:
            0.0392608 = score(doc=4331,freq=2.0), product of:
              0.12684377 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03622214 = queryNorm
              0.30952093 = fieldWeight in 4331, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=4331)
      0.2857143 = coord(2/7)
    
    Abstract
    Das Semantic Web - bzw. Linked Data - hat das Potenzial, die Verfügbarkeit von Daten und Wissen, sowie den Zugriff darauf zu revolutionieren. Einen großen Beitrag dazu können Wissensorganisationssysteme wie Thesauri leisten, die die Daten inhaltlich erschließen und strukturieren. Leider sind immer noch viele dieser Systeme lediglich in Buchform oder in speziellen Anwendungen verfügbar. Wie also lassen sie sich für das Semantic Web nutzen? Das Simple Knowledge Organization System (SKOS) bietet eine Möglichkeit, die Wissensorganisationssysteme in eine Form zu "übersetzen", die im Web zitiert und mit anderen Resourcen verknüpft werden kann.
    Date
    15. 3.2011 19:21:22
    Source
    http://metadaten-twr.org/2011/01/19/skos-simple-knowledge-organisation-system/
  3. Semantic web & linked data : Elemente zukünftiger Informationsinfrastrukturen ; 1. DGI-Konferenz ; 62. Jahrestagung der DGI ; Frankfurt am Main, 7. - 9. Oktober 2010 ; Proceedings / Deutsche Gesellschaft für Informationswissenschaft und Informationspraxis (2010) 0.03
    0.027792523 = product of:
      0.09727383 = sum of:
        0.010978288 = product of:
          0.027445719 = sum of:
            0.01553609 = weight(_text_:retrieval in 1516) [ClassicSimilarity], result of:
              0.01553609 = score(doc=1516,freq=4.0), product of:
                0.109568894 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03622214 = queryNorm
                0.1417929 = fieldWeight in 1516, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1516)
            0.011909628 = weight(_text_:system in 1516) [ClassicSimilarity], result of:
              0.011909628 = score(doc=1516,freq=2.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.104393914 = fieldWeight in 1516, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1516)
          0.4 = coord(2/5)
        0.08629554 = weight(_text_:mehrsprachigkeit in 1516) [ClassicSimilarity], result of:
          0.08629554 = score(doc=1516,freq=2.0), product of:
            0.3070917 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03622214 = queryNorm
            0.28100902 = fieldWeight in 1516, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1516)
      0.2857143 = coord(2/7)
    
    Abstract
    Informationswissenschaft und Informatik nähern sich auf vielen Gebieten bei ihren Projekten und in ihren Zielsetzungen immer stärker an. Ganz besonders deutlich wird dies vor dem Hintergrund von Semantic Web und Linked Data. Die Textanalyse und Fixierung der Essenz einer Publikation in Form ihrer belastbaren Aussagen und zugrunde liegender Daten sowie deren Speicherung und Aufbereitung für die Nutzung gehören zu den professionellen Aktivitäten der in der DGI organisierten Informationsfachleute. Das Semantic Web versprocht, bei entsprechender intellektueller Vorarbeit, diese Tätigkeiten künftig zu erleichtern. Der Tgaungsband macht an vielen Beispielen anschaulich klar, wie die in Informationswissenschaft und -praxis über Jahrzehnte erarbeiteten Grundlagen und Werkzeuge jetzt für den Aufbau der zukünftigen Informationsinfrastrukturen eingesetzt werden können. Er vereint die Textfassungen der Beiträge, die durch das Programmkomitee für die 1. DGI-Konferenz angenommen worden sind, dazu gehören u.a. wissensbasierte Anwendungen in der Wirtschaft, Aufbau von Ontologien mittels Thesauri, Mehrsprachigkeit und Open Data.
    Content
    LINKED DATA IM GEOINFORMATIONSBEREICH - CHANCEN ODER GEFAHR? Geodaten - von der Verantwortung des Dealers / Karsten Neumann - Computergestützte Freizeitplanung basierend auf Points Of Interest / Peter Bäcker und Ugur Macit VON LINKED DATA ZU VERLINKTEN DIALOGEN Die globalisierte Semantic Web Informationswissenschaftlerin / Dierk Eichel - Kommunikation und Kontext. Überlegungen zur Entwicklung virtueller Diskursräume für die Wissenschaft / Ben Kaden und Maxi Kindling - Konzeptstudie: Die informationswissenschaftliche Zeitschrift der Zukunft / Lambert Heller und Heinz Pampel SEMANTIC WEB & LINKED DATA IM BILDUNGSWESEN Einsatz von Semantic Web-Technologien am Informationszentrum Bildung / Carola Carstens und Marc Rittberger - Bedarfsgerecht, kontextbezogen, qualitätsgesichert: Von der Information zum Wertschöpfungsfaktor Wissen am Beispiel einer Wissenslandkarte als dynamisches System zur Repräsentation des Wissens in der Berufsbildungsforschung / Sandra Dücker und Markus Linten - Virtuelle Forschungsumgebungen und Forschungsdaten für Lehre und Forschung: Informationsinfrastrukturen für die (Natur-)Wissenschaften / Matthias Schulze
    RSWK
    Semantic Web / Indexierung <Inhaltserschließung> / Information Retrieval / Kongress / Frankfurt <Main, 2010>
    Subject
    Semantic Web / Indexierung <Inhaltserschließung> / Information Retrieval / Kongress / Frankfurt <Main, 2010>
  4. Borst, T.; Fingerle, B.; Neubert, J.; Seiler, A.: Auf dem Weg in das Semantic Web : Anwendungsbeispiele und Lösungsszenarien in Bibliotheken / Eine Veranstaltung von hbz und ZBW (2010) 0.02
    0.022879425 = product of:
      0.08007798 = sum of:
        0.06228707 = weight(_text_:abfrage in 4398) [ClassicSimilarity], result of:
          0.06228707 = score(doc=4398,freq=2.0), product of:
            0.28580084 = queryWeight, product of:
              7.890225 = idf(docFreq=44, maxDocs=44218)
              0.03622214 = queryNorm
            0.21793872 = fieldWeight in 4398, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.890225 = idf(docFreq=44, maxDocs=44218)
              0.01953125 = fieldNorm(doc=4398)
        0.017790915 = product of:
          0.03558183 = sum of:
            0.03558183 = weight(_text_:zugriff in 4398) [ClassicSimilarity], result of:
              0.03558183 = score(doc=4398,freq=2.0), product of:
                0.2160124 = queryWeight, product of:
                  5.963546 = idf(docFreq=308, maxDocs=44218)
                  0.03622214 = queryNorm
                0.16472124 = fieldWeight in 4398, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.963546 = idf(docFreq=308, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=4398)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Content
    Moderne Web-Anwendungen Wie ein klassischer Bibliothekskatalog in eine moderne Web-Anwendung und in Linked Open Data überführt werden kann, berichtete Anders Söderbäck von der Schwedischen Nationalbibliothek. Unter dem Thema »Linked Applications« entwarf Felix Ostrowski, IT-Entwickler beim hbz, eine tiefe Integration von Ontologien direkt in die verwendete Programmiersprache am Beispiel von Repository Software. Andre Hagenbruch, Mitarbeiter der Universitätsbibliothek Bochum, stellte ein Projekt für ein Linked Data-basiertes Bibliotheksportal vor, das neben bibliografischen Daten auch administrative Informationen wie Personaldaten oder Öffnungszeiten integrieren soll. Interessante Einblicke bot Jürgen Kett von der DNB zur hauseigenen Linked Data-Strategie. Dazu gehört die Veröffentlichung der Personennamendatei (PND) zwecks kostenfreier Weiterverwendung durch andere Bibliotheken. In einem weiteren Vortrag gingen Kett und sein Kollege Svensson auf Details ihres aktuellen Linked Data-Projekts ein, das in einer ersten Stufe URI- und SPARQL-Zugriff auf die SWDund PND-Datensets sowie ihre Verknüpfungen zu internationalen Normdatenbeständen ermöglichen soll. Ein Beta-Dienst soll bis Mitte 2010 online gehen. Timo Borst, Leiter der IT-Entwicklung der ZBW, griff dies auf und demonstrierte die Integration von Thesaurus-und Personennormdaten in den ZBW-Dokumentenserver. Als technischer Zugangspunkt für Linked Data wurde dabei auch ein experimenteller SPARQLEndpoint der DNB genutzt, dessen Abfrage eine automatische Vorschlagsfunktion für PND-Einträge ermöglicht. Praktische Erfahrungen aus den Linked Data-Projekten der ZBW fasste Joachim Neubert, IT-Entwickler bei der ZBW, zusammen, verbunden mit Hinweisen zu Tools, Communities und Best Practice-Beispielen im Web.
  5. Borst, T.; Löhden, A.; Neubert, J.; Pohl, A.: "Linked Open Data" im Fokus : Spannende Themen und Diskussionen bei der SWIB12 (2013) 0.01
    0.012327935 = product of:
      0.08629554 = sum of:
        0.08629554 = weight(_text_:mehrsprachigkeit in 3352) [ClassicSimilarity], result of:
          0.08629554 = score(doc=3352,freq=2.0), product of:
            0.3070917 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03622214 = queryNorm
            0.28100902 = fieldWeight in 3352, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0234375 = fieldNorm(doc=3352)
      0.14285715 = coord(1/7)
    
    Content
    Inzwischen geht es nicht mehr nur um die Veröffentlichung von Linked Data, sondern es werden auch erweiterte Services für BibliotheksnutzerInnen vorgestellt. In der Osloer Deichmanske Bibliothek werden Buchrezensionen strukturiert auffindbar gemacht und den Büchern zugeordnet. An den Regalen bietet, vermittelt durch in den Medien enthaltene RFID-Tags, ein Monitor Medieninformationen aus Bibliotheks- und externen Quellen. Beim LOD-Dienst der Französischen Nationalbibliothek können Nutzerinnen nun auch explizit nach Werken, Editionen und einzelnen Exemplaren recherchieren oder nach Personen, die mit einer bestimmten Rolle an einer Ausgabe beteiligt sind. Bei der Recherche über Themen werden zu einem Stichwort feinere oder gröbere Themen vorgeschlagen. Die semantische Verknüpfung von Termen verbessert die bibliothekarische Erschließung und Recherche, die Anreicherung von Ressourcen sowie Gestaltung von Oberflächen über verschiedene Begriffssysteme und Sprachen hinweg. Mit automatischen Verfahren zur Pflege von Crosskonkordanzen zwischen Thesauri befasst sich unter anderem die »Ontology Alignment Evaluation Initiative« (OAEI), hier wurden Unterschiede von Ontologie-Werkzeugen und Formaten (SKOS, OWL) betrachtet. Ein noch bestehender Mangel an Mehrsprachigkeit von Ontologien und Formaten wurde zum Beispiel im Vortrag der Ontology Engineering Group der Polytechnischen Universität Madrid (OEG-UPM) angedeutet.
  6. Franklin, R.A.: Re-inventing subject access for the semantic web (2003) 0.01
    0.012161928 = product of:
      0.042566746 = sum of:
        0.027843947 = product of:
          0.069609866 = sum of:
            0.02197135 = weight(_text_:retrieval in 2556) [ClassicSimilarity], result of:
              0.02197135 = score(doc=2556,freq=2.0), product of:
                0.109568894 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03622214 = queryNorm
                0.20052543 = fieldWeight in 2556, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2556)
            0.047638513 = weight(_text_:system in 2556) [ClassicSimilarity], result of:
              0.047638513 = score(doc=2556,freq=8.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.41757566 = fieldWeight in 2556, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2556)
          0.4 = coord(2/5)
        0.0147228 = product of:
          0.0294456 = sum of:
            0.0294456 = weight(_text_:22 in 2556) [ClassicSimilarity], result of:
              0.0294456 = score(doc=2556,freq=2.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.23214069 = fieldWeight in 2556, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2556)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    First generation scholarly research on the Web lacked a firm system of authority control. Second generation Web research is beginning to model subject access with library science principles of bibliographic control and cataloguing. Harnessing the Web and organising the intellectual content with standards and controlled vocabulary provides precise search and retrieval capability, increasing relevance and efficient use of technology. Dublin Core metadata standards permit a full evaluation and cataloguing of Web resources appropriate to highly specific research needs and discovery. Current research points to a type of structure based on a system of faceted classification. This system allows the semantic and syntactic relationships to be defined. Controlled vocabulary, such as the Library of Congress Subject Headings, can be assigned, not in a hierarchical structure, but rather as descriptive facets of relating concepts. Web design features such as this are adding value to discovery and filtering out data that lack authority. The system design allows for scalability and extensibility, two technical features that are integral to future development of the digital library and resource discovery.
    Date
    30.12.2008 18:22:46
  7. Tudhope, D.: Knowledge Organization System Services : brief review of NKOS activities and possibility of KOS registries (2007) 0.01
    0.01113523 = product of:
      0.0389733 = sum of:
        0.009527703 = product of:
          0.047638513 = sum of:
            0.047638513 = weight(_text_:system in 100) [ClassicSimilarity], result of:
              0.047638513 = score(doc=100,freq=2.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.41757566 = fieldWeight in 100, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.09375 = fieldNorm(doc=100)
          0.2 = coord(1/5)
        0.0294456 = product of:
          0.0588912 = sum of:
            0.0588912 = weight(_text_:22 in 100) [ClassicSimilarity], result of:
              0.0588912 = score(doc=100,freq=2.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.46428138 = fieldWeight in 100, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=100)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Date
    22. 9.2007 15:41:14
  8. Mayfield, J.; Finin, T.: Information retrieval on the Semantic Web : integrating inference and retrieval 0.01
    0.008182896 = product of:
      0.028640134 = sum of:
        0.011463534 = product of:
          0.05731767 = sum of:
            0.05731767 = weight(_text_:retrieval in 4330) [ClassicSimilarity], result of:
              0.05731767 = score(doc=4330,freq=10.0), product of:
                0.109568894 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03622214 = queryNorm
                0.5231199 = fieldWeight in 4330, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4330)
          0.2 = coord(1/5)
        0.0171766 = product of:
          0.0343532 = sum of:
            0.0343532 = weight(_text_:22 in 4330) [ClassicSimilarity], result of:
              0.0343532 = score(doc=4330,freq=2.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.2708308 = fieldWeight in 4330, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4330)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    One vision of the Semantic Web is that it will be much like the Web we know today, except that documents will be enriched by annotations in machine understandable markup. These annotations will provide metadata about the documents as well as machine interpretable statements capturing some of the meaning of document content. We discuss how the information retrieval paradigm might be recast in such an environment. We suggest that retrieval can be tightly bound to inference. Doing so makes today's Web search engines useful to Semantic Web inference engines, and causes improvements in either retrieval or inference to lead directly to improvements in the other.
    Date
    12. 2.2011 17:35:22
  9. Multimedia content and the Semantic Web : methods, standards, and tools (2005) 0.01
    0.007308805 = product of:
      0.025580816 = sum of:
        0.01495555 = product of:
          0.037388876 = sum of:
            0.027464187 = weight(_text_:retrieval in 150) [ClassicSimilarity], result of:
              0.027464187 = score(doc=150,freq=18.0), product of:
                0.109568894 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03622214 = queryNorm
                0.25065678 = fieldWeight in 150, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=150)
            0.00992469 = weight(_text_:system in 150) [ClassicSimilarity], result of:
              0.00992469 = score(doc=150,freq=2.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.08699492 = fieldWeight in 150, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=150)
          0.4 = coord(2/5)
        0.0106252665 = product of:
          0.021250533 = sum of:
            0.021250533 = weight(_text_:22 in 150) [ClassicSimilarity], result of:
              0.021250533 = score(doc=150,freq=6.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.16753313 = fieldWeight in 150, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=150)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Classification
    006.7 22
    Date
    7. 3.2007 19:30:22
    DDC
    006.7 22
    Footnote
    Rez. in: JASIST 58(2007) no.3, S.457-458 (A.M.A. Ahmad): "The concept of the semantic web has emerged because search engines and text-based searching are no longer adequate, as these approaches involve an extensive information retrieval process. The deployed searching and retrieving descriptors arc naturally subjective and their deployment is often restricted to the specific application domain for which the descriptors were configured. The new era of information technology imposes different kinds of requirements and challenges. Automatic extracted audiovisual features are required, as these features are more objective, domain-independent, and more native to audiovisual content. This book is a useful guide for researchers, experts, students, and practitioners; it is a very valuable reference and can lead them through their exploration and research in multimedia content and the semantic web. The book is well organized, and introduces the concept of the semantic web and multimedia content analysis to the reader through a logical sequence from standards and hypotheses through system examples, presenting relevant tools and methods. But in some chapters readers will need a good technical background to understand some of the details. Readers may attain sufficient knowledge here to start projects or research related to the book's theme; recent results and articles related to the active research area of integrating multimedia with semantic web technologies are included. This book includes full descriptions of approaches to specific problem domains such as content search, indexing, and retrieval. This book will be very useful to researchers in the multimedia content analysis field who wish to explore the benefits of emerging semantic web technologies in applying multimedia content approaches. The first part of the book covers the definition of the two basic terms multimedia content and semantic web. The Moving Picture Experts Group standards MPEG7 and MPEG21 are quoted extensively. In addition, the means of multimedia content description are elaborated upon and schematically drawn. This extensive description is introduced by authors who are actively involved in those standards and have been participating in the work of the International Organization for Standardization (ISO)/MPEG for many years. On the other hand, this results in bias against the ad hoc or nonstandard tools for multimedia description in favor of the standard approaches. This is a general book for multimedia content; more emphasis on the general multimedia description and extraction could be provided.
    The final part of the book discusses research in multimedia content management systems and the semantic web, and presents examples and applications for semantic multimedia analysis in search and retrieval systems. These chapters describe example systems in which current projects have been implemented, and include extensive results and real demonstrations. For example, real case scenarios such as ECommerce medical applications and Web services have been introduced. Topics in natural language, speech and image processing techniques and their application for multimedia indexing, and content-based retrieval have been elaborated upon with extensive examples and deployment methods. The editors of the book themselves provide the readers with a chapter about their latest research results on knowledge-based multimedia content indexing and retrieval. Some interesting applications for multimedia content and the semantic web are introduced. Applications that have taken advantage of the metadata provided by MPEG7 in order to realize advance-access services for multimedia content have been provided. The applications discussed in the third part of the book provide useful guidance to researchers and practitioners properly planning to implement semantic multimedia analysis techniques in new research and development projects in both academia and industry. A fourth part should be added to this book: performance measurements for integrated approaches of multimedia analysis and the semantic web. Performance of the semantic approach is a very sophisticated issue and requires extensive elaboration and effort. Measuring the semantic search is an ongoing research area; several chapters concerning performance measurement and analysis would be required to adequately cover this area and introduce it to readers."
    LCSH
    Information storage and retrieval systems
    RSWK
    Semantic Web / Multimedia / Automatische Indexierung / Information Retrieval
    Subject
    Semantic Web / Multimedia / Automatische Indexierung / Information Retrieval
    Information storage and retrieval systems
  10. Subirats, I.; Prasad, A.R.D.; Keizer, J.; Bagdanov, A.: Implementation of rich metadata formats and demantic tools using DSpace (2008) 0.01
    0.0070448667 = product of:
      0.024657032 = sum of:
        0.014841831 = product of:
          0.037104577 = sum of:
            0.014647567 = weight(_text_:retrieval in 2656) [ClassicSimilarity], result of:
              0.014647567 = score(doc=2656,freq=2.0), product of:
                0.109568894 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03622214 = queryNorm
                0.13368362 = fieldWeight in 2656, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2656)
            0.022457011 = weight(_text_:system in 2656) [ClassicSimilarity], result of:
              0.022457011 = score(doc=2656,freq=4.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.19684705 = fieldWeight in 2656, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2656)
          0.4 = coord(2/5)
        0.0098152 = product of:
          0.0196304 = sum of:
            0.0196304 = weight(_text_:22 in 2656) [ClassicSimilarity], result of:
              0.0196304 = score(doc=2656,freq=2.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.15476047 = fieldWeight in 2656, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2656)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    This poster explores the customization of DSpace to allow the use of the AGRIS Application Profile metadata standard and the AGROVOC thesaurus. The objective is the adaptation of DSpace, through the least invasive code changes either in the form of plug-ins or add-ons, to the specific needs of the Agricultural Sciences and Technology community. Metadata standards such as AGRIS AP, and Knowledge Organization Systems such as the AGROVOC thesaurus, provide mechanisms for sharing information in a standardized manner by recommending the use of common semantics and interoperable syntax (Subirats et al., 2007). AGRIS AP was created to enhance the description, exchange and subsequent retrieval of agricultural Document-like Information Objects (DLIOs). It is a metadata schema which draws from Metadata standards such as Dublin Core (DC), the Australian Government Locator Service Metadata (AGLS) and the Agricultural Metadata Element Set (AgMES) namespaces. It allows sharing of information across dispersed bibliographic systems (FAO, 2005). AGROVOC68 is a multilingual structured thesaurus covering agricultural and related domains. Its main role is to standardize the indexing process in order to make searching simpler and more efficient. AGROVOC is developed by FAO (Lauser et al., 2006). The customization of the DSpace is taking place in several phases. First, the AGRIS AP metadata schema was mapped onto the metadata DSpace model, with several enhancements implemented to support AGRIS AP elements. Next, AGROVOC will be integrated as a controlled vocabulary accessed through a local SKOS or OWL file. Eventually the system will be configurable to access AGROVOC through local files or remotely via webservices. Finally, spell checking and tooltips will be incorporated in the user interface to support metadata editing. Adapting DSpace to support AGRIS AP and annotation using the semantically-rich AGROVOC thesaurus transform DSpace into a powerful, domain-specific system for annotation and exchange of bibliographic metadata in the agricultural domain.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  11. Faaborg, A.; Lagoze, C.: Semantic browsing (2003) 0.01
    0.0063723573 = product of:
      0.02230325 = sum of:
        0.0051266486 = product of:
          0.025633242 = sum of:
            0.025633242 = weight(_text_:retrieval in 1026) [ClassicSimilarity], result of:
              0.025633242 = score(doc=1026,freq=2.0), product of:
                0.109568894 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03622214 = queryNorm
                0.23394634 = fieldWeight in 1026, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1026)
          0.2 = coord(1/5)
        0.0171766 = product of:
          0.0343532 = sum of:
            0.0343532 = weight(_text_:22 in 1026) [ClassicSimilarity], result of:
              0.0343532 = score(doc=1026,freq=2.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.2708308 = fieldWeight in 1026, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1026)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Source
    Research and advanced technology for digital libraries : 7th European Conference, proceedings / ECDL 2003, Trondheim, Norway, August 17-22, 2003
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  12. Metadata and semantics research : 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings (2016) 0.01
    0.0063723573 = product of:
      0.02230325 = sum of:
        0.0051266486 = product of:
          0.025633242 = sum of:
            0.025633242 = weight(_text_:retrieval in 3283) [ClassicSimilarity], result of:
              0.025633242 = score(doc=3283,freq=2.0), product of:
                0.109568894 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03622214 = queryNorm
                0.23394634 = fieldWeight in 3283, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3283)
          0.2 = coord(1/5)
        0.0171766 = product of:
          0.0343532 = sum of:
            0.0343532 = weight(_text_:22 in 3283) [ClassicSimilarity], result of:
              0.0343532 = score(doc=3283,freq=2.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.2708308 = fieldWeight in 3283, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3283)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    This book constitutes the refereed proceedings of the 10th Metadata and Semantics Research Conference, MTSR 2016, held in Göttingen, Germany, in November 2016. The 26 full papers and 6 short papers presented were carefully reviewed and selected from 67 submissions. The papers are organized in several sessions and tracks: Digital Libraries, Information Retrieval, Linked and Social Data, Metadata and Semantics for Open Repositories, Research Information Systems and Data Infrastructures, Metadata and Semantics for Agriculture, Food and Environment, Metadata and Semantics for Cultural Collections and Applications, European and National Projects.
  13. Kara, S.: ¬An ontology-based retrieval system using semantic indexing (2012) 0.01
    0.0052332124 = product of:
      0.036632486 = sum of:
        0.036632486 = product of:
          0.09158121 = sum of:
            0.0439427 = weight(_text_:retrieval in 3829) [ClassicSimilarity], result of:
              0.0439427 = score(doc=3829,freq=8.0), product of:
                0.109568894 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03622214 = queryNorm
                0.40105087 = fieldWeight in 3829, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3829)
            0.047638513 = weight(_text_:system in 3829) [ClassicSimilarity], result of:
              0.047638513 = score(doc=3829,freq=8.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.41757566 = fieldWeight in 3829, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3829)
          0.4 = coord(2/5)
      0.14285715 = coord(1/7)
    
    Abstract
    In this thesis, we present an ontology-based information extraction and retrieval system and its application to soccer domain. In general, we deal with three issues in semantic search, namely, usability, scalability and retrieval performance. We propose a keyword-based semantic retrieval approach. The performance of the system is improved considerably using domain-specific information extraction, inference and rules. Scalability is achieved by adapting a semantic indexing approach. The system is implemented using the state-of-the-art technologies in SemanticWeb and its performance is evaluated against traditional systems as well as the query expansion methods. Furthermore, a detailed evaluation is provided to observe the performance gain due to domain-specific information extraction and inference. Finally, we show how we use semantic indexing to solve simple structural ambiguities.
  14. Dextre Clarke, S.G.: Challenges and opportunities for KOS standards (2007) 0.00
    0.0049076 = product of:
      0.0343532 = sum of:
        0.0343532 = product of:
          0.0687064 = sum of:
            0.0687064 = weight(_text_:22 in 4643) [ClassicSimilarity], result of:
              0.0687064 = score(doc=4643,freq=2.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.5416616 = fieldWeight in 4643, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4643)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    Date
    22. 9.2007 15:41:14
  15. Metadata and semantics research : 7th Research Conference, MTSR 2013 Thessaloniki, Greece, November 19-22, 2013. Proceedings (2013) 0.00
    0.004845403 = product of:
      0.016958911 = sum of:
        0.0048132194 = product of:
          0.024066096 = sum of:
            0.024066096 = weight(_text_:system in 1155) [ClassicSimilarity], result of:
              0.024066096 = score(doc=1155,freq=6.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.21095149 = fieldWeight in 1155, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1155)
          0.2 = coord(1/5)
        0.012145691 = product of:
          0.024291381 = sum of:
            0.024291381 = weight(_text_:22 in 1155) [ClassicSimilarity], result of:
              0.024291381 = score(doc=1155,freq=4.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.19150631 = fieldWeight in 1155, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1155)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    Metadata and semantics are integral to any information system and significant to the sphere of Web data. Research focusing on metadata and semantics is crucial for advancing our understanding and knowledge of metadata; and, more profoundly for being able to effectively discover, use, archive, and repurpose information. In response to this need, researchers are actively examining methods for generating, reusing, and interchanging metadata. Integrated with these developments is research on the application of computational methods, linked data, and data analytics. A growing body of work also targets conceptual and theoretical designs providing foundational frameworks for metadata and semantic applications. There is no doubt that metadata weaves its way into nearly every aspect of our information ecosystem, and there is great motivation for advancing the current state of metadata and semantics. To this end, it is vital that scholars and practitioners convene and share their work.
    The MTSR 2013 program and the contents of these proceedings show a rich diversity of research and practices, drawing on problems from metadata and semantically focused tools and technologies, linked data, cross-language semantics, ontologies, metadata models, and semantic system and metadata standards. The general session of the conference included 18 papers covering a broad spectrum of topics, proving the interdisciplinary field of metadata, and was divided into three main themes: platforms for research data sets, system architecture and data management; metadata and ontology validation, evaluation, mapping and interoperability; and content management. Metadata as a research topic is maturing, and the conference also supported the following five tracks: Metadata and Semantics for Open Repositories, Research Information Systems and Data Infrastructures; Metadata and Semantics for Cultural Collections and Applications; Metadata and Semantics for Agriculture, Food and Environment; Big Data and Digital Libraries in Health, Science and Technology; and European and National Projects, and Project Networking. Each track had a rich selection of papers, giving broader diversity to MTSR, and enabling deeper exploration of significant topics.
    Date
    17.12.2013 12:51:22
  16. Monireh, E.; Sarker, M.K.; Bianchi, F.; Hitzler, P.; Doran, D.; Xie, N.: Reasoning over RDF knowledge bases using deep learning (2018) 0.00
    0.004639679 = product of:
      0.016238876 = sum of:
        0.003969876 = product of:
          0.01984938 = sum of:
            0.01984938 = weight(_text_:system in 4553) [ClassicSimilarity], result of:
              0.01984938 = score(doc=4553,freq=2.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.17398985 = fieldWeight in 4553, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4553)
          0.2 = coord(1/5)
        0.0122690005 = product of:
          0.024538001 = sum of:
            0.024538001 = weight(_text_:22 in 4553) [ClassicSimilarity], result of:
              0.024538001 = score(doc=4553,freq=2.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.19345059 = fieldWeight in 4553, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4553)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    Semantic Web knowledge representation standards, and in particular RDF and OWL, often come endowed with a formal semantics which is considered to be of fundamental importance for the field. Reasoning, i.e., the drawing of logical inferences from knowledge expressed in such standards, is traditionally based on logical deductive methods and algorithms which can be proven to be sound and complete and terminating, i.e. correct in a very strong sense. For various reasons, though, in particular the scalability issues arising from the ever increasing amounts of Semantic Web data available and the inability of deductive algorithms to deal with noise in the data, it has been argued that alternative means of reasoning should be investigated which bear high promise for high scalability and better robustness. From this perspective, deductive algorithms can be considered the gold standard regarding correctness against which alternative methods need to be tested. In this paper, we show that it is possible to train a Deep Learning system on RDF knowledge graphs, such that it is able to perform reasoning over new RDF knowledge graphs, with high precision and recall compared to the deductive gold standard.
    Date
    16.11.2018 14:22:01
  17. Shaw, R.; Buckland, M.: Open identification and linking of the four Ws (2008) 0.00
    0.0043986347 = product of:
      0.01539522 = sum of:
        0.00680692 = product of:
          0.0340346 = sum of:
            0.0340346 = weight(_text_:system in 2665) [ClassicSimilarity], result of:
              0.0340346 = score(doc=2665,freq=12.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.29833046 = fieldWeight in 2665, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=2665)
          0.2 = coord(1/5)
        0.0085883 = product of:
          0.0171766 = sum of:
            0.0171766 = weight(_text_:22 in 2665) [ClassicSimilarity], result of:
              0.0171766 = score(doc=2665,freq=2.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.1354154 = fieldWeight in 2665, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=2665)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    Platforms for social computing connect users via shared references to people with whom they have relationships, events attended, places lived in or traveled to, and topics such as favorite books or movies. Since free text is insufficient for expressing such references precisely and unambiguously, many social computing platforms coin identifiers for topics, places, events, and people and provide interfaces for finding and selecting these identifiers from controlled lists. Using these interfaces, users collaboratively construct a web of links among entities. This model needn't be limited to social networking sites. Understanding an item in a digital library or museum requires context: information about the topics, places, events, and people to which the item is related. Students, journalists and investigators traditionally discover this kind of context by asking "the four Ws": what, where, when and who. The DCMI Kernel Metadata Community has recognized the four Ws as fundamental elements of descriptions (Kunze & Turner, 2007). Making better use of metadata to answer these questions via links to appropriate contextual resources has been our focus in a series of research projects over the past few years. Currently we are building a system for enabling readers of any text to relate any topic, place, event or person mentioned in the text to the best explanatory resources available. This system is being developed with two different corpora: a diverse variety of biographical texts characterized by very rich and dense mentions of people, events, places and activities, and a large collection of newly-scanned books, journals and manuscripts relating to Irish culture and history. Like a social computing platform, our system consists of tools for referring to topics, places, events or people, disambiguating these references by linking them to unique identifiers, and using the disambiguated references to provide useful information in context and to link to related resources. Yet current social computing platforms, while usually amenable to importing and exporting data, tend to mint proprietary identifiers and expect links to be traversed using their own interfaces. We take a different approach, using identifiers from both established and emerging naming authorities, representing relationships using standardized metadata vocabularies, and publishing those representations using standard protocols so that links can be stored and traversed anywhere. Central to our strategy is to move from appearances in a text to naming authorities to the the construction of links for searching or querying trusted resources. Using identifiers from naming authorities, rather than literal values (as in the DCMI Kernel) or keys from a proprietary database, makes it more likely that links constructed using our system will continue to be useful in the future. WorldCat Identities URIs (http://worldcat.org/identities/) linked to Library of Congress and Deutsche Nationalbibliothek authority files for persons and organizations and Geonames (http://geonames.org/) URIs for places are stable identifiers attached to a wealth of useful metadata. Yet no naming authority can be totally comprehensive, so our system can be extended to use new sources of identifiers as needed. For example, we are experimenting with using Freebase (http://freebase.com/) URIs to identify historical events, for which no established naming authority currently exists. Stable identifiers (URIs), standardized hyperlinked data formats (XML), and uniform publishing protocols (HTTP) are key ingredients of the web's open architecture. Our system provides an example of how this open architecture can be exploited to build flexible and useful tools for connecting resources via shared references to topics, places, events, and people.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  18. Web 2.0-Pionier ortet kollektive Intelligenz im Internet : Technologieschübe kommen über mobile Endgeräte und Spracherkennung (2007) 0.00
    0.004313169 = product of:
      0.030192183 = sum of:
        0.030192183 = product of:
          0.060384367 = sum of:
            0.060384367 = weight(_text_:zugriff in 670) [ClassicSimilarity], result of:
              0.060384367 = score(doc=670,freq=4.0), product of:
                0.2160124 = queryWeight, product of:
                  5.963546 = idf(docFreq=308, maxDocs=44218)
                  0.03622214 = queryNorm
                0.2795412 = fieldWeight in 670, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.963546 = idf(docFreq=308, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=670)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    Content
    Die Informationen aus dem Internet werden nach seiner Erkenntnis auf viele verschiedene Arten zugänglich sein. "Wir denken immer, es gibt Milliarden Computer da draußen - aber das stimmt nicht. Es gibt eigentlich nur einen, und darum geht es im Web 2.0. Alles wird mit allem verbunden. Und was wir heute unter einem Computer verstehen ist eigentlich nur ein Zugangsgerät zu dem einen weltweiten elektronischen Gehirn, das wir erschaffen", sagt O'Reilly. Webexperten halten die Visionen von O'Reilly für realistisch: "So wird die Spracherkennung eine Schlüsseltechnologie sein für den Zugriff auf das vernetzte Weltwissen. Wie häufig hätten wir gerne im Alltag genaue Informationen zu Sportergebnissen, Telefonnummern oder Adressen. Unendlich viele Dinge, die wir, wenn wir gerade im Internet wären sofort 'er-googlen' würden. Da das aber in der Freizeit und von unterwegs selten der Fall ist und der Zugriff über das Handy mit Tastatur oder Touch-screen zu mühselig ist, verzichten wir meistens darauf unseren Wissenshunger sofort zu stillen. Anders wäre es, wenn wir mit einfachen gesprochenen Suchbefehlen unsere Anfrage starten und die Suche dann bei Bedarf eingrenzen könnten, genauso, wie wir es derzeitig mit der PC-Tastatur und der Maus tun und das ganze jederzeit und von jedem Ort aus", kommentiert Lupo Pape, Geschäftsführer von SemanticEdge, die Analysen des Web-2.0-Vordenkers.
  19. Broughton, V.: Automatic metadata generation : Digital resource description without human intervention (2007) 0.00
    0.0042065145 = product of:
      0.0294456 = sum of:
        0.0294456 = product of:
          0.0588912 = sum of:
            0.0588912 = weight(_text_:22 in 6048) [ClassicSimilarity], result of:
              0.0588912 = score(doc=6048,freq=2.0), product of:
                0.12684377 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03622214 = queryNorm
                0.46428138 = fieldWeight in 6048, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6048)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    Date
    22. 9.2007 15:41:14
  20. Semantic applications (2018) 0.00
    0.004166863 = product of:
      0.029168038 = sum of:
        0.029168038 = product of:
          0.07292009 = sum of:
            0.04484883 = weight(_text_:retrieval in 5204) [ClassicSimilarity], result of:
              0.04484883 = score(doc=5204,freq=12.0), product of:
                0.109568894 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03622214 = queryNorm
                0.40932083 = fieldWeight in 5204, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5204)
            0.028071264 = weight(_text_:system in 5204) [ClassicSimilarity], result of:
              0.028071264 = score(doc=5204,freq=4.0), product of:
                0.11408355 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03622214 = queryNorm
                0.24605882 = fieldWeight in 5204, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5204)
          0.4 = coord(2/5)
      0.14285715 = coord(1/7)
    
    LCSH
    Information storage and retrieval
    Information Storage and Retrieval
    RSWK
    Wissensbasiertes System
    Information Retrieval
    Subject
    Wissensbasiertes System
    Information Retrieval
    Information storage and retrieval
    Information Storage and Retrieval

Years

Languages

  • e 116
  • d 20

Types

  • a 77
  • el 42
  • m 30
  • s 17
  • x 5
  • n 2
  • r 2
  • More… Less…

Subjects

Classifications