Search (16 results, page 1 of 1)

  • × year_i:[2010 TO 2020}
  • × theme_ss:"Semantic Web"
  1. Keyser, P. de: Indexing : from thesauri to the Semantic Web (2012) 0.07
    0.06706972 = product of:
      0.16767429 = sum of:
        0.1259246 = weight(_text_:thesaurus in 3197) [ClassicSimilarity], result of:
          0.1259246 = score(doc=3197,freq=6.0), product of:
            0.23732872 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.051357865 = queryNorm
            0.5305915 = fieldWeight in 3197, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.046875 = fieldNorm(doc=3197)
        0.04174969 = weight(_text_:22 in 3197) [ClassicSimilarity], result of:
          0.04174969 = score(doc=3197,freq=2.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.23214069 = fieldWeight in 3197, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=3197)
      0.4 = coord(2/5)
    
    Date
    24. 8.2016 14:03:22
    RSWK
    Thesaurus
    Subject
    Thesaurus
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  2. Hooland, S. van; Verborgh, R.; Wilde, M. De; Hercher, J.; Mannens, E.; Wa, R.Van de: Evaluating the success of vocabulary reconciliation for cultural heritage collections (2013) 0.06
    0.057826675 = product of:
      0.14456668 = sum of:
        0.102817 = weight(_text_:thesaurus in 662) [ClassicSimilarity], result of:
          0.102817 = score(doc=662,freq=4.0), product of:
            0.23732872 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.051357865 = queryNorm
            0.4332261 = fieldWeight in 662, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.046875 = fieldNorm(doc=662)
        0.04174969 = weight(_text_:22 in 662) [ClassicSimilarity], result of:
          0.04174969 = score(doc=662,freq=2.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.23214069 = fieldWeight in 662, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=662)
      0.4 = coord(2/5)
    
    Abstract
    The concept of Linked Data has made its entrance in the cultural heritage sector due to its potential use for the integration of heterogeneous collections and deriving additional value out of existing metadata. However, practitioners and researchers alike need a better understanding of what outcome they can reasonably expect of the reconciliation process between their local metadata and established controlled vocabularies which are already a part of the Linked Data cloud. This paper offers an in-depth analysis of how a locally developed vocabulary can be successfully reconciled with the Library of Congress Subject Headings (LCSH) and the Arts and Architecture Thesaurus (AAT) through the help of a general-purpose tool for interactive data transformation (OpenRefine). Issues negatively affecting the reconciliation process are identified and solutions are proposed in order to derive maximum value from existing metadata and controlled vocabularies in an automated manner.
    Date
    22. 3.2013 19:29:20
    Object
    Arts and Architecture Thesaurus
  3. Brunetti, J.M.; Roberto García, R.: User-centered design and evaluation of overview components for semantic data exploration (2014) 0.03
    0.030520609 = product of:
      0.07630152 = sum of:
        0.048468396 = weight(_text_:thesaurus in 1626) [ClassicSimilarity], result of:
          0.048468396 = score(doc=1626,freq=2.0), product of:
            0.23732872 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.051357865 = queryNorm
            0.20422474 = fieldWeight in 1626, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.03125 = fieldNorm(doc=1626)
        0.027833126 = weight(_text_:22 in 1626) [ClassicSimilarity], result of:
          0.027833126 = score(doc=1626,freq=2.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.15476047 = fieldWeight in 1626, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.03125 = fieldNorm(doc=1626)
      0.4 = coord(2/5)
    
    Abstract
    Purpose - The growing volumes of semantic data available in the web result in the need for handling the information overload phenomenon. The potential of this amount of data is enormous but in most cases it is very difficult for users to visualize, explore and use this data, especially for lay-users without experience with Semantic Web technologies. The paper aims to discuss these issues. Design/methodology/approach - The Visual Information-Seeking Mantra "Overview first, zoom and filter, then details-on-demand" proposed by Shneiderman describes how data should be presented in different stages to achieve an effective exploration. The overview is the first user task when dealing with a data set. The objective is that the user is capable of getting an idea about the overall structure of the data set. Different information architecture (IA) components supporting the overview tasks have been developed, so they are automatically generated from semantic data, and evaluated with end-users. Findings - The chosen IA components are well known to web users, as they are present in most web pages: navigation bars, site maps and site indexes. The authors complement them with Treemaps, a visualization technique for displaying hierarchical data. These components have been developed following an iterative User-Centered Design methodology. Evaluations with end-users have shown that they get easily used to them despite the fact that they are generated automatically from structured data, without requiring knowledge about the underlying semantic technologies, and that the different overview components complement each other as they focus on different information search needs. Originality/value - Obtaining semantic data sets overviews cannot be easily done with the current semantic web browsers. Overviews become difficult to achieve with large heterogeneous data sets, which is typical in the Semantic Web, because traditional IA techniques do not easily scale to large data sets. There is little or no support to obtain overview information quickly and easily at the beginning of the exploration of a new data set. This can be a serious limitation when exploring a data set for the first time, especially for lay-users. The proposal is to reuse and adapt existing IA components to provide this overview to users and show that they can be generated automatically from the thesaurus and ontologies that structure semantic data while providing a comparable user experience to traditional web sites.
    Date
    20. 1.2015 18:30:22
  4. Rüther, M.; Fock, J.; Schultz-Krutisch, T.; Bandholtz, T.: Classification and reference vocabulary in linked environment data (2011) 0.03
    0.02518492 = product of:
      0.1259246 = sum of:
        0.1259246 = weight(_text_:thesaurus in 4816) [ClassicSimilarity], result of:
          0.1259246 = score(doc=4816,freq=6.0), product of:
            0.23732872 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.051357865 = queryNorm
            0.5305915 = fieldWeight in 4816, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.046875 = fieldNorm(doc=4816)
      0.2 = coord(1/5)
    
    Abstract
    The Federal Environment Agency (UBA), Germany, has a long tradition in knowledge organization, using a library along with many Web-based information systems. The backbone of this information space is a classification system enhanced by a reference vocabulary which consists of a thesaurus, a gazetteer and a chronicle. Over the years, classification has increasingly been relegated to the background compared with the reference vocabulary indexing and full text search. Bibliographic items are no longer classified directly but tagged with thesaurus terms, with those terms being classified. Since 2010 we have been developing a linked data representation of this knowledge base. While we are linking bibliographic and observation data with the controlled vocabulary in a Resource Desrcription Framework (RDF) representation, the classification may be revisited as a powerful organization system by inference. This also raises questions about the quality and feasibility of an unambiguous classification of thesaurus terms.
  5. Martínez-González, M.M.; Alvite-Díez, M.L.: Thesauri and Semantic Web : discussion of the evolution of thesauri toward their integration with the Semantic Web (2019) 0.02
    0.020987432 = product of:
      0.10493716 = sum of:
        0.10493716 = weight(_text_:thesaurus in 5997) [ClassicSimilarity], result of:
          0.10493716 = score(doc=5997,freq=6.0), product of:
            0.23732872 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.051357865 = queryNorm
            0.44215953 = fieldWeight in 5997, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5997)
      0.2 = coord(1/5)
    
    Abstract
    Thesauri are Knowledge Organization Systems (KOS), that arise from the consensus of wide communities. They have been in use for many years and are regularly updated. Whereas in the past thesauri were designed for information professionals for indexing and searching, today there is a demand for conceptual vocabularies that enable inferencing by machines. The development of the Semantic Web has brought a new opportunity for thesauri, but thesauri also face the challenge of proving that they add value to it. The evolution of thesauri toward their integration with the Semantic Web is examined. Elements and structures in the thesaurus standard, ISO 25964, and SKOS (Simple Knowledge Organization System), the Semantic Web standard for representing KOS, are reviewed and compared. Moreover, the integrity rules of thesauri are contrasted with the axioms of SKOS. How SKOS has been applied to represent some real thesauri is taken into account. Three thesauri are chosen for this aim: AGROVOC, EuroVoc and the UNESCO Thesaurus. Based on the results of this comparison and analysis, the benefits that Semantic Web technologies offer to thesauri, how thesauri can contribute to the Semantic Web, and the challenges that would help to improve their integration with the Semantic Web are discussed.
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  6. Zapilko, B.; Sure, Y.: Neue Möglichkeiten für die Wissensorganisation durch die Kombination von Digital Library Verfahren mit Standards des Semantic Web (2013) 0.01
    0.014540519 = product of:
      0.072702594 = sum of:
        0.072702594 = weight(_text_:thesaurus in 936) [ClassicSimilarity], result of:
          0.072702594 = score(doc=936,freq=2.0), product of:
            0.23732872 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.051357865 = queryNorm
            0.30633712 = fieldWeight in 936, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.046875 = fieldNorm(doc=936)
      0.2 = coord(1/5)
    
    Abstract
    Entwicklungen und Technologien des Semantic Web treffen seit einigen Jahren verstärkt auf die Bibliotheks- und Dokumentationswelt, um das dort seit Jahrzehnten gesammelte und gepflegte Wissen für das Web und seine Nurzer zugänglich und weiter verarbeitbar zu machen. Dabei können beide Lager von einer Öffnung gegenüber den Verfahren des jeweils anderen und den daraus resultierenden Möglichkeiten, beispielsweise einer integrierten Recherche in verteilten und semantisch angereicherten Dokumentbeständen oder der Anreicherung eigener Bestände um Inhalte anderer, frei verfügbarer Bestände, profitieren. Dieses Paper stellt die Reformulierung eines gängigen informationswissenschaftlichen Verfahrens aus der Dokumentations- und Bibliothekswelt, des sogenannten SchalenmodeIls, vor und zeigt neues Potenzial und neue Möglichkeiten für die Wissensorganisation auf, die durch dessen Anwendung im Semantic Web entstehen können. Darüber hinaus werden erste praktische Ergebnisse der Vorarbeiten dieser Reformulierung präsentiert, die Transformation eines Thesaurus ins SKOS-Format.
  7. Papadakis, I. et al.: Highlighting timely information in libraries through social and semantic Web technologies (2016) 0.01
    0.013916564 = product of:
      0.06958282 = sum of:
        0.06958282 = weight(_text_:22 in 2090) [ClassicSimilarity], result of:
          0.06958282 = score(doc=2090,freq=2.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.38690117 = fieldWeight in 2090, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.078125 = fieldNorm(doc=2090)
      0.2 = coord(1/5)
    
    Source
    Metadata and semantics research: 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings. Eds.: E. Garoufallou
  8. Eckert, K.: SKOS: eine Sprache für die Übertragung von Thesauri ins Semantic Web (2011) 0.01
    0.011133251 = product of:
      0.055666253 = sum of:
        0.055666253 = weight(_text_:22 in 4331) [ClassicSimilarity], result of:
          0.055666253 = score(doc=4331,freq=2.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.30952093 = fieldWeight in 4331, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0625 = fieldNorm(doc=4331)
      0.2 = coord(1/5)
    
    Date
    15. 3.2011 19:21:22
  9. Metadata and semantics research : 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings (2016) 0.01
    0.009741595 = product of:
      0.048707973 = sum of:
        0.048707973 = weight(_text_:22 in 3283) [ClassicSimilarity], result of:
          0.048707973 = score(doc=3283,freq=2.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.2708308 = fieldWeight in 3283, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3283)
      0.2 = coord(1/5)
    
  10. Hollink, L.; Assem, M. van: Estimating the relevance of search results in the Culture-Web : a study of semantic distance measures (2010) 0.01
    0.008349938 = product of:
      0.04174969 = sum of:
        0.04174969 = weight(_text_:22 in 4649) [ClassicSimilarity], result of:
          0.04174969 = score(doc=4649,freq=2.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.23214069 = fieldWeight in 4649, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=4649)
      0.2 = coord(1/5)
    
    Date
    26.12.2011 13:40:22
  11. Prud'hommeaux, E.; Gayo, E.: RDF ventures to boldly meet your most pedestrian needs (2015) 0.01
    0.008349938 = product of:
      0.04174969 = sum of:
        0.04174969 = weight(_text_:22 in 2024) [ClassicSimilarity], result of:
          0.04174969 = score(doc=2024,freq=2.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.23214069 = fieldWeight in 2024, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=2024)
      0.2 = coord(1/5)
    
    Source
    Bulletin of the Association for Information Science and Technology. 41(2015) no.4, S.18-22
  12. Firnkes, M.: Schöne neue Welt : der Content der Zukunft wird von Algorithmen bestimmt (2015) 0.01
    0.008349938 = product of:
      0.04174969 = sum of:
        0.04174969 = weight(_text_:22 in 2118) [ClassicSimilarity], result of:
          0.04174969 = score(doc=2118,freq=2.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.23214069 = fieldWeight in 2118, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=2118)
      0.2 = coord(1/5)
    
    Date
    5. 7.2015 22:02:31
  13. Monireh, E.; Sarker, M.K.; Bianchi, F.; Hitzler, P.; Doran, D.; Xie, N.: Reasoning over RDF knowledge bases using deep learning (2018) 0.01
    0.006958282 = product of:
      0.03479141 = sum of:
        0.03479141 = weight(_text_:22 in 4553) [ClassicSimilarity], result of:
          0.03479141 = score(doc=4553,freq=2.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.19345059 = fieldWeight in 4553, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4553)
      0.2 = coord(1/5)
    
    Date
    16.11.2018 14:22:01
  14. Metadata and semantics research : 7th Research Conference, MTSR 2013 Thessaloniki, Greece, November 19-22, 2013. Proceedings (2013) 0.01
    0.006888348 = product of:
      0.03444174 = sum of:
        0.03444174 = weight(_text_:22 in 1155) [ClassicSimilarity], result of:
          0.03444174 = score(doc=1155,freq=4.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.19150631 = fieldWeight in 1155, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1155)
      0.2 = coord(1/5)
    
    Date
    17.12.2013 12:51:22
  15. Borst, T.; Fingerle, B.; Neubert, J.; Seiler, A.: Auf dem Weg in das Semantic Web : Anwendungsbeispiele und Lösungsszenarien in Bibliotheken / Eine Veranstaltung von hbz und ZBW (2010) 0.01
    0.006058549 = product of:
      0.030292746 = sum of:
        0.030292746 = weight(_text_:thesaurus in 4398) [ClassicSimilarity], result of:
          0.030292746 = score(doc=4398,freq=2.0), product of:
            0.23732872 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.051357865 = queryNorm
            0.12764046 = fieldWeight in 4398, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.01953125 = fieldNorm(doc=4398)
      0.2 = coord(1/5)
    
    Content
    Moderne Web-Anwendungen Wie ein klassischer Bibliothekskatalog in eine moderne Web-Anwendung und in Linked Open Data überführt werden kann, berichtete Anders Söderbäck von der Schwedischen Nationalbibliothek. Unter dem Thema »Linked Applications« entwarf Felix Ostrowski, IT-Entwickler beim hbz, eine tiefe Integration von Ontologien direkt in die verwendete Programmiersprache am Beispiel von Repository Software. Andre Hagenbruch, Mitarbeiter der Universitätsbibliothek Bochum, stellte ein Projekt für ein Linked Data-basiertes Bibliotheksportal vor, das neben bibliografischen Daten auch administrative Informationen wie Personaldaten oder Öffnungszeiten integrieren soll. Interessante Einblicke bot Jürgen Kett von der DNB zur hauseigenen Linked Data-Strategie. Dazu gehört die Veröffentlichung der Personennamendatei (PND) zwecks kostenfreier Weiterverwendung durch andere Bibliotheken. In einem weiteren Vortrag gingen Kett und sein Kollege Svensson auf Details ihres aktuellen Linked Data-Projekts ein, das in einer ersten Stufe URI- und SPARQL-Zugriff auf die SWDund PND-Datensets sowie ihre Verknüpfungen zu internationalen Normdatenbeständen ermöglichen soll. Ein Beta-Dienst soll bis Mitte 2010 online gehen. Timo Borst, Leiter der IT-Entwicklung der ZBW, griff dies auf und demonstrierte die Integration von Thesaurus-und Personennormdaten in den ZBW-Dokumentenserver. Als technischer Zugangspunkt für Linked Data wurde dabei auch ein experimenteller SPARQLEndpoint der DNB genutzt, dessen Abfrage eine automatische Vorschlagsfunktion für PND-Einträge ermöglicht. Praktische Erfahrungen aus den Linked Data-Projekten der ZBW fasste Joachim Neubert, IT-Entwickler bei der ZBW, zusammen, verbunden mit Hinweisen zu Tools, Communities und Best Practice-Beispielen im Web.
  16. Zhitomirsky-Geffet, M.; Bar-Ilan, J.: Towards maximal unification of semantically diverse ontologies for controversial domains (2014) 0.01
    0.0055666254 = product of:
      0.027833126 = sum of:
        0.027833126 = weight(_text_:22 in 1634) [ClassicSimilarity], result of:
          0.027833126 = score(doc=1634,freq=2.0), product of:
            0.1798465 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.051357865 = queryNorm
            0.15476047 = fieldWeight in 1634, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.03125 = fieldNorm(doc=1634)
      0.2 = coord(1/5)
    
    Date
    20. 1.2015 18:30:22