Search (313 results, page 1 of 16)

  • × theme_ss:"Semantic Web"
  1. Papadakis, I. et al.: Highlighting timely information in libraries through social and semantic Web technologies (2016) 0.03
    0.03236657 = product of:
      0.06473314 = sum of:
        0.06473314 = sum of:
          0.0063317944 = weight(_text_:a in 2090) [ClassicSimilarity], result of:
            0.0063317944 = score(doc=2090,freq=2.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.12739488 = fieldWeight in 2090, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.078125 = fieldNorm(doc=2090)
          0.058401346 = weight(_text_:22 in 2090) [ClassicSimilarity], result of:
            0.058401346 = score(doc=2090,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.38690117 = fieldWeight in 2090, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.078125 = fieldNorm(doc=2090)
      0.5 = coord(1/2)
    
    Source
    Metadata and semantics research: 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings. Eds.: E. Garoufallou
    Type
    a
  2. Baker, T.; Sutton, S.A.: Linked data and the charm of weak semantics : Introduction: the strengths of weak semantics (2015) 0.03
    0.029890416 = sum of:
      0.02514157 = product of:
        0.15084942 = sum of:
          0.15084942 = weight(_text_:baker in 2022) [ClassicSimilarity], result of:
            0.15084942 = score(doc=2022,freq=2.0), product of:
              0.34308222 = queryWeight, product of:
                7.9592175 = idf(docFreq=41, maxDocs=44218)
                0.043105017 = queryNorm
              0.4396888 = fieldWeight in 2022, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                7.9592175 = idf(docFreq=41, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2022)
        0.16666667 = coord(1/6)
      0.004748846 = product of:
        0.009497692 = sum of:
          0.009497692 = weight(_text_:a in 2022) [ClassicSimilarity], result of:
            0.009497692 = score(doc=2022,freq=18.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.19109234 = fieldWeight in 2022, product of:
                4.2426405 = tf(freq=18.0), with freq of:
                  18.0 = termFreq=18.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2022)
        0.5 = coord(1/2)
    
    Abstract
    Logic and precision are fundamental to ontologies underlying the semantic web and, by extension, to linked data. This special section focuses on the interaction of semantics, ontologies and linked data. The discussion presents the Simple Knowledge Organization Scheme (SKOS) as a less formal strategy for expressing concept hierarchies and associations and questions the value of deep domain ontologies in favor of simpler vocabularies that are more open to reuse, albeit risking illogical outcomes. RDF ontologies harbor another unexpected drawback. While structurally sound, they leave validation gaps permitting illogical uses, a problem being addressed by a W3C Working Group. Data models based on RDF graphs and properties may replace traditional library catalog models geared to predefined entities, with relationships between RDF classes providing the semantic connections. The BIBFRAME Initiative takes a different and streamlined approach to linking data, building rich networks of information resources rather than relying on a strict underlying structure and vocabulary. Taken together, the articles illustrate the trend toward a pragmatic approach to a Semantic Web, sacrificing some specificity for greater flexibility and partial interoperability.
    Footnote
    Introduction to a special section "Linked data and the charm of weak semantics".
    Type
    a
  3. Isaac, A.; Baker, T.: Linked data practice at different levels of semantic precision : the perspective of libraries, archives and museums (2015) 0.03
    0.029018987 = sum of:
      0.02514157 = product of:
        0.15084942 = sum of:
          0.15084942 = weight(_text_:baker in 2026) [ClassicSimilarity], result of:
            0.15084942 = score(doc=2026,freq=2.0), product of:
              0.34308222 = queryWeight, product of:
                7.9592175 = idf(docFreq=41, maxDocs=44218)
                0.043105017 = queryNorm
              0.4396888 = fieldWeight in 2026, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                7.9592175 = idf(docFreq=41, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2026)
        0.16666667 = coord(1/6)
      0.0038774167 = product of:
        0.0077548334 = sum of:
          0.0077548334 = weight(_text_:a in 2026) [ClassicSimilarity], result of:
            0.0077548334 = score(doc=2026,freq=12.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.15602624 = fieldWeight in 2026, product of:
                3.4641016 = tf(freq=12.0), with freq of:
                  12.0 = termFreq=12.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2026)
        0.5 = coord(1/2)
    
    Abstract
    Libraries, archives and museums rely on structured schemas and vocabularies to indicate classes in which a resource may belong. In the context of linked data, key organizational components are the RDF data model, element schemas and value vocabularies, with simple ontologies having minimally defined classes and properties in order to facilitate reuse and interoperability. Simplicity over formal semantics is a tenet of the open-world assumption underlying ontology languages central to the Semantic Web, but the result is a lack of constraints, data quality checks and validation capacity. Inconsistent use of vocabularies and ontologies that do not follow formal semantics rules and logical concept hierarchies further complicate the use of Semantic Web technologies. The Simple Knowledge Organization System (SKOS) helps make existing value vocabularies available in the linked data environment, but it exchanges precision for simplicity. Incompatibilities between simple organized vocabularies, Resource Description Framework Schemas and OWL ontologies and even basic notions of subjects and concepts prevent smooth translations and challenge the conversion of cultural institutions' unique legacy vocabularies for linked data. Adopting the linked data vision requires accepting loose semantic interpretations. To avoid semantic inconsistencies and illogical results, cultural organizations following the linked data path must be careful to choose the level of semantics that best suits their domain and needs.
    Footnote
    Contribution to a special section "Linked data and the charm of weak semantics".
    Type
    a
  4. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.03
    0.028193437 = sum of:
      0.022820732 = product of:
        0.13692439 = sum of:
          0.13692439 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
            0.13692439 = score(doc=701,freq=2.0), product of:
              0.3654448 = queryWeight, product of:
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.043105017 = queryNorm
              0.3746787 = fieldWeight in 701, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.03125 = fieldNorm(doc=701)
        0.16666667 = coord(1/6)
      0.005372706 = product of:
        0.010745412 = sum of:
          0.010745412 = weight(_text_:a in 701) [ClassicSimilarity], result of:
            0.010745412 = score(doc=701,freq=36.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.2161963 = fieldWeight in 701, product of:
                6.0 = tf(freq=36.0), with freq of:
                  36.0 = termFreq=36.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.03125 = fieldNorm(doc=701)
        0.5 = coord(1/2)
    
    Abstract
    By the explosion of possibilities for a ubiquitous content production, the information overload problem reaches the level of complexity which cannot be managed by traditional modelling approaches anymore. Due to their pure syntactical nature traditional information retrieval approaches did not succeed in treating content itself (i.e. its meaning, and not its representation). This leads to a very low usefulness of the results of a retrieval process for a user's task at hand. In the last ten years ontologies have been emerged from an interesting conceptualisation paradigm to a very promising (semantic) modelling technology, especially in the context of the Semantic Web. From the information retrieval point of view, ontologies enable a machine-understandable form of content description, such that the retrieval process can be driven by the meaning of the content. However, the very ambiguous nature of the retrieval process in which a user, due to the unfamiliarity with the underlying repository and/or query syntax, just approximates his information need in a query, implies a necessity to include the user in the retrieval process more actively in order to close the gap between the meaning of the content and the meaning of a user's query (i.e. his information need). This thesis lays foundation for such an ontology-based interactive retrieval process, in which the retrieval system interacts with a user in order to conceptually interpret the meaning of his query, whereas the underlying domain ontology drives the conceptualisation process. In that way the retrieval process evolves from a query evaluation process into a highly interactive cooperation between a user and the retrieval system, in which the system tries to anticipate the user's information need and to deliver the relevant content proactively. Moreover, the notion of content relevance for a user's query evolves from a content dependent artefact to the multidimensional context-dependent structure, strongly influenced by the user's preferences. This cooperation process is realized as the so-called Librarian Agent Query Refinement Process. In order to clarify the impact of an ontology on the retrieval process (regarding its complexity and quality), a set of methods and tools for different levels of content and query formalisation is developed, ranging from pure ontology-based inferencing to keyword-based querying in which semantics automatically emerges from the results. Our evaluation studies have shown that the possibilities to conceptualize a user's information need in the right manner and to interpret the retrieval results accordingly are key issues for realizing much more meaningful information retrieval systems.
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  5. Synak, M.; Dabrowski, M.; Kruk, S.R.: Semantic Web and ontologies (2009) 0.03
    0.025893256 = product of:
      0.051786512 = sum of:
        0.051786512 = sum of:
          0.0050654355 = weight(_text_:a in 3376) [ClassicSimilarity], result of:
            0.0050654355 = score(doc=3376,freq=2.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.10191591 = fieldWeight in 3376, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0625 = fieldNorm(doc=3376)
          0.04672108 = weight(_text_:22 in 3376) [ClassicSimilarity], result of:
            0.04672108 = score(doc=3376,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.30952093 = fieldWeight in 3376, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=3376)
      0.5 = coord(1/2)
    
    Date
    31. 7.2010 16:58:22
    Type
    a
  6. Faaborg, A.; Lagoze, C.: Semantic browsing (2003) 0.03
    0.025868854 = product of:
      0.051737707 = sum of:
        0.051737707 = sum of:
          0.010856766 = weight(_text_:a in 1026) [ClassicSimilarity], result of:
            0.010856766 = score(doc=1026,freq=12.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.21843673 = fieldWeight in 1026, product of:
                3.4641016 = tf(freq=12.0), with freq of:
                  12.0 = termFreq=12.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0546875 = fieldNorm(doc=1026)
          0.04088094 = weight(_text_:22 in 1026) [ClassicSimilarity], result of:
            0.04088094 = score(doc=1026,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.2708308 = fieldWeight in 1026, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=1026)
      0.5 = coord(1/2)
    
    Abstract
    We have created software applications that allow users to both author and use Semantic Web metadata. To create and use a layer of semantic content on top of the existing Web, we have (1) implemented a user interface that expedites the task of attributing metadata to resources on the Web, and (2) augmented a Web browser to leverage this semantic metadata to provide relevant information and tasks to the user. This project provides a framework for annotating and reorganizing existing files, pages, and sites on the Web that is similar to Vannevar Bushrsquos original concepts of trail blazing and associative indexing.
    Source
    Research and advanced technology for digital libraries : 7th European Conference, proceedings / ECDL 2003, Trondheim, Norway, August 17-22, 2003
    Type
    a
  7. Heflin, J.; Hendler, J.: Semantic interoperability on the Web (2000) 0.02
    0.024872728 = product of:
      0.049745455 = sum of:
        0.049745455 = sum of:
          0.008864513 = weight(_text_:a in 759) [ClassicSimilarity], result of:
            0.008864513 = score(doc=759,freq=8.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.17835285 = fieldWeight in 759, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0546875 = fieldNorm(doc=759)
          0.04088094 = weight(_text_:22 in 759) [ClassicSimilarity], result of:
            0.04088094 = score(doc=759,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.2708308 = fieldWeight in 759, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=759)
      0.5 = coord(1/2)
    
    Abstract
    XML will have a profound impact on the way data is exchanged on the Internet. An important feature of this language is the separation of content from presentation, which makes it easier to select and/or reformat the data. However, due to the likelihood of numerous industry and domain specific DTDs, those who wish to integrate information will still be faced with the problem of semantic interoperability. In this paper we discuss why this problem is not solved by XML, and then discuss why the Resource Description Framework is only a partial solution. We then present the SHOE language, which we feel has many of the features necessary to enable a semantic web, and describe an existing set of tools that make it easy to use the language.
    Date
    11. 5.2013 19:22:18
    Type
    a
  8. Malmsten, M.: Making a library catalogue part of the Semantic Web (2008) 0.02
    0.024278916 = product of:
      0.048557833 = sum of:
        0.048557833 = sum of:
          0.0076768934 = weight(_text_:a in 2640) [ClassicSimilarity], result of:
            0.0076768934 = score(doc=2640,freq=6.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.1544581 = fieldWeight in 2640, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2640)
          0.04088094 = weight(_text_:22 in 2640) [ClassicSimilarity], result of:
            0.04088094 = score(doc=2640,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.2708308 = fieldWeight in 2640, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2640)
      0.5 = coord(1/2)
    
    Abstract
    Library catalogues contain an enormous amount of structured, high-quality data, however, this data is generally not made available to semantic web applications. In this paper we describe the tools and techniques used to make the Swedish Union Catalogue (LIBRIS) part of the Semantic Web and Linked Data. The focus is on links to and between resources and the mechanisms used to make data available, rather than perfect description of the individual resources. We also present a method of creating links between records of the same work.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
    Type
    a
  9. Blumauer, A.; Pellegrini, T.: Semantic Web Revisited : Eine kurze Einführung in das Social Semantic Web (2009) 0.02
    0.024278916 = product of:
      0.048557833 = sum of:
        0.048557833 = sum of:
          0.0076768934 = weight(_text_:a in 4855) [ClassicSimilarity], result of:
            0.0076768934 = score(doc=4855,freq=6.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.1544581 = fieldWeight in 4855, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0546875 = fieldNorm(doc=4855)
          0.04088094 = weight(_text_:22 in 4855) [ClassicSimilarity], result of:
            0.04088094 = score(doc=4855,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.2708308 = fieldWeight in 4855, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=4855)
      0.5 = coord(1/2)
    
    Pages
    S.3-22
    Source
    Social Semantic Web: Web 2.0, was nun? Hrsg.: A. Blumauer u. T. Pellegrini
    Type
    a
  10. Schneider, R.: Web 3.0 ante portas? : Integration von Social Web und Semantic Web (2008) 0.02
    0.02357455 = product of:
      0.0471491 = sum of:
        0.0471491 = sum of:
          0.006268157 = weight(_text_:a in 4184) [ClassicSimilarity], result of:
            0.006268157 = score(doc=4184,freq=4.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.12611452 = fieldWeight in 4184, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0546875 = fieldNorm(doc=4184)
          0.04088094 = weight(_text_:22 in 4184) [ClassicSimilarity], result of:
            0.04088094 = score(doc=4184,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.2708308 = fieldWeight in 4184, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=4184)
      0.5 = coord(1/2)
    
    Date
    22. 1.2011 10:38:28
    Source
    Kommunikation, Partizipation und Wirkungen im Social Web, Band 1. Hrsg.: A. Zerfaß u.a
    Type
    a
  11. Baker, T.; Bermès, E.; Coyle, K.; Dunsire, G.; Isaac, A.; Murray, P.; Panzer, M.; Schneider, J.; Singer, R.; Summers, E.; Waites, W.; Young, J.; Zeng, M.: Library Linked Data Incubator Group Final Report (2011) 0.02
    0.022645973 = sum of:
      0.020113256 = product of:
        0.12067953 = sum of:
          0.12067953 = weight(_text_:baker in 4796) [ClassicSimilarity], result of:
            0.12067953 = score(doc=4796,freq=2.0), product of:
              0.34308222 = queryWeight, product of:
                7.9592175 = idf(docFreq=41, maxDocs=44218)
                0.043105017 = queryNorm
              0.35175103 = fieldWeight in 4796, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                7.9592175 = idf(docFreq=41, maxDocs=44218)
                0.03125 = fieldNorm(doc=4796)
        0.16666667 = coord(1/6)
      0.0025327178 = product of:
        0.0050654355 = sum of:
          0.0050654355 = weight(_text_:a in 4796) [ClassicSimilarity], result of:
            0.0050654355 = score(doc=4796,freq=8.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.10191591 = fieldWeight in 4796, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.03125 = fieldNorm(doc=4796)
        0.5 = coord(1/2)
    
    Abstract
    The mission of the W3C Library Linked Data Incubator Group, chartered from May 2010 through August 2011, has been "to help increase global interoperability of library data on the Web, by bringing together people involved in Semantic Web activities - focusing on Linked Data - in the library community and beyond, building on existing initiatives, and identifying collaboration tracks for the future." In Linked Data [LINKEDDATA], data is expressed using standards such as Resource Description Framework (RDF) [RDF], which specifies relationships between things, and Uniform Resource Identifiers (URIs, or "Web addresses") [URI]. This final report of the Incubator Group examines how Semantic Web standards and Linked Data principles can be used to make the valuable information assets that library create and curate - resources such as bibliographic data, authorities, and concept schemes - more visible and re-usable outside of their original library context on the wider Web. The Incubator Group began by eliciting reports on relevant activities from parties ranging from small, independent projects to national library initiatives (see the separate report, Library Linked Data Incubator Group: Use Cases) [USECASE]. These use cases provided the starting point for the work summarized in the report: an analysis of the benefits of library Linked Data, a discussion of current issues with regard to traditional library data, existing library Linked Data initiatives, and legal rights over library data; and recommendations for next steps. The report also summarizes the results of a survey of current Linked Data technologies and an inventory of library Linked Data resources available today (see also the more detailed report, Library Linked Data Incubator Group: Datasets, Value Vocabularies, and Metadata Element Sets) [VOCABDATASET].
    Key recommendations of the report are: - That library leaders identify sets of data as possible candidates for early exposure as Linked Data and foster a discussion about Open Data and rights; - That library standards bodies increase library participation in Semantic Web standardization, develop library data standards that are compatible with Linked Data, and disseminate best-practice design patterns tailored to library Linked Data; - That data and systems designers design enhanced user services based on Linked Data capabilities, create URIs for the items in library datasets, develop policies for managing RDF vocabularies and their URIs, and express library data by re-using or mapping to existing Linked Data vocabularies; - That librarians and archivists preserve Linked Data element sets and value vocabularies and apply library experience in curation and long-term preservation to Linked Data datasets.
  12. Franklin, R.A.: Re-inventing subject access for the semantic web (2003) 0.02
    0.022173304 = product of:
      0.04434661 = sum of:
        0.04434661 = sum of:
          0.009305799 = weight(_text_:a in 2556) [ClassicSimilarity], result of:
            0.009305799 = score(doc=2556,freq=12.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.18723148 = fieldWeight in 2556, product of:
                3.4641016 = tf(freq=12.0), with freq of:
                  12.0 = termFreq=12.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=2556)
          0.035040807 = weight(_text_:22 in 2556) [ClassicSimilarity], result of:
            0.035040807 = score(doc=2556,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.23214069 = fieldWeight in 2556, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2556)
      0.5 = coord(1/2)
    
    Abstract
    First generation scholarly research on the Web lacked a firm system of authority control. Second generation Web research is beginning to model subject access with library science principles of bibliographic control and cataloguing. Harnessing the Web and organising the intellectual content with standards and controlled vocabulary provides precise search and retrieval capability, increasing relevance and efficient use of technology. Dublin Core metadata standards permit a full evaluation and cataloguing of Web resources appropriate to highly specific research needs and discovery. Current research points to a type of structure based on a system of faceted classification. This system allows the semantic and syntactic relationships to be defined. Controlled vocabulary, such as the Library of Congress Subject Headings, can be assigned, not in a hierarchical structure, but rather as descriptive facets of relating concepts. Web design features such as this are adding value to discovery and filtering out data that lack authority. The system design allows for scalability and extensibility, two technical features that are integral to future development of the digital library and resource discovery.
    Date
    30.12.2008 18:22:46
    Type
    a
  13. Jacobs, I.: From chaos, order: W3C standard helps organize knowledge : SKOS Connects Diverse Knowledge Organization Systems to Linked Data (2009) 0.02
    0.021890612 = sum of:
      0.017599098 = product of:
        0.10559458 = sum of:
          0.10559458 = weight(_text_:baker in 3062) [ClassicSimilarity], result of:
            0.10559458 = score(doc=3062,freq=2.0), product of:
              0.34308222 = queryWeight, product of:
                7.9592175 = idf(docFreq=41, maxDocs=44218)
                0.043105017 = queryNorm
              0.30778214 = fieldWeight in 3062, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                7.9592175 = idf(docFreq=41, maxDocs=44218)
                0.02734375 = fieldNorm(doc=3062)
        0.16666667 = coord(1/6)
      0.004291514 = product of:
        0.008583028 = sum of:
          0.008583028 = weight(_text_:a in 3062) [ClassicSimilarity], result of:
            0.008583028 = score(doc=3062,freq=30.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.17268941 = fieldWeight in 3062, product of:
                5.477226 = tf(freq=30.0), with freq of:
                  30.0 = termFreq=30.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.02734375 = fieldNorm(doc=3062)
        0.5 = coord(1/2)
    
    Abstract
    18 August 2009 -- Today W3C announces a new standard that builds a bridge between the world of knowledge organization systems - including thesauri, classifications, subject headings, taxonomies, and folksonomies - and the linked data community, bringing benefits to both. Libraries, museums, newspapers, government portals, enterprises, social networking applications, and other communities that manage large collections of books, historical artifacts, news reports, business glossaries, blog entries, and other items can now use Simple Knowledge Organization System (SKOS) to leverage the power of linked data. As different communities with expertise and established vocabularies use SKOS to integrate them into the Semantic Web, they increase the value of the information for everyone.
    Content
    SKOS Adapts to the Diversity of Knowledge Organization Systems A useful starting point for understanding the role of SKOS is the set of subject headings published by the US Library of Congress (LOC) for categorizing books, videos, and other library resources. These headings can be used to broaden or narrow queries for discovering resources. For instance, one can narrow a query about books on "Chinese literature" to "Chinese drama," or further still to "Chinese children's plays." Library of Congress subject headings have evolved within a community of practice over a period of decades. By now publishing these subject headings in SKOS, the Library of Congress has made them available to the linked data community, which benefits from a time-tested set of concepts to re-use in their own data. This re-use adds value ("the network effect") to the collection. When people all over the Web re-use the same LOC concept for "Chinese drama," or a concept from some other vocabulary linked to it, this creates many new routes to the discovery of information, and increases the chances that relevant items will be found. As an example of mapping one vocabulary to another, a combined effort from the STITCH, TELplus and MACS Projects provides links between LOC concepts and RAMEAU, a collection of French subject headings used by the Bibliothèque Nationale de France and other institutions. SKOS can be used for subject headings but also many other approaches to organizing knowledge. Because different communities are comfortable with different organization schemes, SKOS is designed to port diverse knowledge organization systems to the Web. "Active participation from the library and information science community in the development of SKOS over the past seven years has been key to ensuring that SKOS meets a variety of needs," said Thomas Baker, co-chair of the Semantic Web Deployment Working Group, which published SKOS. "One goal in creating SKOS was to provide new uses for well-established knowledge organization systems by providing a bridge to the linked data cloud." SKOS is part of the Semantic Web technology stack. Like the Web Ontology Language (OWL), SKOS can be used to define vocabularies. But the two technologies were designed to meet different needs. SKOS is a simple language with just a few features, tuned for sharing and linking knowledge organization systems such as thesauri and classification schemes. OWL offers a general and powerful framework for knowledge representation, where additional "rigor" can afford additional benefits (for instance, business rule processing). To get started with SKOS, see the SKOS Primer.
  14. Hooland, S. van; Verborgh, R.; Wilde, M. De; Hercher, J.; Mannens, E.; Wa, R.Van de: Evaluating the success of vocabulary reconciliation for cultural heritage collections (2013) 0.02
    0.021767901 = product of:
      0.043535803 = sum of:
        0.043535803 = sum of:
          0.008494995 = weight(_text_:a in 662) [ClassicSimilarity], result of:
            0.008494995 = score(doc=662,freq=10.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.1709182 = fieldWeight in 662, product of:
                3.1622777 = tf(freq=10.0), with freq of:
                  10.0 = termFreq=10.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=662)
          0.035040807 = weight(_text_:22 in 662) [ClassicSimilarity], result of:
            0.035040807 = score(doc=662,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.23214069 = fieldWeight in 662, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=662)
      0.5 = coord(1/2)
    
    Abstract
    The concept of Linked Data has made its entrance in the cultural heritage sector due to its potential use for the integration of heterogeneous collections and deriving additional value out of existing metadata. However, practitioners and researchers alike need a better understanding of what outcome they can reasonably expect of the reconciliation process between their local metadata and established controlled vocabularies which are already a part of the Linked Data cloud. This paper offers an in-depth analysis of how a locally developed vocabulary can be successfully reconciled with the Library of Congress Subject Headings (LCSH) and the Arts and Architecture Thesaurus (AAT) through the help of a general-purpose tool for interactive data transformation (OpenRefine). Issues negatively affecting the reconciliation process are identified and solutions are proposed in order to derive maximum value from existing metadata and controlled vocabularies in an automated manner.
    Date
    22. 3.2013 19:29:20
    Type
    a
  15. Gendt, M. van; Isaac, I.; Meij, L. van der; Schlobach, S.: Semantic Web techniques for multiple views on heterogeneous collections : a case study (2006) 0.02
    0.02131948 = product of:
      0.04263896 = sum of:
        0.04263896 = sum of:
          0.007598154 = weight(_text_:a in 2418) [ClassicSimilarity], result of:
            0.007598154 = score(doc=2418,freq=8.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.15287387 = fieldWeight in 2418, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=2418)
          0.035040807 = weight(_text_:22 in 2418) [ClassicSimilarity], result of:
            0.035040807 = score(doc=2418,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.23214069 = fieldWeight in 2418, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2418)
      0.5 = coord(1/2)
    
    Abstract
    Integrated digital access to multiple collections is a prominent issue for many Cultural Heritage institutions. The metadata describing diverse collections must be interoperable, which requires aligning the controlled vocabularies that are used to annotate objects from these collections. In this paper, we present an experiment where we match the vocabularies of two collections by applying the Knowledge Representation techniques established in recent Semantic Web research. We discuss the steps that are required for such matching, namely formalising the initial resources using Semantic Web languages, and running ontology mapping tools on the resulting representations. In addition, we present a prototype that enables the user to browse the two collections using the obtained alignment while still providing her with the original vocabulary structures.
    Source
    Research and advanced technology for digital libraries : 10th European conference, proceedings / ECDL 2006, Alicante, Spain, September 17 - 22, 2006
    Type
    a
  16. Prud'hommeaux, E.; Gayo, E.: RDF ventures to boldly meet your most pedestrian needs (2015) 0.02
    0.0208105 = product of:
      0.041621 = sum of:
        0.041621 = sum of:
          0.0065801945 = weight(_text_:a in 2024) [ClassicSimilarity], result of:
            0.0065801945 = score(doc=2024,freq=6.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.13239266 = fieldWeight in 2024, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=2024)
          0.035040807 = weight(_text_:22 in 2024) [ClassicSimilarity], result of:
            0.035040807 = score(doc=2024,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.23214069 = fieldWeight in 2024, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2024)
      0.5 = coord(1/2)
    
    Abstract
    Defined in 1999 and paired with XML, the Resource Description Framework (RDF) has been cast as an RDF Schema, producing data that is well-structured but not validated, permitting certain illogical relationships. When stakeholders convened in 2014 to consider solutions to the data validation challenge, a W3C working group proposed Resource Shapes and Shape Expressions to describe the properties expected for an RDF node. Resistance rose from concerns about data and schema reuse, key principles in RDF. Ideally data types and properties are designed for broad use, but they are increasingly adopted with local restrictions for specific purposes. Resource Shapes are commonly treated as record classes, standing in for data structures but losing flexibility for later reuse. Of various solutions to the resulting tensions, the concept of record classes may be the most reasonable basis for agreement, satisfying stakeholders' objectives while allowing for variations with constraints.
    Footnote
    Contribution to a special section "Linked data and the charm of weak semantics".
    Source
    Bulletin of the Association for Information Science and Technology. 41(2015) no.4, S.18-22
    Type
    a
  17. Dextre Clarke, S.G.: Challenges and opportunities for KOS standards (2007) 0.02
    0.02044047 = product of:
      0.04088094 = sum of:
        0.04088094 = product of:
          0.08176188 = sum of:
            0.08176188 = weight(_text_:22 in 4643) [ClassicSimilarity], result of:
              0.08176188 = score(doc=4643,freq=2.0), product of:
                0.15094642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043105017 = queryNorm
                0.5416616 = fieldWeight in 4643, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4643)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 9.2007 15:41:14
  18. Zeng, M.L.; Fan, W.; Lin, X.: SKOS for an integrated vocabulary structure (2008) 0.02
    0.020317473 = product of:
      0.040634945 = sum of:
        0.040634945 = sum of:
          0.007598154 = weight(_text_:a in 2654) [ClassicSimilarity], result of:
            0.007598154 = score(doc=2654,freq=18.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.15287387 = fieldWeight in 2654, product of:
                4.2426405 = tf(freq=18.0), with freq of:
                  18.0 = termFreq=18.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.03125 = fieldNorm(doc=2654)
          0.03303679 = weight(_text_:22 in 2654) [ClassicSimilarity], result of:
            0.03303679 = score(doc=2654,freq=4.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.21886435 = fieldWeight in 2654, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=2654)
      0.5 = coord(1/2)
    
    Abstract
    In order to transfer the Chinese Classified Thesaurus (CCT) into a machine-processable format and provide CCT-based Web services, a pilot study has been conducted in which a variety of selected CCT classes and mapped thesaurus entries are encoded with SKOS. OWL and RDFS are also used to encode the same contents for the purposes of feasibility and cost-benefit comparison. CCT is a collected effort led by the National Library of China. It is an integration of the national standards Chinese Library Classification (CLC) 4th edition and Chinese Thesaurus (CT). As a manually created mapping product, CCT provides for each of the classes the corresponding thesaurus terms, and vice versa. The coverage of CCT includes four major clusters: philosophy, social sciences and humanities, natural sciences and technologies, and general works. There are 22 main-classes, 52,992 sub-classes and divisions, 110,837 preferred thesaurus terms, 35,690 entry terms (non-preferred terms), and 59,738 pre-coordinated headings (Chinese Classified Thesaurus, 2005) Major challenges of encoding this large vocabulary comes from its integrated structure. CCT is a result of the combination of two structures (illustrated in Figure 1): a thesaurus that uses ISO-2788 standardized structure and a classification scheme that is basically enumerative, but provides some flexibility for several kinds of synthetic mechanisms Other challenges include the complex relationships caused by differences of granularities of two original schemes and their presentation with various levels of SKOS elements; as well as the diverse coordination of entries due to the use of auxiliary tables and pre-coordinated headings derived from combining classes, subdivisions, and thesaurus terms, which do not correspond to existing unique identifiers. The poster reports the progress, shares the sample SKOS entries, and summarizes problems identified during the SKOS encoding process. Although OWL Lite and OWL Full provide richer expressiveness, the cost-benefit issues and the final purposes of encoding CCT raise questions of using such approaches.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
    Type
    a
  19. Hollink, L.; Assem, M. van: Estimating the relevance of search results in the Culture-Web : a study of semantic distance measures (2010) 0.02
    0.020206757 = product of:
      0.040413514 = sum of:
        0.040413514 = sum of:
          0.005372706 = weight(_text_:a in 4649) [ClassicSimilarity], result of:
            0.005372706 = score(doc=4649,freq=4.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.10809815 = fieldWeight in 4649, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=4649)
          0.035040807 = weight(_text_:22 in 4649) [ClassicSimilarity], result of:
            0.035040807 = score(doc=4649,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.23214069 = fieldWeight in 4649, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=4649)
      0.5 = coord(1/2)
    
    Abstract
    More and more cultural heritage institutions publish their collections, vocabularies and metadata on the Web. The resulting Web of linked cultural data opens up exciting new possibilities for searching and browsing through these cultural heritage collections. We report on ongoing work in which we investigate the estimation of relevance in this Web of Culture. We study existing measures of semantic distance and how they apply to two use cases. The use cases relate to the structured, multilingual and multimodal nature of the Culture Web. We distinguish between measures using the Web, such as Google distance and PMI, and measures using the Linked Data Web, i.e. the semantic structure of metadata vocabularies. We perform a small study in which we compare these semantic distance measures to human judgements of relevance. Although it is too early to draw any definitive conclusions, the study provides new insights into the applicability of semantic distance measures to the Web of Culture, and clear starting points for further research.
    Date
    26.12.2011 13:40:22
  20. Keyser, P. de: Indexing : from thesauri to the Semantic Web (2012) 0.02
    0.019419942 = product of:
      0.038839884 = sum of:
        0.038839884 = sum of:
          0.003799077 = weight(_text_:a in 3197) [ClassicSimilarity], result of:
            0.003799077 = score(doc=3197,freq=2.0), product of:
              0.049702108 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.043105017 = queryNorm
              0.07643694 = fieldWeight in 3197, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=3197)
          0.035040807 = weight(_text_:22 in 3197) [ClassicSimilarity], result of:
            0.035040807 = score(doc=3197,freq=2.0), product of:
              0.15094642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043105017 = queryNorm
              0.23214069 = fieldWeight in 3197, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=3197)
      0.5 = coord(1/2)
    
    Abstract
    Indexing consists of both novel and more traditional techniques. Cutting-edge indexing techniques, such as automatic indexing, ontologies, and topic maps, were developed independently of older techniques such as thesauri, but it is now recognized that these older methods also hold expertise. Indexing describes various traditional and novel indexing techniques, giving information professionals and students of library and information sciences a broad and comprehensible introduction to indexing. This title consists of twelve chapters: an Introduction to subject readings and theasauri; Automatic indexing versus manual indexing; Techniques applied in automatic indexing of text material; Automatic indexing of images; The black art of indexing moving images; Automatic indexing of music; Taxonomies and ontologies; Metadata formats and indexing; Tagging; Topic maps; Indexing the web; and The Semantic Web.
    Date
    24. 8.2016 14:03:22

Years

Languages

  • e 241
  • d 69
  • f 1
  • More… Less…

Types

  • a 213
  • el 80
  • m 43
  • s 17
  • n 10
  • x 6
  • r 2
  • More… Less…

Subjects

Classifications