Search (192 results, page 1 of 10)

  • × theme_ss:"Semantic Web"
  1. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.18
    0.17578289 = product of:
      0.3955115 = sum of:
        0.039551154 = product of:
          0.118653454 = sum of:
            0.118653454 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.118653454 = score(doc=701,freq=2.0), product of:
                0.31668055 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.037353165 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
        0.118653454 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.118653454 = score(doc=701,freq=2.0), product of:
            0.31668055 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037353165 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.118653454 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.118653454 = score(doc=701,freq=2.0), product of:
            0.31668055 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037353165 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.118653454 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.118653454 = score(doc=701,freq=2.0), product of:
            0.31668055 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037353165 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
      0.44444445 = coord(4/9)
    
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  2. Metadata and semantics research : 8th Research Conference, MTSR 2014, Karlsruhe, Germany, November 27-29, 2014, Proceedings (2014) 0.07
    0.07228309 = product of:
      0.21684925 = sum of:
        0.103783086 = weight(_text_:germany in 2192) [ClassicSimilarity], result of:
          0.103783086 = score(doc=2192,freq=4.0), product of:
            0.22275731 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.037353165 = queryNorm
            0.46590203 = fieldWeight in 2192, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2192)
        0.065243915 = weight(_text_:data in 2192) [ClassicSimilarity], result of:
          0.065243915 = score(doc=2192,freq=20.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.5523875 = fieldWeight in 2192, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2192)
        0.047822252 = weight(_text_:processing in 2192) [ClassicSimilarity], result of:
          0.047822252 = score(doc=2192,freq=4.0), product of:
            0.15121111 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.037353165 = queryNorm
            0.3162615 = fieldWeight in 2192, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2192)
      0.33333334 = coord(3/9)
    
    Abstract
    This book constitutes the refereed proceedings of the 8th Metadata and Semantics Research Conference, MTSR 2014, held in Karlsruhe, Germany, in November 2014. The 23 full papers and 9 short papers presented were carefully reviewed and selected from 57 submissions. The papers are organized in several sessions and tracks. They cover the following topics: metadata and linked data: tools and models; (meta) data quality assessment and curation; semantic interoperability, ontology-based data access and representation; big data and digital libraries in health, science and technology; metadata and semantics for open repositories, research information systems and data infrastructure; metadata and semantics for cultural collections and applications; semantics for agriculture, food and environment.
    Content
    Metadata and linked data.- Tools and models.- (Meta)data quality assessment and curation.- Semantic interoperability, ontology-based data access and representation.- Big data and digital libraries in health, science and technology.- Metadata and semantics for open repositories, research information systems and data infrastructure.- Metadata and semantics for cultural collections and applications.- Semantics for agriculture, food and environment.
    LCSH
    Text processing (Computer science)
    Subject
    Text processing (Computer science)
  3. Metadata and semantics research : 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings (2016) 0.07
    0.0679528 = product of:
      0.20385839 = sum of:
        0.1452963 = weight(_text_:germany in 3283) [ClassicSimilarity], result of:
          0.1452963 = score(doc=3283,freq=4.0), product of:
            0.22275731 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.037353165 = queryNorm
            0.6522628 = fieldWeight in 3283, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3283)
        0.040849157 = weight(_text_:data in 3283) [ClassicSimilarity], result of:
          0.040849157 = score(doc=3283,freq=4.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.34584928 = fieldWeight in 3283, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3283)
        0.017712934 = product of:
          0.035425868 = sum of:
            0.035425868 = weight(_text_:22 in 3283) [ClassicSimilarity], result of:
              0.035425868 = score(doc=3283,freq=2.0), product of:
                0.13080442 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037353165 = queryNorm
                0.2708308 = fieldWeight in 3283, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3283)
          0.5 = coord(1/2)
      0.33333334 = coord(3/9)
    
    Abstract
    This book constitutes the refereed proceedings of the 10th Metadata and Semantics Research Conference, MTSR 2016, held in Göttingen, Germany, in November 2016. The 26 full papers and 6 short papers presented were carefully reviewed and selected from 67 submissions. The papers are organized in several sessions and tracks: Digital Libraries, Information Retrieval, Linked and Social Data, Metadata and Semantics for Open Repositories, Research Information Systems and Data Infrastructures, Metadata and Semantics for Agriculture, Food and Environment, Metadata and Semantics for Cultural Collections and Applications, European and National Projects.
  4. Malmsten, M.: Making a library catalogue part of the Semantic Web (2008) 0.06
    0.059407458 = product of:
      0.17822237 = sum of:
        0.102740005 = weight(_text_:germany in 2640) [ClassicSimilarity], result of:
          0.102740005 = score(doc=2640,freq=2.0), product of:
            0.22275731 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.037353165 = queryNorm
            0.46121946 = fieldWeight in 2640, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2640)
        0.057769425 = weight(_text_:data in 2640) [ClassicSimilarity], result of:
          0.057769425 = score(doc=2640,freq=8.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.48910472 = fieldWeight in 2640, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2640)
        0.017712934 = product of:
          0.035425868 = sum of:
            0.035425868 = weight(_text_:22 in 2640) [ClassicSimilarity], result of:
              0.035425868 = score(doc=2640,freq=2.0), product of:
                0.13080442 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037353165 = queryNorm
                0.2708308 = fieldWeight in 2640, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2640)
          0.5 = coord(1/2)
      0.33333334 = coord(3/9)
    
    Abstract
    Library catalogues contain an enormous amount of structured, high-quality data, however, this data is generally not made available to semantic web applications. In this paper we describe the tools and techniques used to make the Swedish Union Catalogue (LIBRIS) part of the Semantic Web and Linked Data. The focus is on links to and between resources and the mechanisms used to make data available, rather than perfect description of the individual resources. We also present a method of creating links between records of the same work.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  5. Willer, M.; Dunsire, G.: ISBD, the UNIMARC bibliographic format, and RDA : interoperability issues in namespaces and the linked data environment (2014) 0.05
    0.047414273 = product of:
      0.14224282 = sum of:
        0.04487142 = weight(_text_:cataloging in 1999) [ClassicSimilarity], result of:
          0.04487142 = score(doc=1999,freq=2.0), product of:
            0.14721331 = queryWeight, product of:
              3.9411201 = idf(docFreq=2334, maxDocs=44218)
              0.037353165 = queryNorm
            0.30480546 = fieldWeight in 1999, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9411201 = idf(docFreq=2334, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1999)
        0.050029792 = weight(_text_:data in 1999) [ClassicSimilarity], result of:
          0.050029792 = score(doc=1999,freq=6.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.42357713 = fieldWeight in 1999, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1999)
        0.047341615 = weight(_text_:processing in 1999) [ClassicSimilarity], result of:
          0.047341615 = score(doc=1999,freq=2.0), product of:
            0.15121111 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.037353165 = queryNorm
            0.3130829 = fieldWeight in 1999, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1999)
      0.33333334 = coord(3/9)
    
    Abstract
    The article is an updated and expanded version of a paper presented to International Federation of Library Associations and Institutions in 2013. It describes recent work involving the representation of International Standard for Bibliographic Description (ISBD) and UNIMARC (UNIversal MARC) in Resource Description Framework (RDF), the basis of the Semantic Web and linked data. The UNIMARC Bibliographic format is used to illustrate issues arising from the development of a bibliographic element set and its semantic alignment with ISBD. The article discusses the use of such alignments in the automated processing of linked data for interoperability, using examples from ISBD, UNIMARC, and Resource Description and Access.
    Source
    Cataloging and classification quarterly. 52(2014) no.8, S.888-913
  6. Aberer, K. et al.: ¬The Semantic Web : 6th International Semantic Web Conference, 2nd Asian Semantic Web Conference, ISWC 2007 + ASWC 2007, Busan, Korea, November 11-15, 2007 : proceedings (2007) 0.04
    0.043854866 = product of:
      0.1973469 = sum of:
        0.043669585 = weight(_text_:data in 2477) [ClassicSimilarity], result of:
          0.043669585 = score(doc=2477,freq=14.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.36972845 = fieldWeight in 2477, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=2477)
        0.15367731 = product of:
          0.30735463 = sum of:
            0.30735463 = weight(_text_:congresses in 2477) [ClassicSimilarity], result of:
              0.30735463 = score(doc=2477,freq=16.0), product of:
                0.30305997 = queryWeight, product of:
                  8.113368 = idf(docFreq=35, maxDocs=44218)
                  0.037353165 = queryNorm
                1.014171 = fieldWeight in 2477, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  8.113368 = idf(docFreq=35, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2477)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Abstract
    This book constitutes the refereed proceedings of the joint 6th International Semantic Web Conference, ISWC 2007, and the 2nd Asian Semantic Web Conference, ASWC 2007, held in Busan, Korea, in November 2007. The 50 revised full academic papers and 12 revised application papers presented together with 5 Semantic Web Challenge papers and 12 selected doctoral consortium articles were carefully reviewed and selected from a total of 257 submitted papers to the academic track and 29 to the applications track. The papers address all current issues in the field of the semantic Web, ranging from theoretical and foundational aspects to various applied topics such as management of semantic Web data, ontologies, semantic Web architecture, social semantic Web, as well as applications of the semantic Web. Short descriptions of the top five winning applications submitted to the Semantic Web Challenge competition conclude the volume.
    LCSH
    Semantic Web / Congresses
    Web site development / Congresses
    Knowledge management / Congresses
    Ontology / Congresses
    Data mining
    Data Mining and Knowledge Discovery
    RSWK
    Semantic Web / Metadatenmodell / Data Mining / Ontologie <Wissensverarbeitung> / Kongress / Pusan <2007> (BVB)
    Subject
    Semantic Web / Congresses
    Web site development / Congresses
    Knowledge management / Congresses
    Ontology / Congresses
    Data mining
    Data Mining and Knowledge Discovery
    Semantic Web / Metadatenmodell / Data Mining / Ontologie <Wissensverarbeitung> / Kongress / Pusan <2007> (BVB)
  7. Papadakis, I. et al.: Highlighting timely information in libraries through social and semantic Web technologies (2016) 0.04
    0.03823903 = product of:
      0.17207563 = sum of:
        0.14677143 = weight(_text_:germany in 2090) [ClassicSimilarity], result of:
          0.14677143 = score(doc=2090,freq=2.0), product of:
            0.22275731 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.037353165 = queryNorm
            0.65888494 = fieldWeight in 2090, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.078125 = fieldNorm(doc=2090)
        0.02530419 = product of:
          0.05060838 = sum of:
            0.05060838 = weight(_text_:22 in 2090) [ClassicSimilarity], result of:
              0.05060838 = score(doc=2090,freq=2.0), product of:
                0.13080442 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037353165 = queryNorm
                0.38690117 = fieldWeight in 2090, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2090)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Source
    Metadata and semantics research: 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings. Eds.: E. Garoufallou
  8. Miller, E.; Schloss. B.; Lassila, O.; Swick, R.R.: Resource Description Framework (RDF) : model and syntax (1997) 0.03
    0.033079453 = product of:
      0.09923836 = sum of:
        0.02243571 = weight(_text_:cataloging in 5903) [ClassicSimilarity], result of:
          0.02243571 = score(doc=5903,freq=2.0), product of:
            0.14721331 = queryWeight, product of:
              3.9411201 = idf(docFreq=2334, maxDocs=44218)
              0.037353165 = queryNorm
            0.15240273 = fieldWeight in 5903, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9411201 = idf(docFreq=2334, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5903)
        0.04332707 = weight(_text_:data in 5903) [ClassicSimilarity], result of:
          0.04332707 = score(doc=5903,freq=18.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.36682853 = fieldWeight in 5903, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5903)
        0.033475578 = weight(_text_:processing in 5903) [ClassicSimilarity], result of:
          0.033475578 = score(doc=5903,freq=4.0), product of:
            0.15121111 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.037353165 = queryNorm
            0.22138305 = fieldWeight in 5903, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5903)
      0.33333334 = coord(3/9)
    
    Abstract
    RDF - the Resource Description Framework - is a foundation for processing metadata; it provides interoperability between applications that exchange machine-understandable information on the Web. RDF emphasizes facilities to enable automated processing of Web resources. RDF metadata can be used in a variety of application areas; for example: in resource discovery to provide better search engine capabilities; in cataloging for describing the content and content relationships available at a particular Web site, page, or digital library; by intelligent software agents to facilitate knowledge sharing and exchange; in content rating; in describing collections of pages that represent a single logical "document"; for describing intellectual property rights of Web pages, and in many others. RDF with digital signatures will be key to building the "Web of Trust" for electronic commerce, collaboration, and other applications. Metadata is "data about data" or specifically in the context of RDF "data describing web resources." The distinction between "data" and "metadata" is not an absolute one; it is a distinction created primarily by a particular application. Many times the same resource will be interpreted in both ways simultaneously. RDF encourages this view by using XML as the encoding syntax for the metadata. The resources being described by RDF are, in general, anything that can be named via a URI. The broad goal of RDF is to define a mechanism for describing resources that makes no assumptions about a particular application domain, nor defines the semantics of any application domain. The definition of the mechanism should be domain neutral, yet the mechanism should be suitable for describing information about any domain. This document introduces a model for representing RDF metadata and one syntax for expressing and transporting this metadata in a manner that maximizes the interoperability of independently developed web servers and clients. The syntax described in this document is best considered as a "serialization syntax" for the underlying RDF representation model. The serialization syntax is XML, XML being the W3C's work-in-progress to define a richer Web syntax for a variety of applications. RDF and XML are complementary; there will be alternate ways to represent the same RDF data model, some more suitable for direct human authoring. Future work may lead to including such alternatives in this document.
    Content
    RDF Data Model At the core of RDF is a model for representing named properties and their values. These properties serve both to represent attributes of resources (and in this sense correspond to usual attribute-value-pairs) and to represent relationships between resources. The RDF data model is a syntax-independent way of representing RDF statements. RDF statements that are syntactically very different could mean the same thing. This concept of equivalence in meaning is very important when performing queries, aggregation and a number of other tasks at which RDF is aimed. The equivalence is defined in a clean machine understandable way. Two pieces of RDF are equivalent if and only if their corresponding data model representations are the same. Table of contents 1. Introduction 2. RDF Data Model 3. RDF Grammar 4. Signed RDF 5. Examples 6. Appendix A: Brief Explanation of XML Namespaces
  9. Rüther, M.; Fock, J.; Schultz-Krutisch, T.; Bandholtz, T.: Classification and reference vocabulary in linked environment data (2011) 0.03
    0.029099008 = product of:
      0.13094553 = sum of:
        0.08806286 = weight(_text_:germany in 4816) [ClassicSimilarity], result of:
          0.08806286 = score(doc=4816,freq=2.0), product of:
            0.22275731 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.037353165 = queryNorm
            0.39533097 = fieldWeight in 4816, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.046875 = fieldNorm(doc=4816)
        0.04288268 = weight(_text_:data in 4816) [ClassicSimilarity], result of:
          0.04288268 = score(doc=4816,freq=6.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.3630661 = fieldWeight in 4816, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046875 = fieldNorm(doc=4816)
      0.22222222 = coord(2/9)
    
    Abstract
    The Federal Environment Agency (UBA), Germany, has a long tradition in knowledge organization, using a library along with many Web-based information systems. The backbone of this information space is a classification system enhanced by a reference vocabulary which consists of a thesaurus, a gazetteer and a chronicle. Over the years, classification has increasingly been relegated to the background compared with the reference vocabulary indexing and full text search. Bibliographic items are no longer classified directly but tagged with thesaurus terms, with those terms being classified. Since 2010 we have been developing a linked data representation of this knowledge base. While we are linking bibliographic and observation data with the controlled vocabulary in a Resource Desrcription Framework (RDF) representation, the classification may be revisited as a powerful organization system by inference. This also raises questions about the quality and feasibility of an unambiguous classification of thesaurus terms.
  10. Shoffner, M.; Greenberg, J.; Kramer-Duffield, J.; Woodbury, D.: Web 2.0 semantic systems : collaborative learning in science (2008) 0.03
    0.028445266 = product of:
      0.0853358 = sum of:
        0.058708575 = weight(_text_:germany in 2661) [ClassicSimilarity], result of:
          0.058708575 = score(doc=2661,freq=2.0), product of:
            0.22275731 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.037353165 = queryNorm
            0.26355398 = fieldWeight in 2661, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.03125 = fieldNorm(doc=2661)
        0.01650555 = weight(_text_:data in 2661) [ClassicSimilarity], result of:
          0.01650555 = score(doc=2661,freq=2.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.1397442 = fieldWeight in 2661, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=2661)
        0.010121676 = product of:
          0.020243352 = sum of:
            0.020243352 = weight(_text_:22 in 2661) [ClassicSimilarity], result of:
              0.020243352 = score(doc=2661,freq=2.0), product of:
                0.13080442 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037353165 = queryNorm
                0.15476047 = fieldWeight in 2661, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2661)
          0.5 = coord(1/2)
      0.33333334 = coord(3/9)
    
    Abstract
    The basic goal of education within a discipline is to transform a novice into an expert. This entails moving the novice toward the "semantic space" that the expert inhabits-the space of concepts, meanings, vocabularies, and other intellectual constructs that comprise the discipline. Metadata is significant to this goal in digitally mediated education environments. Encoding the experts' semantic space not only enables the sharing of semantics among discipline scientists, but also creates an environment that bridges the semantic gap between the common vocabulary of the novice and the granular descriptive language of the seasoned scientist (Greenberg, et al, 2005). Developments underlying the Semantic Web, where vocabularies are formalized in the Web Ontology Language (OWL), and Web 2.0 approaches of user-generated folksonomies provide an infrastructure for linking vocabulary systems and promoting group learning via metadata literacy. Group learning is a pedagogical approach to teaching that harnesses the phenomenon of "collective intelligence" to increase learning by means of collaboration. Learning a new semantic system can be daunting for a novice, and yet it is integral to advance one's knowledge in a discipline and retain interest. These ideas are key to the "BOT 2.0: Botany through Web 2.0, the Memex and Social Learning" project (Bot 2.0).72 Bot 2.0 is a collaboration involving the North Carolina Botanical Garden, the UNC SILS Metadata Research center, and the Renaissance Computing Institute (RENCI). Bot 2.0 presents a curriculum utilizing a memex as a way for students to link and share digital information, working asynchronously in an environment beyond the traditional classroom. Our conception of a memex is not a centralized black box but rather a flexible, distributed framework that uses the most salient and easiest-to-use collaborative platforms (e.g., Facebook, Flickr, wiki and blog technology) for personal information management. By meeting students "where they live" digitally, we hope to attract students to the study of botanical science. A key aspect is to teach students scientific terminology and about the value of metadata, an inherent function in several of the technologies and in the instructional approach we are utilizing. This poster will report on a study examining the value of both folksonomies and taxonomies for post-secondary college students learning plant identification. Our data is drawn from a curriculum involving a virtual independent learning portion and a "BotCamp" weekend at UNC, where students work with digital plan specimens that they have captured. Results provide some insight into the importance of collaboration and shared vocabulary for gaining confidence and for student progression from novice to expert in botany.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  11. Coyle, K.: Understanding the Semantic Web : bibliographic data and metadata (2010) 0.03
    0.028097574 = product of:
      0.12643908 = sum of:
        0.07692243 = weight(_text_:cataloging in 4169) [ClassicSimilarity], result of:
          0.07692243 = score(doc=4169,freq=2.0), product of:
            0.14721331 = queryWeight, product of:
              3.9411201 = idf(docFreq=2334, maxDocs=44218)
              0.037353165 = queryNorm
            0.52252364 = fieldWeight in 4169, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9411201 = idf(docFreq=2334, maxDocs=44218)
              0.09375 = fieldNorm(doc=4169)
        0.04951665 = weight(_text_:data in 4169) [ClassicSimilarity], result of:
          0.04951665 = score(doc=4169,freq=2.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.4192326 = fieldWeight in 4169, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.09375 = fieldNorm(doc=4169)
      0.22222222 = coord(2/9)
    
    Footnote
    Rez. in: Cataloging & Classification Quarterly, Volume 49(2011) no.1, S.51-52 (Marcia Salmon).
  12. Shaw, R.; Buckland, M.: Open identification and linking of the four Ws (2008) 0.03
    0.026883684 = product of:
      0.08065105 = sum of:
        0.051370002 = weight(_text_:germany in 2665) [ClassicSimilarity], result of:
          0.051370002 = score(doc=2665,freq=2.0), product of:
            0.22275731 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.037353165 = queryNorm
            0.23060973 = fieldWeight in 2665, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2665)
        0.020424578 = weight(_text_:data in 2665) [ClassicSimilarity], result of:
          0.020424578 = score(doc=2665,freq=4.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.17292464 = fieldWeight in 2665, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2665)
        0.008856467 = product of:
          0.017712934 = sum of:
            0.017712934 = weight(_text_:22 in 2665) [ClassicSimilarity], result of:
              0.017712934 = score(doc=2665,freq=2.0), product of:
                0.13080442 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037353165 = queryNorm
                0.1354154 = fieldWeight in 2665, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=2665)
          0.5 = coord(1/2)
      0.33333334 = coord(3/9)
    
    Abstract
    Platforms for social computing connect users via shared references to people with whom they have relationships, events attended, places lived in or traveled to, and topics such as favorite books or movies. Since free text is insufficient for expressing such references precisely and unambiguously, many social computing platforms coin identifiers for topics, places, events, and people and provide interfaces for finding and selecting these identifiers from controlled lists. Using these interfaces, users collaboratively construct a web of links among entities. This model needn't be limited to social networking sites. Understanding an item in a digital library or museum requires context: information about the topics, places, events, and people to which the item is related. Students, journalists and investigators traditionally discover this kind of context by asking "the four Ws": what, where, when and who. The DCMI Kernel Metadata Community has recognized the four Ws as fundamental elements of descriptions (Kunze & Turner, 2007). Making better use of metadata to answer these questions via links to appropriate contextual resources has been our focus in a series of research projects over the past few years. Currently we are building a system for enabling readers of any text to relate any topic, place, event or person mentioned in the text to the best explanatory resources available. This system is being developed with two different corpora: a diverse variety of biographical texts characterized by very rich and dense mentions of people, events, places and activities, and a large collection of newly-scanned books, journals and manuscripts relating to Irish culture and history. Like a social computing platform, our system consists of tools for referring to topics, places, events or people, disambiguating these references by linking them to unique identifiers, and using the disambiguated references to provide useful information in context and to link to related resources. Yet current social computing platforms, while usually amenable to importing and exporting data, tend to mint proprietary identifiers and expect links to be traversed using their own interfaces. We take a different approach, using identifiers from both established and emerging naming authorities, representing relationships using standardized metadata vocabularies, and publishing those representations using standard protocols so that links can be stored and traversed anywhere. Central to our strategy is to move from appearances in a text to naming authorities to the the construction of links for searching or querying trusted resources. Using identifiers from naming authorities, rather than literal values (as in the DCMI Kernel) or keys from a proprietary database, makes it more likely that links constructed using our system will continue to be useful in the future. WorldCat Identities URIs (http://worldcat.org/identities/) linked to Library of Congress and Deutsche Nationalbibliothek authority files for persons and organizations and Geonames (http://geonames.org/) URIs for places are stable identifiers attached to a wealth of useful metadata. Yet no naming authority can be totally comprehensive, so our system can be extended to use new sources of identifiers as needed. For example, we are experimenting with using Freebase (http://freebase.com/) URIs to identify historical events, for which no established naming authority currently exists. Stable identifiers (URIs), standardized hyperlinked data formats (XML), and uniform publishing protocols (HTTP) are key ingredients of the web's open architecture. Our system provides an example of how this open architecture can be exploited to build flexible and useful tools for connecting resources via shared references to topics, places, events, and people.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  13. Piscitelli, F.A.: Library linked data models : library data in the Semantic Web (2019) 0.02
    0.024571309 = product of:
      0.11057089 = sum of:
        0.04532698 = weight(_text_:cataloging in 5478) [ClassicSimilarity], result of:
          0.04532698 = score(doc=5478,freq=4.0), product of:
            0.14721331 = queryWeight, product of:
              3.9411201 = idf(docFreq=2334, maxDocs=44218)
              0.037353165 = queryNorm
            0.3079 = fieldWeight in 5478, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9411201 = idf(docFreq=2334, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5478)
        0.065243915 = weight(_text_:data in 5478) [ClassicSimilarity], result of:
          0.065243915 = score(doc=5478,freq=20.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.5523875 = fieldWeight in 5478, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5478)
      0.22222222 = coord(2/9)
    
    Abstract
    This exploratory study examined Linked Data (LD) schemas/ontologies and data models proposed or in use by libraries around the world using MAchine Readable Cataloging (MARC) as a basis for comparison of the scope and extensibility of these potential new standards. The researchers selected 14 libraries from national libraries, academic libraries, government libraries, public libraries, multi-national libraries, and cultural heritage centers currently developing Library Linked Data (LLD) schemas. The choices of models, schemas, and elements used in each library's LD can create interoperability issues for LD services because of substantial differences between schemas and data models evolving via local decisions. The researchers observed that a wide variety of vocabularies and ontologies were used for LLD including common web schemas such as Dublin Core (DC)/DCTerms, Schema.org and Resource Description Framework (RDF), as well as deprecated schemas such as MarcOnt and rdagroup1elements. A sharp divide existed as well between LLD schemas using variations of the Functional Requirements for Bibliographic Records (FRBR) data model and those with different data models or even with no listed data model. Libraries worldwide are not using the same elements or even the same ontologies, schemas and data models to describe the same materials using the same general concepts.
    Source
    Cataloging and classification quarterly. 57(2019) no.5, S.261-277
  14. Bianchini, C.; Willer, M.: ISBD resource and Its description in the context of the Semantic Web (2014) 0.02
    0.024324356 = product of:
      0.1094596 = sum of:
        0.04487142 = weight(_text_:cataloging in 1998) [ClassicSimilarity], result of:
          0.04487142 = score(doc=1998,freq=2.0), product of:
            0.14721331 = queryWeight, product of:
              3.9411201 = idf(docFreq=2334, maxDocs=44218)
              0.037353165 = queryNorm
            0.30480546 = fieldWeight in 1998, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9411201 = idf(docFreq=2334, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1998)
        0.06458818 = weight(_text_:data in 1998) [ClassicSimilarity], result of:
          0.06458818 = score(doc=1998,freq=10.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.5468357 = fieldWeight in 1998, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1998)
      0.22222222 = coord(2/9)
    
    Abstract
    This article explores the question "What is an International Standard for Bibliographic Description (ISBD) resource in the context of the Semantic Web, and what is the relationship of its description to the linked data?" This question is discussed against the background of the dichotomy between the description and access using the Semantic Web differentiation of the three logical layers: real-world objects, web of data, and special purpose (bibliographic) data. The representation of bibliographic data as linked data is discussed, distinguishing the description of a resource from the iconic/objective and the informational/subjective viewpoints. In the conclusion, the authors give views on possible directions of future development of the ISBD.
    Source
    Cataloging and classification quarterly. 52(2014) no.8, S.869-887
  15. Wang, H.; Liu, Q.; Penin, T.; Fu, L.; Zhang, L.; Tran, T.; Yu, Y.; Pan, Y.: Semplore: a scalable IR approach to search the Web of Data (2009) 0.02
    0.023573978 = product of:
      0.1060829 = sum of:
        0.06550438 = weight(_text_:data in 1638) [ClassicSimilarity], result of:
          0.06550438 = score(doc=1638,freq=14.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.55459267 = fieldWeight in 1638, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046875 = fieldNorm(doc=1638)
        0.040578526 = weight(_text_:processing in 1638) [ClassicSimilarity], result of:
          0.040578526 = score(doc=1638,freq=2.0), product of:
            0.15121111 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.037353165 = queryNorm
            0.26835677 = fieldWeight in 1638, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.046875 = fieldNorm(doc=1638)
      0.22222222 = coord(2/9)
    
    Abstract
    The Web of Data keeps growing rapidly. However, the full exploitation of this large amount of structured data faces numerous challenges like usability, scalability, imprecise information needs and data change. We present Semplore, an IR-based system that aims at addressing these issues. Semplore supports intuitive faceted search and complex queries both on text and structured data. It combines imprecise keyword search and precise structured query in a unified ranking scheme. Scalable query processing is supported by leveraging inverted indexes traditionally used in IR systems. This is combined with a novel block-based index structure to support efficient index update when data changes. The experimental results show that Semplore is an efficient and effective system for searching the Web of Data and can be used as a basic infrastructure for Web-scale Semantic Web search engines.
  16. Metadata and semantics research : 9th Research Conference, MTSR 2015, Manchester, UK, September 9-11, 2015, Proceedings (2015) 0.02
    0.022282086 = product of:
      0.100269385 = sum of:
        0.04288268 = weight(_text_:data in 3274) [ClassicSimilarity], result of:
          0.04288268 = score(doc=3274,freq=6.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.3630661 = fieldWeight in 3274, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046875 = fieldNorm(doc=3274)
        0.057386704 = weight(_text_:processing in 3274) [ClassicSimilarity], result of:
          0.057386704 = score(doc=3274,freq=4.0), product of:
            0.15121111 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.037353165 = queryNorm
            0.3795138 = fieldWeight in 3274, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.046875 = fieldNorm(doc=3274)
      0.22222222 = coord(2/9)
    
    Content
    The papers are organized in several sessions and tracks: general track on ontology evolution, engineering, and frameworks, semantic Web and metadata extraction, modelling, interoperability and exploratory search, data analysis, reuse and visualization; track on digital libraries, information retrieval, linked and social data; track on metadata and semantics for open repositories, research information systems and data infrastructure; track on metadata and semantics for agriculture, food and environment; track on metadata and semantics for cultural collections and applications; track on European and national projects.
    LCSH
    Text processing (Computer science)
    Subject
    Text processing (Computer science)
  17. Blanco, L.; Bronzi, M.; Crescenzi, V.; Merialdo, P.; Papotti, P.: Flint: from Web pages to probabilistic semantic data (2012) 0.02
    0.022013191 = product of:
      0.09905936 = sum of:
        0.065243915 = weight(_text_:data in 437) [ClassicSimilarity], result of:
          0.065243915 = score(doc=437,freq=20.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.5523875 = fieldWeight in 437, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0390625 = fieldNorm(doc=437)
        0.03381544 = weight(_text_:processing in 437) [ClassicSimilarity], result of:
          0.03381544 = score(doc=437,freq=2.0), product of:
            0.15121111 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.037353165 = queryNorm
            0.22363065 = fieldWeight in 437, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0390625 = fieldNorm(doc=437)
      0.22222222 = coord(2/9)
    
    Abstract
    The Web is a surprisingly extensive source of information: it offers a huge number of sites containing data about a disparate range of topics. Although Web pages are built for human fruition, not for automatic processing of the data, we observe that an increasing number of Web sites deliver pages containing structured information about recognizable concepts, relevant to specific application domains, such as movies, finance, sport, products, etc. The development of scalable techniques to discover, extract, and integrate data from fairly structured large corpora available on the Web is a challenging issue, because to face the Web scale, these activities should be accomplished automatically by domain-independent techniques. To cope with the complexity and the heterogeneity of Web data, state-of-the-art approaches focus on information organized according to specific patterns that frequently occur on the Web. Meaningful examples are WebTables, which focuses on data published in HTML tables, and information extraction systems, such as TextRunner, which exploits lexical-syntactic patterns. As noticed by Cafarella et al., even if a small fraction of the Web is organized according to these patterns, due to the Web scale, the amount of data involved is impressive. In this chapter, we focus on methods and techniques to wring out value from the data delivered by large data-intensive Web sites.
    Series
    Data-centric systems and applications
  18. Sakr, S.; Wylot, M.; Mutharaju, R.; Le-Phuoc, D.; Fundulaki, I.: Linked data : storing, querying, and reasoning (2018) 0.02
    0.020217352 = product of:
      0.09097808 = sum of:
        0.06392573 = weight(_text_:data in 5329) [ClassicSimilarity], result of:
          0.06392573 = score(doc=5329,freq=30.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.54122704 = fieldWeight in 5329, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=5329)
        0.027052352 = weight(_text_:processing in 5329) [ClassicSimilarity], result of:
          0.027052352 = score(doc=5329,freq=2.0), product of:
            0.15121111 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.037353165 = queryNorm
            0.17890452 = fieldWeight in 5329, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.03125 = fieldNorm(doc=5329)
      0.22222222 = coord(2/9)
    
    Abstract
    This book describes efficient and effective techniques for harnessing the power of Linked Data by tackling the various aspects of managing its growing volume: storing, querying, reasoning, provenance management and benchmarking. To this end, Chapter 1 introduces the main concepts of the Semantic Web and Linked Data and provides a roadmap for the book. Next, Chapter 2 briefly presents the basic concepts underpinning Linked Data technologies that are discussed in the book. Chapter 3 then offers an overview of various techniques and systems for centrally querying RDF datasets, and Chapter 4 outlines various techniques and systems for efficiently querying large RDF datasets in distributed environments. Subsequently, Chapter 5 explores how streaming requirements are addressed in current, state-of-the-art RDF stream data processing. Chapter 6 covers performance and scaling issues of distributed RDF reasoning systems, while Chapter 7 details benchmarks for RDF query engines and instance matching systems. Chapter 8 addresses the provenance management for Linked Data and presents the different provenance models developed. Lastly, Chapter 9 offers a brief summary, highlighting and providing insights into some of the open challenges and research directions. Providing an updated overview of methods, technologies and systems related to Linked Data this book is mainly intended for students and researchers who are interested in the Linked Data domain. It enables students to gain an understanding of the foundations and underpinning technologies and standards for Linked Data, while researchers benefit from the in-depth coverage of the emerging and ongoing advances in Linked Data storing, querying, reasoning, and provenance management systems. Further, it serves as a starting point to tackle the next research challenges in the domain of Linked Data management.
    LCSH
    Linked data
    RSWK
    Linked Data
    Subject
    Linked Data
    Linked data
  19. Cali, A.: Ontology querying : datalog strikes back (2017) 0.02
    0.018254451 = product of:
      0.08214503 = sum of:
        0.024758326 = weight(_text_:data in 3928) [ClassicSimilarity], result of:
          0.024758326 = score(doc=3928,freq=2.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.2096163 = fieldWeight in 3928, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046875 = fieldNorm(doc=3928)
        0.057386704 = weight(_text_:processing in 3928) [ClassicSimilarity], result of:
          0.057386704 = score(doc=3928,freq=4.0), product of:
            0.15121111 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.037353165 = queryNorm
            0.3795138 = fieldWeight in 3928, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.046875 = fieldNorm(doc=3928)
      0.22222222 = coord(2/9)
    
    Abstract
    In this tutorial we address the problem of ontology querying, that is, the problem of answering queries against a theory constituted by facts (the data) and inference rules (the ontology). A varied landscape of ontology languages exists in the scientific literature, with several degrees of complexity of query processing. We argue that Datalog±, a family of languages derived from Datalog, is a powerful tool for ontology querying. To illustrate the impact of this comeback of Datalog, we present the basic paradigms behind the main Datalog± as well as some recent extensions. We also present some efficient query processing techniques for some cases.
  20. Brunetti, J.M.; Roberto García, R.: User-centered design and evaluation of overview components for semantic data exploration (2014) 0.02
    0.01692086 = product of:
      0.076143876 = sum of:
        0.0660222 = weight(_text_:data in 1626) [ClassicSimilarity], result of:
          0.0660222 = score(doc=1626,freq=32.0), product of:
            0.118112594 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.037353165 = queryNorm
            0.5589768 = fieldWeight in 1626, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=1626)
        0.010121676 = product of:
          0.020243352 = sum of:
            0.020243352 = weight(_text_:22 in 1626) [ClassicSimilarity], result of:
              0.020243352 = score(doc=1626,freq=2.0), product of:
                0.13080442 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037353165 = queryNorm
                0.15476047 = fieldWeight in 1626, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1626)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Abstract
    Purpose - The growing volumes of semantic data available in the web result in the need for handling the information overload phenomenon. The potential of this amount of data is enormous but in most cases it is very difficult for users to visualize, explore and use this data, especially for lay-users without experience with Semantic Web technologies. The paper aims to discuss these issues. Design/methodology/approach - The Visual Information-Seeking Mantra "Overview first, zoom and filter, then details-on-demand" proposed by Shneiderman describes how data should be presented in different stages to achieve an effective exploration. The overview is the first user task when dealing with a data set. The objective is that the user is capable of getting an idea about the overall structure of the data set. Different information architecture (IA) components supporting the overview tasks have been developed, so they are automatically generated from semantic data, and evaluated with end-users. Findings - The chosen IA components are well known to web users, as they are present in most web pages: navigation bars, site maps and site indexes. The authors complement them with Treemaps, a visualization technique for displaying hierarchical data. These components have been developed following an iterative User-Centered Design methodology. Evaluations with end-users have shown that they get easily used to them despite the fact that they are generated automatically from structured data, without requiring knowledge about the underlying semantic technologies, and that the different overview components complement each other as they focus on different information search needs. Originality/value - Obtaining semantic data sets overviews cannot be easily done with the current semantic web browsers. Overviews become difficult to achieve with large heterogeneous data sets, which is typical in the Semantic Web, because traditional IA techniques do not easily scale to large data sets. There is little or no support to obtain overview information quickly and easily at the beginning of the exploration of a new data set. This can be a serious limitation when exploring a data set for the first time, especially for lay-users. The proposal is to reuse and adapt existing IA components to provide this overview to users and show that they can be generated automatically from the thesaurus and ontologies that structure semantic data while providing a comparable user experience to traditional web sites.
    Date
    20. 1.2015 18:30:22

Years

Languages

  • e 168
  • d 23

Types

  • a 119
  • el 44
  • m 42
  • s 18
  • n 5
  • x 5
  • r 1
  • More… Less…

Subjects

Classifications