Search (93 results, page 1 of 5)

  • × theme_ss:"Semantic Web"
  • × type_ss:"a"
  • × year_i:[2010 TO 2020}
  1. Oliveira Machado, L.M.; Souza, R.R.; Simões, M. da Graça: Semantic web or web of data? : a diachronic study (1999 to 2017) of the publications of Tim Berners-Lee and the World Wide Web Consortium (2019) 0.07
    0.07356448 = product of:
      0.09808597 = sum of:
        0.010507616 = weight(_text_:a in 5300) [ClassicSimilarity], result of:
          0.010507616 = score(doc=5300,freq=16.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.18016359 = fieldWeight in 5300, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5300)
        0.08012097 = weight(_text_:70 in 5300) [ClassicSimilarity], result of:
          0.08012097 = score(doc=5300,freq=2.0), product of:
            0.27085114 = queryWeight, product of:
              5.354766 = idf(docFreq=567, maxDocs=44218)
              0.05058132 = queryNorm
            0.29581183 = fieldWeight in 5300, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.354766 = idf(docFreq=567, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5300)
        0.0074573862 = product of:
          0.0149147725 = sum of:
            0.0149147725 = weight(_text_:information in 5300) [ClassicSimilarity], result of:
              0.0149147725 = score(doc=5300,freq=6.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.16796975 = fieldWeight in 5300, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5300)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Abstract
    The web has been, in the last decades, the place where information retrieval achieved its maximum importance, given its ubiquity and the sheer volume of information. However, its exponential growth made the retrieval task increasingly hard, relying in its effectiveness on idiosyncratic and somewhat biased ranking algorithms. To deal with this problem, a "new" web, called the Semantic Web (SW), was proposed, bringing along concepts like "Web of Data" and "Linked Data," although the definitions and connections among these concepts are often unclear. Based on a qualitative approach built over a literature review, a definition of SW is presented, discussing the related concepts sometimes used as synonyms. It concludes that the SW is a comprehensive and ambitious construct that includes the great purpose of making the web a global database. It also follows the specifications developed and/or associated with its operationalization and the necessary procedures for the connection of data in an open format on the web. The goals of this comprehensive SW are the union of two outcomes still tenuously connected: the virtually unlimited possibility of connections between data-the web domain-with the potentiality of the automated inference of "intelligent" systems-the semantic component.
    Source
    Journal of the Association for Information Science and Technology. 70(2019) no.7, S.701-714
    Type
    a
  2. Fensel, A.: Towards semantic APIs for research data services (2017) 0.06
    0.05976234 = product of:
      0.11952468 = sum of:
        0.0073553314 = weight(_text_:a in 4439) [ClassicSimilarity], result of:
          0.0073553314 = score(doc=4439,freq=4.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.12611452 = fieldWeight in 4439, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4439)
        0.11216935 = weight(_text_:70 in 4439) [ClassicSimilarity], result of:
          0.11216935 = score(doc=4439,freq=2.0), product of:
            0.27085114 = queryWeight, product of:
              5.354766 = idf(docFreq=567, maxDocs=44218)
              0.05058132 = queryNorm
            0.41413653 = fieldWeight in 4439, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.354766 = idf(docFreq=567, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4439)
      0.5 = coord(2/4)
    
    Source
    Mitteilungen der Vereinigung Österreichischer Bibliothekarinnen und Bibliothekare. 70(2017) H.2, S.157-169
    Type
    a
  3. Papadakis, I. et al.: Highlighting timely information in libraries through social and semantic Web technologies (2016) 0.05
    0.050158218 = product of:
      0.100316435 = sum of:
        0.007430006 = weight(_text_:a in 2090) [ClassicSimilarity], result of:
          0.007430006 = score(doc=2090,freq=2.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.12739488 = fieldWeight in 2090, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=2090)
        0.092886426 = sum of:
          0.024355719 = weight(_text_:information in 2090) [ClassicSimilarity], result of:
            0.024355719 = score(doc=2090,freq=4.0), product of:
              0.088794395 = queryWeight, product of:
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.05058132 = queryNorm
              0.27429342 = fieldWeight in 2090, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.078125 = fieldNorm(doc=2090)
          0.06853071 = weight(_text_:22 in 2090) [ClassicSimilarity], result of:
            0.06853071 = score(doc=2090,freq=2.0), product of:
              0.17712717 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05058132 = queryNorm
              0.38690117 = fieldWeight in 2090, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.078125 = fieldNorm(doc=2090)
      0.5 = coord(2/4)
    
    Series
    Communications in computer and information science; 672
    Source
    Metadata and semantics research: 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings. Eds.: E. Garoufallou
    Type
    a
  4. Neubauer, G.: Visualization of typed links in linked data (2017) 0.04
    0.041917987 = product of:
      0.083835974 = sum of:
        0.003715003 = weight(_text_:a in 3912) [ClassicSimilarity], result of:
          0.003715003 = score(doc=3912,freq=2.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.06369744 = fieldWeight in 3912, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3912)
        0.08012097 = weight(_text_:70 in 3912) [ClassicSimilarity], result of:
          0.08012097 = score(doc=3912,freq=2.0), product of:
            0.27085114 = queryWeight, product of:
              5.354766 = idf(docFreq=567, maxDocs=44218)
              0.05058132 = queryNorm
            0.29581183 = fieldWeight in 3912, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.354766 = idf(docFreq=567, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3912)
      0.5 = coord(2/4)
    
    Source
    Mitteilungen der Vereinigung Österreichischer Bibliothekarinnen und Bibliothekare. 70(2017) H.2, S.179-199
    Type
    a
  5. Hooland, S. van; Verborgh, R.; Wilde, M. De; Hercher, J.; Mannens, E.; Wa, R.Van de: Evaluating the success of vocabulary reconciliation for cultural heritage collections (2013) 0.03
    0.030710042 = product of:
      0.061420083 = sum of:
        0.0099684 = weight(_text_:a in 662) [ClassicSimilarity], result of:
          0.0099684 = score(doc=662,freq=10.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.1709182 = fieldWeight in 662, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=662)
        0.051451683 = sum of:
          0.010333257 = weight(_text_:information in 662) [ClassicSimilarity], result of:
            0.010333257 = score(doc=662,freq=2.0), product of:
              0.088794395 = queryWeight, product of:
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.05058132 = queryNorm
              0.116372846 = fieldWeight in 662, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.046875 = fieldNorm(doc=662)
          0.041118424 = weight(_text_:22 in 662) [ClassicSimilarity], result of:
            0.041118424 = score(doc=662,freq=2.0), product of:
              0.17712717 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05058132 = queryNorm
              0.23214069 = fieldWeight in 662, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=662)
      0.5 = coord(2/4)
    
    Abstract
    The concept of Linked Data has made its entrance in the cultural heritage sector due to its potential use for the integration of heterogeneous collections and deriving additional value out of existing metadata. However, practitioners and researchers alike need a better understanding of what outcome they can reasonably expect of the reconciliation process between their local metadata and established controlled vocabularies which are already a part of the Linked Data cloud. This paper offers an in-depth analysis of how a locally developed vocabulary can be successfully reconciled with the Library of Congress Subject Headings (LCSH) and the Arts and Architecture Thesaurus (AAT) through the help of a general-purpose tool for interactive data transformation (OpenRefine). Issues negatively affecting the reconciliation process are identified and solutions are proposed in order to derive maximum value from existing metadata and controlled vocabularies in an automated manner.
    Date
    22. 3.2013 19:29:20
    Source
    Journal of the American Society for Information Science and Technology. 64(2013) no.3, S.464-479
    Type
    a
  6. Prud'hommeaux, E.; Gayo, E.: RDF ventures to boldly meet your most pedestrian needs (2015) 0.03
    0.029586587 = product of:
      0.059173174 = sum of:
        0.0077214893 = weight(_text_:a in 2024) [ClassicSimilarity], result of:
          0.0077214893 = score(doc=2024,freq=6.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.13239266 = fieldWeight in 2024, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=2024)
        0.051451683 = sum of:
          0.010333257 = weight(_text_:information in 2024) [ClassicSimilarity], result of:
            0.010333257 = score(doc=2024,freq=2.0), product of:
              0.088794395 = queryWeight, product of:
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.05058132 = queryNorm
              0.116372846 = fieldWeight in 2024, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.046875 = fieldNorm(doc=2024)
          0.041118424 = weight(_text_:22 in 2024) [ClassicSimilarity], result of:
            0.041118424 = score(doc=2024,freq=2.0), product of:
              0.17712717 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05058132 = queryNorm
              0.23214069 = fieldWeight in 2024, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2024)
      0.5 = coord(2/4)
    
    Abstract
    Defined in 1999 and paired with XML, the Resource Description Framework (RDF) has been cast as an RDF Schema, producing data that is well-structured but not validated, permitting certain illogical relationships. When stakeholders convened in 2014 to consider solutions to the data validation challenge, a W3C working group proposed Resource Shapes and Shape Expressions to describe the properties expected for an RDF node. Resistance rose from concerns about data and schema reuse, key principles in RDF. Ideally data types and properties are designed for broad use, but they are increasingly adopted with local restrictions for specific purposes. Resource Shapes are commonly treated as record classes, standing in for data structures but losing flexibility for later reuse. Of various solutions to the resulting tensions, the concept of record classes may be the most reasonable basis for agreement, satisfying stakeholders' objectives while allowing for variations with constraints.
    Footnote
    Contribution to a special section "Linked data and the charm of weak semantics".
    Source
    Bulletin of the Association for Information Science and Technology. 41(2015) no.4, S.18-22
    Type
    a
  7. Brunetti, J.M.; Roberto García, R.: User-centered design and evaluation of overview components for semantic data exploration (2014) 0.03
    0.026074799 = product of:
      0.052149598 = sum of:
        0.00786318 = weight(_text_:a in 1626) [ClassicSimilarity], result of:
          0.00786318 = score(doc=1626,freq=14.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.13482209 = fieldWeight in 1626, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=1626)
        0.04428642 = sum of:
          0.016874136 = weight(_text_:information in 1626) [ClassicSimilarity], result of:
            0.016874136 = score(doc=1626,freq=12.0), product of:
              0.088794395 = queryWeight, product of:
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.05058132 = queryNorm
              0.19003606 = fieldWeight in 1626, product of:
                3.4641016 = tf(freq=12.0), with freq of:
                  12.0 = termFreq=12.0
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.03125 = fieldNorm(doc=1626)
          0.027412282 = weight(_text_:22 in 1626) [ClassicSimilarity], result of:
            0.027412282 = score(doc=1626,freq=2.0), product of:
              0.17712717 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05058132 = queryNorm
              0.15476047 = fieldWeight in 1626, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1626)
      0.5 = coord(2/4)
    
    Abstract
    Purpose - The growing volumes of semantic data available in the web result in the need for handling the information overload phenomenon. The potential of this amount of data is enormous but in most cases it is very difficult for users to visualize, explore and use this data, especially for lay-users without experience with Semantic Web technologies. The paper aims to discuss these issues. Design/methodology/approach - The Visual Information-Seeking Mantra "Overview first, zoom and filter, then details-on-demand" proposed by Shneiderman describes how data should be presented in different stages to achieve an effective exploration. The overview is the first user task when dealing with a data set. The objective is that the user is capable of getting an idea about the overall structure of the data set. Different information architecture (IA) components supporting the overview tasks have been developed, so they are automatically generated from semantic data, and evaluated with end-users. Findings - The chosen IA components are well known to web users, as they are present in most web pages: navigation bars, site maps and site indexes. The authors complement them with Treemaps, a visualization technique for displaying hierarchical data. These components have been developed following an iterative User-Centered Design methodology. Evaluations with end-users have shown that they get easily used to them despite the fact that they are generated automatically from structured data, without requiring knowledge about the underlying semantic technologies, and that the different overview components complement each other as they focus on different information search needs. Originality/value - Obtaining semantic data sets overviews cannot be easily done with the current semantic web browsers. Overviews become difficult to achieve with large heterogeneous data sets, which is typical in the Semantic Web, because traditional IA techniques do not easily scale to large data sets. There is little or no support to obtain overview information quickly and easily at the beginning of the exploration of a new data set. This can be a serious limitation when exploring a data set for the first time, especially for lay-users. The proposal is to reuse and adapt existing IA components to provide this overview to users and show that they can be generated automatically from the thesaurus and ontologies that structure semantic data while providing a comparable user experience to traditional web sites.
    Date
    20. 1.2015 18:30:22
    Source
    Aslib journal of information management. 66(2014) no.5, S.519-536
    Type
    a
  8. Zhitomirsky-Geffet, M.; Bar-Ilan, J.: Towards maximal unification of semantically diverse ontologies for controversial domains (2014) 0.02
    0.024137393 = product of:
      0.048274785 = sum of:
        0.011120215 = weight(_text_:a in 1634) [ClassicSimilarity], result of:
          0.011120215 = score(doc=1634,freq=28.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.19066721 = fieldWeight in 1634, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=1634)
        0.03715457 = sum of:
          0.009742287 = weight(_text_:information in 1634) [ClassicSimilarity], result of:
            0.009742287 = score(doc=1634,freq=4.0), product of:
              0.088794395 = queryWeight, product of:
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.05058132 = queryNorm
              0.10971737 = fieldWeight in 1634, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.03125 = fieldNorm(doc=1634)
          0.027412282 = weight(_text_:22 in 1634) [ClassicSimilarity], result of:
            0.027412282 = score(doc=1634,freq=2.0), product of:
              0.17712717 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05058132 = queryNorm
              0.15476047 = fieldWeight in 1634, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1634)
      0.5 = coord(2/4)
    
    Abstract
    Purpose - Ontologies are prone to wide semantic variability due to subjective points of view of their composers. The purpose of this paper is to propose a new approach for maximal unification of diverse ontologies for controversial domains by their relations. Design/methodology/approach - Effective matching or unification of multiple ontologies for a specific domain is crucial for the success of many semantic web applications, such as semantic information retrieval and organization, document tagging, summarization and search. To this end, numerous automatic and semi-automatic techniques were proposed in the past decade that attempt to identify similar entities, mostly classes, in diverse ontologies for similar domains. Apparently, matching individual entities cannot result in full integration of ontologies' semantics without matching their inter-relations with all other-related classes (and instances). However, semantic matching of ontological relations still constitutes a major research challenge. Therefore, in this paper the authors propose a new paradigm for assessment of maximal possible matching and unification of ontological relations. To this end, several unification rules for ontological relations were devised based on ontological reference rules, and lexical and textual entailment. These rules were semi-automatically implemented to extend a given ontology with semantically matching relations from another ontology for a similar domain. Then, the ontologies were unified through these similar pairs of relations. The authors observe that these rules can be also facilitated to reveal the contradictory relations in different ontologies. Findings - To assess the feasibility of the approach two experiments were conducted with different sets of multiple personal ontologies on controversial domains constructed by trained subjects. The results for about 50 distinct ontology pairs demonstrate a good potential of the methodology for increasing inter-ontology agreement. Furthermore, the authors show that the presented methodology can lead to a complete unification of multiple semantically heterogeneous ontologies. Research limitations/implications - This is a conceptual study that presents a new approach for semantic unification of ontologies by a devised set of rules along with the initial experimental evidence of its feasibility and effectiveness. However, this methodology has to be fully automatically implemented and tested on a larger dataset in future research. Practical implications - This result has implication for semantic search, since a richer ontology, comprised of multiple aspects and viewpoints of the domain of knowledge, enhances discoverability and improves search results. Originality/value - To the best of the knowledge, this is the first study to examine and assess the maximal level of semantic relation-based ontology unification.
    Date
    20. 1.2015 18:30:22
    Source
    Aslib journal of information management. 66(2014) no.5, S.494-518
    Type
    a
  9. Zhang, L.: Linking information through function (2014) 0.01
    0.013611285 = product of:
      0.02722257 = sum of:
        0.012609138 = weight(_text_:a in 1526) [ClassicSimilarity], result of:
          0.012609138 = score(doc=1526,freq=16.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.2161963 = fieldWeight in 1526, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=1526)
        0.014613431 = product of:
          0.029226862 = sum of:
            0.029226862 = weight(_text_:information in 1526) [ClassicSimilarity], result of:
              0.029226862 = score(doc=1526,freq=16.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.3291521 = fieldWeight in 1526, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1526)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    How information resources can be meaningfully related has been addressed in contexts from bibliographic entries to hyperlinks and, more recently, linked data. The genre structure and relationships among genre structure constituents shed new light on organizing information by purpose or function. This study examines the relationships among a set of functional units previously constructed in a taxonomy, each of which is a chunk of information embedded in a document and is distinct in terms of its communicative function. Through a card-sort study, relationships among functional units were identified with regard to their occurrence and function. The findings suggest that a group of functional units can be identified, collocated, and navigated by particular relationships. Understanding how functional units are related to each other is significant in linking information pieces in documents to support finding, aggregating, and navigating information in a distributed information environment.
    Source
    Journal of the Association for Information Science and Technology. 65(2014) no.11, S.2293-2305
    Type
    a
  10. Monireh, E.; Sarker, M.K.; Bianchi, F.; Hitzler, P.; Doran, D.; Xie, N.: Reasoning over RDF knowledge bases using deep learning (2018) 0.01
    0.0122813415 = product of:
      0.024562683 = sum of:
        0.007430006 = weight(_text_:a in 4553) [ClassicSimilarity], result of:
          0.007430006 = score(doc=4553,freq=8.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.12739488 = fieldWeight in 4553, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4553)
        0.017132677 = product of:
          0.034265354 = sum of:
            0.034265354 = weight(_text_:22 in 4553) [ClassicSimilarity], result of:
              0.034265354 = score(doc=4553,freq=2.0), product of:
                0.17712717 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05058132 = queryNorm
                0.19345059 = fieldWeight in 4553, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4553)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Semantic Web knowledge representation standards, and in particular RDF and OWL, often come endowed with a formal semantics which is considered to be of fundamental importance for the field. Reasoning, i.e., the drawing of logical inferences from knowledge expressed in such standards, is traditionally based on logical deductive methods and algorithms which can be proven to be sound and complete and terminating, i.e. correct in a very strong sense. For various reasons, though, in particular the scalability issues arising from the ever increasing amounts of Semantic Web data available and the inability of deductive algorithms to deal with noise in the data, it has been argued that alternative means of reasoning should be investigated which bear high promise for high scalability and better robustness. From this perspective, deductive algorithms can be considered the gold standard regarding correctness against which alternative methods need to be tested. In this paper, we show that it is possible to train a Deep Learning system on RDF knowledge graphs, such that it is able to perform reasoning over new RDF knowledge graphs, with high precision and recall compared to the deductive gold standard.
    Date
    16.11.2018 14:22:01
    Type
    a
  11. Rüther, M.; Fock, J.; Schultz-Krutisch, T.; Bandholtz, T.: Classification and reference vocabulary in linked environment data (2011) 0.01
    0.011374847 = product of:
      0.022749694 = sum of:
        0.015442979 = weight(_text_:a in 4816) [ClassicSimilarity], result of:
          0.015442979 = score(doc=4816,freq=24.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.26478532 = fieldWeight in 4816, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=4816)
        0.0073067155 = product of:
          0.014613431 = sum of:
            0.014613431 = weight(_text_:information in 4816) [ClassicSimilarity], result of:
              0.014613431 = score(doc=4816,freq=4.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.16457605 = fieldWeight in 4816, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4816)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    The Federal Environment Agency (UBA), Germany, has a long tradition in knowledge organization, using a library along with many Web-based information systems. The backbone of this information space is a classification system enhanced by a reference vocabulary which consists of a thesaurus, a gazetteer and a chronicle. Over the years, classification has increasingly been relegated to the background compared with the reference vocabulary indexing and full text search. Bibliographic items are no longer classified directly but tagged with thesaurus terms, with those terms being classified. Since 2010 we have been developing a linked data representation of this knowledge base. While we are linking bibliographic and observation data with the controlled vocabulary in a Resource Desrcription Framework (RDF) representation, the classification may be revisited as a powerful organization system by inference. This also raises questions about the quality and feasibility of an unambiguous classification of thesaurus terms.
    Source
    Classification and ontology: formal approaches and access to knowledge: proceedings of the International UDC Seminar, 19-20 September 2011, The Hague, The Netherlands. Eds.: A. Slavic u. E. Civallero
    Type
    a
  12. San Segundo, R.; Ávila, D.M.: New conceptual structures for the digital environment : from KOS to the semantic interconnection (2012) 0.01
    0.010740098 = product of:
      0.021480195 = sum of:
        0.012869149 = weight(_text_:a in 850) [ClassicSimilarity], result of:
          0.012869149 = score(doc=850,freq=24.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.22065444 = fieldWeight in 850, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=850)
        0.008611047 = product of:
          0.017222093 = sum of:
            0.017222093 = weight(_text_:information in 850) [ClassicSimilarity], result of:
              0.017222093 = score(doc=850,freq=8.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.19395474 = fieldWeight in 850, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=850)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Primitive thinking forms affected the organization of knowledge, and at a later date writing also affected organization. Currently, the web requires new forms of learning and knowledge; with the globalization of information, connectivity and virtuality have a bearing on human thought. Digital thinking is shaping our reality and its organizational form. Natural memory, considered to be a process that requires the structure of natural language and human capabilities, is interwoven with a subject and a conscience; memory preserved through writing required other tools to assist it, and classifications, cataloguing, organization or other KOS were created. The new tool for recovering digital memory is the semantic web. This points to information's future on the Internet and seems to approach the utopia of global, organized information and attempts to give the website greater significance. The Web 3.0 incorporates a proliferation of languages, concepts and tools that are difficult to govern and are created by users. The semantic web seems to be a natural evolution of the participative web in which we find ourselves, and if an effective combination is achieved between the inclusion of semantic content in web pages and the use of artificial intelligence it will be a revolution; semantic codification will be a fact when it is totally automated. Based on this, a collective digital intelligence is being constituted. We find ourselves before intelligent multitudes with broad access to enormous amounts of information. The intelligent multitude emerges when technologies interconnect. In this global interconnection of semantic information an exponential pattern of technological growth can take place.
    Source
    Categories, contexts and relations in knowledge organization: Proceedings of the Twelfth International ISKO Conference 6-9 August 2012, Mysore, India. Eds.: Neelameghan, A. u. K.S. Raghavan
    Type
    a
  13. Eiter, T.; Kaminski, T.; Redl, C.; Schüller, P.; Weinzierl, A.: Answer set programming with external source access (2017) 0.01
    0.010179459 = product of:
      0.020358918 = sum of:
        0.0117478715 = weight(_text_:a in 3938) [ClassicSimilarity], result of:
          0.0117478715 = score(doc=3938,freq=20.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.20142901 = fieldWeight in 3938, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3938)
        0.008611047 = product of:
          0.017222093 = sum of:
            0.017222093 = weight(_text_:information in 3938) [ClassicSimilarity], result of:
              0.017222093 = score(doc=3938,freq=8.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.19395474 = fieldWeight in 3938, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3938)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Access to external information is an important need for Answer Set Programming (ASP), which is a booming declarative problem solving approach these days. External access not only includes data in different formats, but more general also the results of computations, and possibly in a two-way information exchange. Providing such access is a major challenge, and in particular if it should be supported at a generic level, both regarding the semantics and efficient computation. In this article, we consider problem solving with ASP under external information access using the dlvhex system. The latter facilitates this access through special external atoms, which are two-way API style interfaces between the rules of the program and an external source. The dlvhex system has a flexible plugin architecture that allows one to use multiple predefined and user-defined external atoms which can be implemented, e.g., in Python or C++. We consider how to solve problems using the ASP paradigm, and specifically discuss how to use external atoms in this context, illustrated by examples. As a showcase, we demonstrate the development of a hex program for a concrete real-world problem using Semantic Web technologies, and discuss specifics of the implementation process.
    Series
    Lecture Notes in Computer Scienc;10370) (Information Systems and Applications, incl. Internet/Web, and HCI
    Type
    a
  14. Rajabi, E.; Sanchez-Alonso, S.; Sicilia, M.-A.: Analyzing broken links on the web of data : An experiment with DBpedia (2014) 0.01
    0.010077151 = product of:
      0.020154301 = sum of:
        0.011629799 = weight(_text_:a in 1330) [ClassicSimilarity], result of:
          0.011629799 = score(doc=1330,freq=10.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.19940455 = fieldWeight in 1330, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1330)
        0.008524502 = product of:
          0.017049003 = sum of:
            0.017049003 = weight(_text_:information in 1330) [ClassicSimilarity], result of:
              0.017049003 = score(doc=1330,freq=4.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.1920054 = fieldWeight in 1330, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1330)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Linked open data allow interlinking and integrating any kind of data on the web. Links between various data sources play a key role insofar as they allow software applications (e.g., browsers, search engines) to operate over the aggregated data space as if it was a unique local database. In this new data space, where DBpedia, a data set including structured information from Wikipedia, seems to be the central hub, we analyzed and highlighted outgoing links from this hub in an effort to discover broken links. The paper reports on an experiment to examine the causes of broken links and proposes some treatments for solving this problem.
    Source
    Journal of the Association for Information Science and Technology. 65(2014) no.8, S.1721-1727
    Type
    a
  15. Lassalle, E.; Lassalle, E.: Semantic models in information retrieval (2012) 0.01
    0.009889329 = product of:
      0.019778658 = sum of:
        0.012321272 = weight(_text_:a in 97) [ClassicSimilarity], result of:
          0.012321272 = score(doc=97,freq=22.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.21126054 = fieldWeight in 97, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=97)
        0.0074573862 = product of:
          0.0149147725 = sum of:
            0.0149147725 = weight(_text_:information in 97) [ClassicSimilarity], result of:
              0.0149147725 = score(doc=97,freq=6.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.16796975 = fieldWeight in 97, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=97)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Robertson and Spärck Jones pioneered experimental probabilistic models (Binary Independence Model) with both a typology generalizing the Boolean model, a frequency counting to calculate elementary weightings, and their combination into a global probabilistic estimation. However, this model did not consider indexing terms dependencies. An extension to mixture models (e.g., using a 2-Poisson law) made it possible to take into account these dependencies from a macroscopic point of view (BM25), as well as a shallow linguistic processing of co-references. New approaches (language models, for example "bag of words" models, probabilistic dependencies between requests and documents, and consequently Bayesian inference using Dirichlet prior conjugate) furnished new solutions for documents structuring (categorization) and for index smoothing. Presently, in these probabilistic models the main issues have been addressed from a formal point of view only. Thus, linguistic properties are neglected in the indexing language. The authors examine how a linguistic and semantic modeling can be integrated in indexing languages and set up a hybrid model that makes it possible to deal with different information retrieval problems in a unified way.
    Source
    Next generation search engines: advanced models for information retrieval. Eds.: C. Jouis, u.a
    Type
    a
  16. Padmavathi, T.; Krishnamurthy, M.: Semantic Web tools and techniques for knowledge organization : an overview (2017) 0.01
    0.009724372 = product of:
      0.019448744 = sum of:
        0.009008404 = weight(_text_:a in 3618) [ClassicSimilarity], result of:
          0.009008404 = score(doc=3618,freq=6.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.1544581 = fieldWeight in 3618, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3618)
        0.01044034 = product of:
          0.02088068 = sum of:
            0.02088068 = weight(_text_:information in 3618) [ClassicSimilarity], result of:
              0.02088068 = score(doc=3618,freq=6.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.23515764 = fieldWeight in 3618, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3618)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    The enormous amount of information generated every day and spread across the web is diversified in nature far beyond human consumption. To overcome this difficulty, the transformation of current unstructured information into a structured form called a "Semantic Web" was proposed by Tim Berners-Lee in 1989 to enable computers to understand and interpret the information they store. The aim of the semantic web is the integration of heterogeneous and distributed data spread across the web for knowledge discovery. The core of sematic web technologies includes knowledge representation languages RDF and OWL, ontology editors and reasoning tools, and ontology query languages such as SPARQL have also been discussed.
    Type
    a
  17. Kara, S.: ¬An ontology-based retrieval system using semantic indexing (2012) 0.01
    0.009624632 = product of:
      0.019249264 = sum of:
        0.008916007 = weight(_text_:a in 3829) [ClassicSimilarity], result of:
          0.008916007 = score(doc=3829,freq=8.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.15287387 = fieldWeight in 3829, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
        0.010333257 = product of:
          0.020666514 = sum of:
            0.020666514 = weight(_text_:information in 3829) [ClassicSimilarity], result of:
              0.020666514 = score(doc=3829,freq=8.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.23274569 = fieldWeight in 3829, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3829)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    In this thesis, we present an ontology-based information extraction and retrieval system and its application to soccer domain. In general, we deal with three issues in semantic search, namely, usability, scalability and retrieval performance. We propose a keyword-based semantic retrieval approach. The performance of the system is improved considerably using domain-specific information extraction, inference and rules. Scalability is achieved by adapting a semantic indexing approach. The system is implemented using the state-of-the-art technologies in SemanticWeb and its performance is evaluated against traditional systems as well as the query expansion methods. Furthermore, a detailed evaluation is provided to observe the performance gain due to domain-specific information extraction and inference. Finally, we show how we use semantic indexing to solve simple structural ambiguities.
    Source
    Information Systems. 37(2012) no. 4, S.294-305
    Type
    a
  18. Ilik, V.: Distributed person data : using Semantic Web compliant data in subject name headings (2015) 0.01
    0.009559331 = product of:
      0.019118663 = sum of:
        0.010507616 = weight(_text_:a in 2292) [ClassicSimilarity], result of:
          0.010507616 = score(doc=2292,freq=16.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.18016359 = fieldWeight in 2292, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2292)
        0.008611047 = product of:
          0.017222093 = sum of:
            0.017222093 = weight(_text_:information in 2292) [ClassicSimilarity], result of:
              0.017222093 = score(doc=2292,freq=8.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.19395474 = fieldWeight in 2292, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2292)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Providing efficient access to information is a crucial library mission. Subject classification is one of the major pillars that guarantees the accessibility of records in libraries. In this paper we discuss the need to associate person IDs and URIs with subjects when a named person happens to be the subject of the document. This is often the case with biographies, schools of thought in philosophy, politics, art, and literary criticism. Using Semantic Web compliant data in subject name headings enhances the ability to collocate topics about a person. Also, in retrieval, books about a person would be easily linked to works by that same person. In the context of the Semantic Web, it is expected that, as the available information grows, one would be more effective in the task of information retrieval. Information about a person or, as in the case of this paper, about a researcher exist in various databases, which can be discipline specific or publishers' databases, and in such cases they have an assigned identifier. They also exist in institutional directory databases. We argue that these various databases can be leveraged to support improved discoverability and retrieval of research output for individual authors and institutions, as well as works about those authors.
    Source
    Classification and authority control: expanding resource discovery: proceedings of the International UDC Seminar 2015, 29-30 October 2015, Lisbon, Portugal. Eds.: Slavic, A. u. M.I. Cordeiro
    Type
    a
  19. Corcho, O.; Poveda-Villalón, M.; Gómez-Pérez, A.: Ontology engineering in the era of linked data (2015) 0.01
    0.00938377 = product of:
      0.01876754 = sum of:
        0.012739806 = weight(_text_:a in 3293) [ClassicSimilarity], result of:
          0.012739806 = score(doc=3293,freq=12.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.21843673 = fieldWeight in 3293, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3293)
        0.006027733 = product of:
          0.012055466 = sum of:
            0.012055466 = weight(_text_:information in 3293) [ClassicSimilarity], result of:
              0.012055466 = score(doc=3293,freq=2.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.13576832 = fieldWeight in 3293, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3293)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Ontology engineering encompasses the method, tools and techniques used to develop ontologies. Without requiring ontologies, linked data is driving a paradigm shift, bringing benefits and drawbacks to the publishing world. Ontologies may be heavyweight, supporting deep understanding of a domain, or lightweight, suited to simple classification of concepts and more adaptable for linked data. They also vary in domain specificity, usability and reusabilty. Hybrid vocabularies drawing elements from diverse sources often suffer from internally incompatible semantics. To serve linked data purposes, ontology engineering teams require a range of skills in philosophy, computer science, web development, librarianship and domain expertise.
    Footnote
    Contribution to a special section "Linked data and the charm of weak semantics".
    Source
    Bulletin of the Association for Information Science and Technology. 41(2015) no.4, S.13-17
    Type
    a
  20. Djioua, B.; Desclés, J.-P.; Alrahabi, M.: Searching and mining with semantic categories (2012) 0.01
    0.009220011 = product of:
      0.018440021 = sum of:
        0.0098289745 = weight(_text_:a in 99) [ClassicSimilarity], result of:
          0.0098289745 = score(doc=99,freq=14.0), product of:
            0.05832264 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.05058132 = queryNorm
            0.1685276 = fieldWeight in 99, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=99)
        0.008611047 = product of:
          0.017222093 = sum of:
            0.017222093 = weight(_text_:information in 99) [ClassicSimilarity], result of:
              0.017222093 = score(doc=99,freq=8.0), product of:
                0.088794395 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.05058132 = queryNorm
                0.19395474 = fieldWeight in 99, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=99)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    A new model is proposed to retrieve information by building automatically a semantic metatext structure for texts that allow searching and extracting discourse and semantic information according to certain linguistic categorizations. This paper presents approaches for searching and mining full text with semantic categories. The model is built up from two engines: The first one, called EXCOM (Djioua et al., 2006; Alrahabi, 2010), is an automatic system for text annotation, related to discourse and semantic maps, which are specification of general linguistic ontologies founded on the Applicative and Cognitive Grammar. The annotation layer uses a linguistic method called Contextual Exploration, which handles the polysemic values of a term in texts. Several 'semantic maps' underlying 'point of views' for text mining guide this automatic annotation process. The second engine uses semantic annotated texts, produced previously in order to create a semantic inverted index, which is able to retrieve relevant documents for queries associated with discourse and semantic categories such as definition, quotation, causality, relations between concepts, etc. (Djioua & Desclés, 2007). This semantic indexation process builds a metatext layer for textual contents. Some data and linguistic rules sets as well as the general architecture that extend third-party software are expressed as supplementary information.
    Source
    Next generation search engines: advanced models for information retrieval. Eds.: C. Jouis, u.a
    Type
    a

Languages

  • e 71
  • d 21
  • f 1
  • More… Less…

Types