Search (501 results, page 1 of 26)

  • × theme_ss:"Wissensrepräsentation"
  1. Zeng, Q.; Yu, M.; Yu, W.; Xiong, J.; Shi, Y.; Jiang, M.: Faceted hierarchy : a new graph type to organize scientific concepts and a construction method (2019) 0.46
    0.46407837 = product of:
      0.7955629 = sum of:
        0.054483652 = product of:
          0.16345096 = sum of:
            0.16345096 = weight(_text_:3a in 400) [ClassicSimilarity], result of:
              0.16345096 = score(doc=400,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.56201804 = fieldWeight in 400, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=400)
          0.33333334 = coord(1/3)
        0.16345096 = weight(_text_:2f in 400) [ClassicSimilarity], result of:
          0.16345096 = score(doc=400,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
        0.0070079383 = weight(_text_:information in 400) [ClassicSimilarity], result of:
          0.0070079383 = score(doc=400,freq=2.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.116372846 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
        0.08026751 = weight(_text_:extraction in 400) [ClassicSimilarity], result of:
          0.08026751 = score(doc=400,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.39384598 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
        0.16345096 = weight(_text_:2f in 400) [ClassicSimilarity], result of:
          0.16345096 = score(doc=400,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
        0.16345096 = weight(_text_:2f in 400) [ClassicSimilarity], result of:
          0.16345096 = score(doc=400,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
        0.16345096 = weight(_text_:2f in 400) [ClassicSimilarity], result of:
          0.16345096 = score(doc=400,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
      0.5833333 = coord(7/12)
    
    Abstract
    On a scientific concept hierarchy, a parent concept may have a few attributes, each of which has multiple values being a group of child concepts. We call these attributes facets: classification has a few facets such as application (e.g., face recognition), model (e.g., svm, knn), and metric (e.g., precision). In this work, we aim at building faceted concept hierarchies from scientific literature. Hierarchy construction methods heavily rely on hypernym detection, however, the faceted relations are parent-to-child links but the hypernym relation is a multi-hop, i.e., ancestor-to-descendent link with a specific facet "type-of". We use information extraction techniques to find synonyms, sibling concepts, and ancestor-descendent relations from a data science corpus. And we propose a hierarchy growth algorithm to infer the parent-child links from the three types of relationships. It resolves conflicts by maintaining the acyclic structure of a hierarchy.
    Content
    Vgl.: https%3A%2F%2Faclanthology.org%2FD19-5317.pdf&usg=AOvVaw0ZZFyq5wWTtNTvNkrvjlGA.
  2. Xiong, C.: Knowledge based text representations for information retrieval (2016) 0.40
    0.39724264 = product of:
      0.6809874 = sum of:
        0.036322437 = product of:
          0.10896731 = sum of:
            0.10896731 = weight(_text_:3a in 5820) [ClassicSimilarity], result of:
              0.10896731 = score(doc=5820,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.3746787 = fieldWeight in 5820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.33333334 = coord(1/3)
        0.15410304 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.15410304 = score(doc=5820,freq=4.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.013214295 = weight(_text_:information in 5820) [ClassicSimilarity], result of:
          0.013214295 = score(doc=5820,freq=16.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.21943474 = fieldWeight in 5820, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.15410304 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.15410304 = score(doc=5820,freq=4.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.15410304 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.15410304 = score(doc=5820,freq=4.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.015038553 = weight(_text_:system in 5820) [ClassicSimilarity], result of:
          0.015038553 = score(doc=5820,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.13919188 = fieldWeight in 5820, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.15410304 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.15410304 = score(doc=5820,freq=4.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
      0.5833333 = coord(7/12)
    
    Abstract
    The successes of information retrieval (IR) in recent decades were built upon bag-of-words representations. Effective as it is, bag-of-words is only a shallow text understanding; there is a limited amount of information for document ranking in the word space. This dissertation goes beyond words and builds knowledge based text representations, which embed the external and carefully curated information from knowledge bases, and provide richer and structured evidence for more advanced information retrieval systems. This thesis research first builds query representations with entities associated with the query. Entities' descriptions are used by query expansion techniques that enrich the query with explanation terms. Then we present a general framework that represents a query with entities that appear in the query, are retrieved by the query, or frequently show up in the top retrieved documents. A latent space model is developed to jointly learn the connections from query to entities and the ranking of documents, modeling the external evidence from knowledge bases and internal ranking features cooperatively. To further improve the quality of relevant entities, a defining factor of our query representations, we introduce learning to rank to entity search and retrieve better entities from knowledge bases. In the document representation part, this thesis research also moves one step forward with a bag-of-entities model, in which documents are represented by their automatic entity annotations, and the ranking is performed in the entity space.
    This proposal includes plans to improve the quality of relevant entities with a co-learning framework that learns from both entity labels and document labels. We also plan to develop a hybrid ranking system that combines word based and entity based representations together with their uncertainties considered. At last, we plan to enrich the text representations with connections between entities. We propose several ways to infer entity graph representations for texts, and to rank documents using their structure representations. This dissertation overcomes the limitation of word based representations with external and carefully curated information from knowledge bases. We believe this thesis research is a solid start towards the new generation of intelligent, semantic, and structured information retrieval.
    Content
    Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Language and Information Technologies. Vgl.: https%3A%2F%2Fwww.cs.cmu.edu%2F~cx%2Fpapers%2Fknowledge_based_text_representation.pdf&usg=AOvVaw0SaTSvhWLTh__Uz_HtOtl3.
  3. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.36
    0.35672647 = product of:
      0.5350897 = sum of:
        0.036322437 = product of:
          0.10896731 = sum of:
            0.10896731 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.10896731 = score(doc=701,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
        0.022834593 = weight(_text_:web in 701) [ClassicSimilarity], result of:
          0.022834593 = score(doc=701,freq=4.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.2039694 = fieldWeight in 701, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.10896731 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.10896731 = score(doc=701,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.014015876 = weight(_text_:information in 701) [ClassicSimilarity], result of:
          0.014015876 = score(doc=701,freq=18.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.23274568 = fieldWeight in 701, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.10896731 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.10896731 = score(doc=701,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.10896731 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.10896731 = score(doc=701,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.026047539 = weight(_text_:system in 701) [ClassicSimilarity], result of:
          0.026047539 = score(doc=701,freq=6.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24108742 = fieldWeight in 701, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.10896731 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.10896731 = score(doc=701,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
      0.6666667 = coord(8/12)
    
    Abstract
    By the explosion of possibilities for a ubiquitous content production, the information overload problem reaches the level of complexity which cannot be managed by traditional modelling approaches anymore. Due to their pure syntactical nature traditional information retrieval approaches did not succeed in treating content itself (i.e. its meaning, and not its representation). This leads to a very low usefulness of the results of a retrieval process for a user's task at hand. In the last ten years ontologies have been emerged from an interesting conceptualisation paradigm to a very promising (semantic) modelling technology, especially in the context of the Semantic Web. From the information retrieval point of view, ontologies enable a machine-understandable form of content description, such that the retrieval process can be driven by the meaning of the content. However, the very ambiguous nature of the retrieval process in which a user, due to the unfamiliarity with the underlying repository and/or query syntax, just approximates his information need in a query, implies a necessity to include the user in the retrieval process more actively in order to close the gap between the meaning of the content and the meaning of a user's query (i.e. his information need). This thesis lays foundation for such an ontology-based interactive retrieval process, in which the retrieval system interacts with a user in order to conceptually interpret the meaning of his query, whereas the underlying domain ontology drives the conceptualisation process. In that way the retrieval process evolves from a query evaluation process into a highly interactive cooperation between a user and the retrieval system, in which the system tries to anticipate the user's information need and to deliver the relevant content proactively. Moreover, the notion of content relevance for a user's query evolves from a content dependent artefact to the multidimensional context-dependent structure, strongly influenced by the user's preferences. This cooperation process is realized as the so-called Librarian Agent Query Refinement Process. In order to clarify the impact of an ontology on the retrieval process (regarding its complexity and quality), a set of methods and tools for different levels of content and query formalisation is developed, ranging from pure ontology-based inferencing to keyword-based querying in which semantics automatically emerges from the results. Our evaluation studies have shown that the possibilities to conceptualize a user's information need in the right manner and to interpret the retrieval results accordingly are key issues for realizing much more meaningful information retrieval systems.
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
    Theme
    Semantic Web
  4. Semantische Technologien : Grundlagen - Konzepte - Anwendungen (2012) 0.25
    0.25433332 = product of:
      0.436 = sum of:
        0.03996054 = weight(_text_:web in 167) [ClassicSimilarity], result of:
          0.03996054 = score(doc=167,freq=16.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.35694647 = fieldWeight in 167, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.02734375 = fieldNorm(doc=167)
        0.008175928 = weight(_text_:information in 167) [ClassicSimilarity], result of:
          0.008175928 = score(doc=167,freq=8.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.13576832 = fieldWeight in 167, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=167)
        0.06621732 = weight(_text_:extraction in 167) [ClassicSimilarity], result of:
          0.06621732 = score(doc=167,freq=4.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.32490635 = fieldWeight in 167, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.02734375 = fieldNorm(doc=167)
        0.08760671 = weight(_text_:suche in 167) [ClassicSimilarity], result of:
          0.08760671 = score(doc=167,freq=14.0), product of:
            0.17138755 = queryWeight, product of:
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.03430388 = queryNorm
            0.51116145 = fieldWeight in 167, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.02734375 = fieldNorm(doc=167)
        0.14549215 = weight(_text_:wissensbasiertes in 167) [ClassicSimilarity], result of:
          0.14549215 = score(doc=167,freq=4.0), product of:
            0.30209768 = queryWeight, product of:
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.03430388 = queryNorm
            0.48160633 = fieldWeight in 167, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.02734375 = fieldNorm(doc=167)
        0.01860926 = weight(_text_:system in 167) [ClassicSimilarity], result of:
          0.01860926 = score(doc=167,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.17224117 = fieldWeight in 167, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02734375 = fieldNorm(doc=167)
        0.06993809 = product of:
          0.13987619 = sum of:
            0.13987619 = weight(_text_:aufsatzsammlung in 167) [ClassicSimilarity], result of:
              0.13987619 = score(doc=167,freq=12.0), product of:
                0.2250708 = queryWeight, product of:
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.03430388 = queryNorm
                0.62147635 = fieldWeight in 167, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=167)
          0.5 = coord(1/2)
      0.5833333 = coord(7/12)
    
    Abstract
    Dieses Lehrbuch bietet eine umfassende Einführung in Grundlagen, Potentiale und Anwendungen Semantischer Technologien. Es richtet sich an Studierende der Informatik und angrenzender Fächer sowie an Entwickler, die Semantische Technologien am Arbeitsplatz oder in verteilten Applikationen nutzen möchten. Mit seiner an praktischen Beispielen orientierten Darstellung gibt es aber auch Anwendern und Entscheidern in Unternehmen einen breiten Überblick über Nutzen und Möglichkeiten dieser Technologie. Semantische Technologien versetzen Computer in die Lage, Informationen nicht nur zu speichern und wieder zu finden, sondern sie ihrer Bedeutung entsprechend auszuwerten, zu verbinden, zu Neuem zu verknüpfen, und so flexibel und zielgerichtet nützliche Leistungen zu erbringen. Das vorliegende Buch stellt im ersten Teil die als Semantische Technologien bezeichneten Techniken, Sprachen und Repräsentationsformalismen vor. Diese Elemente erlauben es, das in Informationen enthaltene Wissen formal und damit für den Computer verarbeitbar zu beschreiben, Konzepte und Beziehungen darzustellen und schließlich Inhalte zu erfragen, zu erschließen und in Netzen zugänglich zu machen. Der zweite Teil beschreibt, wie mit Semantischen Technologien elementare Funktionen und umfassende Dienste der Informations- und Wissensverarbeitung realisiert werden können. Hierzu gehören etwa die Annotation und das Erschließen von Information, die Suche in den resultierenden Strukturen, das Erklären von Bedeutungszusammenhängen sowie die Integration einzelner Komponenten in komplexe Ablaufprozesse und Anwendungslösungen. Der dritte Teil beschreibt schließlich vielfältige Anwendungsbeispiele in unterschiedlichen Bereichen und illustriert so Mehrwert, Potenzial und Grenzen von Semantischen Technologien. Die dargestellten Systeme reichen von Werkzeugen für persönliches, individuelles Informationsmanagement über Unterstützungsfunktionen für Gruppen bis hin zu neuen Ansätzen im Internet der Dinge und Dienste, einschließlich der Integration verschiedener Medien und Anwendungen von Medizin bis Musik.
    Content
    Inhalt: 1. Einleitung (A. Dengel, A. Bernardi) 2. Wissensrepräsentation (A. Dengel, A. Bernardi, L. van Elst) 3. Semantische Netze, Thesauri und Topic Maps (O. Rostanin, G. Weber) 4. Das Ressource Description Framework (T. Roth-Berghofer) 5. Ontologien und Ontologie-Abgleich in verteilten Informationssystemen (L. van Elst) 6. Anfragesprachen und Reasoning (M. Sintek) 7. Linked Open Data, Semantic Web Datensätze (G.A. Grimnes, O. Hartig, M. Kiesel, M. Liwicki) 8. Semantik in der Informationsextraktion (B. Adrian, B. Endres-Niggemeyer) 9. Semantische Suche (K. Schumacher, B. Forcher, T. Tran) 10. Erklärungsfähigkeit semantischer Systeme (B. Forcher, T. Roth-Berghofer, S. Agne) 11. Semantische Webservices zur Steuerung von Prooduktionsprozessen (M. Loskyll, J. Schlick, S. Hodeck, L. Ollinger, C. Maxeiner) 12. Wissensarbeit am Desktop (S. Schwarz, H. Maus, M. Kiesel, L. Sauermann) 13. Semantische Suche für medizinische Bilder (MEDICO) (M. Möller, M. Sintek) 14. Semantische Musikempfehlungen (S. Baumann, A. Passant) 15. Optimierung von Instandhaltungsprozessen durch Semantische Technologien (P. Stephan, M. Loskyll, C. Stahl, J. Schlick)
    Footnote
    Auch als digitale Ausgabe verfügbar. Auf S. 5 befindet sich der Satz: "Wissen ist Information, die in Aktion umgesetzt wird".
    RSWK
    Wissensrepräsentation / Semantisches Netz / RDF <Informatik> / OWL <Informatik> / Aufsatzsammlung
    Semantic Web / Information Extraction / Suche / Wissensbasiertes System / Aufsatzsammlung
    Semantic Web / Web Services / Semantische Modellierung / Ontologie <Wissensverarbeitung> / Suche / Navigieren / Anwendungsbereich / Aufsatzsammlung
    Subject
    Wissensrepräsentation / Semantisches Netz / RDF <Informatik> / OWL <Informatik> / Aufsatzsammlung
    Semantic Web / Information Extraction / Suche / Wissensbasiertes System / Aufsatzsammlung
    Semantic Web / Web Services / Semantische Modellierung / Ontologie <Wissensverarbeitung> / Suche / Navigieren / Anwendungsbereich / Aufsatzsammlung
    Theme
    Semantic Web
  5. Handbook on ontologies (2004) 0.16
    0.15931618 = product of:
      0.38235885 = sum of:
        0.03495819 = weight(_text_:web in 1952) [ClassicSimilarity], result of:
          0.03495819 = score(doc=1952,freq=6.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.3122631 = fieldWeight in 1952, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1952)
        0.013058522 = weight(_text_:information in 1952) [ClassicSimilarity], result of:
          0.013058522 = score(doc=1952,freq=10.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.21684799 = fieldWeight in 1952, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1952)
        0.20784593 = weight(_text_:wissensbasiertes in 1952) [ClassicSimilarity], result of:
          0.20784593 = score(doc=1952,freq=4.0), product of:
            0.30209768 = queryWeight, product of:
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.03430388 = queryNorm
            0.688009 = fieldWeight in 1952, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1952)
        0.026584659 = weight(_text_:system in 1952) [ClassicSimilarity], result of:
          0.026584659 = score(doc=1952,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24605882 = fieldWeight in 1952, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1952)
        0.099911556 = product of:
          0.19982311 = sum of:
            0.19982311 = weight(_text_:aufsatzsammlung in 1952) [ClassicSimilarity], result of:
              0.19982311 = score(doc=1952,freq=12.0), product of:
                0.2250708 = queryWeight, product of:
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.03430388 = queryNorm
                0.88782334 = fieldWeight in 1952, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1952)
          0.5 = coord(1/2)
      0.41666666 = coord(5/12)
    
    LCSH
    Knowledge representation (Information theory)
    Conceptual structures (Information theory)
    RSWK
    Informationssystem / Wissenstechnik / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Semantic Web / Wissensbasiertes System / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Informationssystem / Konzeptionelle Modellierung / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Series
    International handbook on information systems
    Subject
    Informationssystem / Wissenstechnik / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Semantic Web / Wissensbasiertes System / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Informationssystem / Konzeptionelle Modellierung / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Knowledge representation (Information theory)
    Conceptual structures (Information theory)
    Theme
    Semantic Web
  6. Semantic applications (2018) 0.10
    0.10164173 = product of:
      0.30492517 = sum of:
        0.049438346 = weight(_text_:web in 5204) [ClassicSimilarity], result of:
          0.049438346 = score(doc=5204,freq=12.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.4416067 = fieldWeight in 5204, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
        0.021056235 = weight(_text_:information in 5204) [ClassicSimilarity], result of:
          0.021056235 = score(doc=5204,freq=26.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.34965688 = fieldWeight in 5204, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
        0.20784593 = weight(_text_:wissensbasiertes in 5204) [ClassicSimilarity], result of:
          0.20784593 = score(doc=5204,freq=4.0), product of:
            0.30209768 = queryWeight, product of:
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.03430388 = queryNorm
            0.688009 = fieldWeight in 5204, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
        0.026584659 = weight(_text_:system in 5204) [ClassicSimilarity], result of:
          0.026584659 = score(doc=5204,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24605882 = fieldWeight in 5204, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
      0.33333334 = coord(4/12)
    
    Content
    Introduction.- Ontology Development.- Compliance using Metadata.- Variety Management for Big Data.- Text Mining in Economics.- Generation of Natural Language Texts.- Sentiment Analysis.- Building Concise Text Corpora from Web Contents.- Ontology-Based Modelling of Web Content.- Personalized Clinical Decision Support for Cancer Care.- Applications of Temporal Conceptual Semantic Systems.- Context-Aware Documentation in the Smart Factory.- Knowledge-Based Production Planning for Industry 4.0.- Information Exchange in Jurisdiction.- Supporting Automated License Clearing.- Managing cultural assets: Implementing typical cultural heritage archive's usage scenarios via Semantic Web technologies.- Semantic Applications for Process Management.- Domain-Specific Semantic Search Applications.
    LCSH
    Information storage and retrieval
    Management information systems
    Information Systems Applications (incl. Internet)
    Management of Computing and Information Systems
    Information Storage and Retrieval
    RSWK
    Wissensbasiertes System
    Information Retrieval
    Semantic Web
    Subject
    Wissensbasiertes System
    Information Retrieval
    Semantic Web
    Information storage and retrieval
    Management information systems
    Information Systems Applications (incl. Internet)
    Management of Computing and Information Systems
    Information Storage and Retrieval
    Theme
    Semantic Web
  7. Kara, S.: ¬An ontology-based retrieval system using semantic indexing (2012) 0.07
    0.07412624 = product of:
      0.22237872 = sum of:
        0.024219744 = weight(_text_:web in 3829) [ClassicSimilarity], result of:
          0.024219744 = score(doc=3829,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.21634221 = fieldWeight in 3829, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
        0.014015877 = weight(_text_:information in 3829) [ClassicSimilarity], result of:
          0.014015877 = score(doc=3829,freq=8.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.23274569 = fieldWeight in 3829, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
        0.13902743 = weight(_text_:extraction in 3829) [ClassicSimilarity], result of:
          0.13902743 = score(doc=3829,freq=6.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.68216133 = fieldWeight in 3829, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
        0.04511566 = weight(_text_:system in 3829) [ClassicSimilarity], result of:
          0.04511566 = score(doc=3829,freq=8.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.41757566 = fieldWeight in 3829, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
      0.33333334 = coord(4/12)
    
    Abstract
    In this thesis, we present an ontology-based information extraction and retrieval system and its application to soccer domain. In general, we deal with three issues in semantic search, namely, usability, scalability and retrieval performance. We propose a keyword-based semantic retrieval approach. The performance of the system is improved considerably using domain-specific information extraction, inference and rules. Scalability is achieved by adapting a semantic indexing approach. The system is implemented using the state-of-the-art technologies in SemanticWeb and its performance is evaluated against traditional systems as well as the query expansion methods. Furthermore, a detailed evaluation is provided to observe the performance gain due to domain-specific information extraction and inference. Finally, we show how we use semantic indexing to solve simple structural ambiguities.
    Source
    Information Systems. 37(2012) no. 4, S.294-305
    Theme
    Semantic Web
  8. Schutz, A.; Buitelaar, P.: RelExt: a tool for relation extraction from text in ontology extension (2005) 0.05
    0.052661 = product of:
      0.157983 = sum of:
        0.028543243 = weight(_text_:web in 1078) [ClassicSimilarity], result of:
          0.028543243 = score(doc=1078,freq=4.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.25496176 = fieldWeight in 1078, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1078)
        0.008258934 = weight(_text_:information in 1078) [ClassicSimilarity], result of:
          0.008258934 = score(doc=1078,freq=4.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.13714671 = fieldWeight in 1078, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1078)
        0.09459618 = weight(_text_:extraction in 1078) [ClassicSimilarity], result of:
          0.09459618 = score(doc=1078,freq=4.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.46415195 = fieldWeight in 1078, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1078)
        0.026584659 = weight(_text_:system in 1078) [ClassicSimilarity], result of:
          0.026584659 = score(doc=1078,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24605882 = fieldWeight in 1078, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1078)
      0.33333334 = coord(4/12)
    
    Abstract
    Domain ontologies very rarely model verbs as relations holding between concepts. However, the role of the verb as a central connecting element between concepts is undeniable. Verbs specify the interaction between the participants of some action or event by expressing relations between them. In parallel, it can be argued from an ontology engineering point of view that verbs express a relation between two classes that specify domain and range. The work described here is concerned with relation extraction for ontology extension along these lines. We describe a system (RelExt) that is capable of automatically identifying highly relevant triples (pairs of concepts connected by a relation) over concepts from an existing ontology. RelExt works by extracting relevant verbs and their grammatical arguments (i.e. terms) from a domain-specific text collection and computing corresponding relations through a combination of linguistic and statistical processing. The paper includes a detailed description of the system architecture and evaluation results on a constructed benchmark. RelExt has been developed in the context of the SmartWeb project, which aims at providing intelligent information services via mobile broadband devices on the FIFA World Cup that will be hosted in Germany in 2006. Such services include location based navigational information as well as question answering in the football domain.
    Source
    Semantic Web - ISWC 2005, 4th International Semantic Web Conference, ISWC 2005, Galway, Ireland, November 6-10, 2005, Proceedings. Eds.: Yolanda Gil, Enrico Motta, V. Richard Benjamins, Mark A. Musen
  9. Engels, R.H.P.; Lech, T.Ch.: Generating ontologies for the Semantic Web : OntoBuilder (2004) 0.05
    0.050609346 = product of:
      0.15182804 = sum of:
        0.06850378 = weight(_text_:web in 4404) [ClassicSimilarity], result of:
          0.06850378 = score(doc=4404,freq=36.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.6119082 = fieldWeight in 4404, product of:
              6.0 = tf(freq=36.0), with freq of:
                36.0 = termFreq=36.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=4404)
        0.014774032 = weight(_text_:information in 4404) [ClassicSimilarity], result of:
          0.014774032 = score(doc=4404,freq=20.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.2453355 = fieldWeight in 4404, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=4404)
        0.053511675 = weight(_text_:extraction in 4404) [ClassicSimilarity], result of:
          0.053511675 = score(doc=4404,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.26256397 = fieldWeight in 4404, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03125 = fieldNorm(doc=4404)
        0.015038553 = weight(_text_:system in 4404) [ClassicSimilarity], result of:
          0.015038553 = score(doc=4404,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.13919188 = fieldWeight in 4404, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=4404)
      0.33333334 = coord(4/12)
    
    Abstract
    Significant progress has been made in technologies for publishing and distributing knowledge and information on the web. However, much of the published information is not organized, and it is hard to find answers to questions that require more than a keyword search. In general, one can say that the web is organizing itself. Information is often published in relatively ad hoc fashion. Typically, concern about the presentation of content has been limited to purely layout issues. This, combined with the fact that the representation language used on the World Wide Web (HTML) is mainly format-oriented, makes publishing on the WWW easy, giving it an enormous expressiveness. People add private, educational or organizational content to the web that is of an immensely diverse nature. Content on the web is growing closer to a real universal knowledge base, with one problem relatively undefined; the problem of the interpretation of its contents. Although widely acknowledged for its general and universal advantages, the increasing popularity of the web also shows us some major drawbacks. The developments of the information content on the web during the last year alone, clearly indicates the need for some changes. Perhaps one of the most significant problems with the web as a distributed information system is the difficulty of finding and comparing information.
    Thus, there is a clear need for the web to become more semantic. The aim of introducing semantics into the web is to enhance the precision of search, but also enable the use of logical reasoning on web contents in order to answer queries. The CORPORUM OntoBuilder toolset is developed specifically for this task. It consists of a set of applications that can fulfil a variety of tasks, either as stand-alone tools, or augmenting each other. Important tasks that are dealt with by CORPORUM are related to document and information retrieval (find relevant documents, or support the user finding them), as well as information extraction (building a knowledge base from web documents to answer queries), information dissemination (summarizing strategies and information visualization), and automated document classification strategies. First versions of the toolset are encouraging in that they show large potential as a supportive technology for building up the Semantic Web. In this chapter, methods for transforming the current web into a semantic web are discussed, as well as a technical solution that can perform this task: the CORPORUM tool set. First, the toolset is introduced; followed by some pragmatic issues relating to the approach; then there will be a short overview of the theory in relation to CognIT's vision; and finally, a discussion on some of the applications that arose from the project.
    Source
    Towards the semantic Web: ontology-driven knowledge management. Eds.: J. Davies, u.a
    Theme
    Semantic Web
  10. Beierle, C.; Kern-Isberner, G.: Methoden wissensbasierter Systeme : Grundlagen, Algorithmen, Anwendungen (2008) 0.05
    0.048537906 = product of:
      0.19415163 = sum of:
        0.0066071474 = weight(_text_:information in 4622) [ClassicSimilarity], result of:
          0.0066071474 = score(doc=4622,freq=4.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.10971737 = fieldWeight in 4622, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=4622)
        0.16627675 = weight(_text_:wissensbasiertes in 4622) [ClassicSimilarity], result of:
          0.16627675 = score(doc=4622,freq=4.0), product of:
            0.30209768 = queryWeight, product of:
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.03430388 = queryNorm
            0.55040723 = fieldWeight in 4622, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.03125 = fieldNorm(doc=4622)
        0.021267725 = weight(_text_:system in 4622) [ClassicSimilarity], result of:
          0.021267725 = score(doc=4622,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.19684705 = fieldWeight in 4622, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=4622)
      0.25 = coord(3/12)
    
    LCSH
    Information theory
    RSWK
    Wissensbasiertes System / Lehrbuch
    Subject
    Wissensbasiertes System / Lehrbuch
    Information theory
  11. Rajasurya, S.; Muralidharan, T.; Devi, S.; Swamynathan, S.: Semantic information retrieval using ontology in university domain (2012) 0.05
    0.047985192 = product of:
      0.14395557 = sum of:
        0.04036624 = weight(_text_:web in 2861) [ClassicSimilarity], result of:
          0.04036624 = score(doc=2861,freq=8.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.36057037 = fieldWeight in 2861, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2861)
        0.010115089 = weight(_text_:information in 2861) [ClassicSimilarity], result of:
          0.010115089 = score(doc=2861,freq=6.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.16796975 = fieldWeight in 2861, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2861)
        0.06688959 = weight(_text_:extraction in 2861) [ClassicSimilarity], result of:
          0.06688959 = score(doc=2861,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.32820496 = fieldWeight in 2861, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2861)
        0.026584659 = weight(_text_:system in 2861) [ClassicSimilarity], result of:
          0.026584659 = score(doc=2861,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24605882 = fieldWeight in 2861, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2861)
      0.33333334 = coord(4/12)
    
    Abstract
    Today's conventional search engines hardly do provide the essential content relevant to the user's search query. This is because the context and semantics of the request made by the user is not analyzed to the full extent. So here the need for a semantic web search arises. SWS is upcoming in the area of web search which combines Natural Language Processing and Artificial Intelligence. The objective of the work done here is to design, develop and implement a semantic search engine- SIEU(Semantic Information Extraction in University Domain) confined to the university domain. SIEU uses ontology as a knowledge base for the information retrieval process. It is not just a mere keyword search. It is one layer above what Google or any other search engines retrieve by analyzing just the keywords. Here the query is analyzed both syntactically and semantically. The developed system retrieves the web results more relevant to the user query through keyword expansion. The results obtained here will be accurate enough to satisfy the request made by the user. The level of accuracy will be enhanced since the query is analyzed semantically. The system will be of great use to the developers and researchers who work on web. The Google results are re-ranked and optimized for providing the relevant links. For ranking an algorithm has been applied which fetches more apt results for the user query.
  12. Kiryakov, A.; Popov, B.; Terziev, I.; Manov, D.; Ognyanoff, D.: Semantic annotation, indexing, and retrieval (2004) 0.05
    0.04528837 = product of:
      0.1358651 = sum of:
        0.045669187 = weight(_text_:web in 700) [ClassicSimilarity], result of:
          0.045669187 = score(doc=700,freq=16.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.4079388 = fieldWeight in 700, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=700)
        0.0066071474 = weight(_text_:information in 700) [ClassicSimilarity], result of:
          0.0066071474 = score(doc=700,freq=4.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.10971737 = fieldWeight in 700, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=700)
        0.053511675 = weight(_text_:extraction in 700) [ClassicSimilarity], result of:
          0.053511675 = score(doc=700,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.26256397 = fieldWeight in 700, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03125 = fieldNorm(doc=700)
        0.030077105 = weight(_text_:system in 700) [ClassicSimilarity], result of:
          0.030077105 = score(doc=700,freq=8.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.27838376 = fieldWeight in 700, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=700)
      0.33333334 = coord(4/12)
    
    Abstract
    The Semantic Web realization depends on the availability of a critical mass of metadata for the web content, associated with the respective formal knowledge about the world. We claim that the Semantic Web, at its current stage of development, is in a state of a critical need of metadata generation and usage schemata that are specific, well-defined and easy to understand. This paper introduces our vision for a holistic architecture for semantic annotation, indexing, and retrieval of documents with regard to extensive semantic repositories. A system (called KIM), implementing this concept, is presented in brief and it is used for the purposes of evaluation and demonstration. A particular schema for semantic annotation with respect to real-world entities is proposed. The underlying philosophy is that a practical semantic annotation is impossible without some particular knowledge modelling commitments. Our understanding is that a system for such semantic annotation should be based upon a simple model of real-world entity classes, complemented with extensive instance knowledge. To ensure the efficiency, ease of sharing, and reusability of the metadata, we introduce an upper-level ontology (of about 250 classes and 100 properties), which starts with some basic philosophical distinctions and then goes down to the most common entity types (people, companies, cities, etc.). Thus it encodes many of the domain-independent commonsense concepts and allows straightforward domain-specific extensions. On the basis of the ontology, a large-scale knowledge base of entity descriptions is bootstrapped, and further extended and maintained. Currently, the knowledge bases usually scales between 105 and 106 descriptions. Finally, this paper presents a semantically enhanced information extraction system, which provides automatic semantic annotation with references to classes in the ontology and to instances. The system has been running over a continuously growing document collection (currently about 0.5 million news articles), so it has been under constant testing and evaluation for some time now. On the basis of these semantic annotations, we perform semantic based indexing and retrieval where users can mix traditional information retrieval (IR) queries and ontology-based ones. We argue that such large-scale, fully automatic methods are essential for the transformation of the current largely textual web into a Semantic Web.
    Source
    Web semantics: science, services and agents on the World Wide Web. 2(2004) no.1, S.49-79
    Theme
    Semantic Web
  13. Vlachidis, A.; Tudhope, D.: ¬A knowledge-based approach to information extraction for semantic interoperability in the archaeology domain (2016) 0.05
    0.045011945 = product of:
      0.18004778 = sum of:
        0.011679897 = weight(_text_:information in 2895) [ClassicSimilarity], result of:
          0.011679897 = score(doc=2895,freq=8.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.19395474 = fieldWeight in 2895, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2895)
        0.14956969 = weight(_text_:extraction in 2895) [ClassicSimilarity], result of:
          0.14956969 = score(doc=2895,freq=10.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.7338887 = fieldWeight in 2895, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2895)
        0.018798191 = weight(_text_:system in 2895) [ClassicSimilarity], result of:
          0.018798191 = score(doc=2895,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.17398985 = fieldWeight in 2895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2895)
      0.25 = coord(3/12)
    
    Abstract
    The article presents a method for automatic semantic indexing of archaeological grey-literature reports using empirical (rule-based) Information Extraction techniques in combination with domain-specific knowledge organization systems. The semantic annotation system (OPTIMA) performs the tasks of Named Entity Recognition, Relation Extraction, Negation Detection, and Word-Sense Disambiguation using hand-crafted rules and terminological resources for associating contextual abstractions with classes of the standard ontology CIDOC Conceptual Reference Model (CRM) for cultural heritage and its archaeological extension, CRM-EH. Relation Extraction (RE) performance benefits from a syntactic-based definition of RE patterns derived from domain oriented corpus analysis. The evaluation also shows clear benefit in the use of assistive natural language processing (NLP) modules relating to Word-Sense Disambiguation, Negation Detection, and Noun Phrase Validation, together with controlled thesaurus expansion. The semantic indexing results demonstrate the capacity of rule-based Information Extraction techniques to deliver interoperable semantic abstractions (semantic annotations) with respect to the CIDOC CRM and archaeological thesauri. Major contributions include recognition of relevant entities using shallow parsing NLP techniques driven by a complimentary use of ontological and terminological domain resources and empirical derivation of context-driven RE rules for the recognition of semantic relationships from phrases of unstructured text.
    Source
    Journal of the Association for Information Science and Technology. 67(2016) no.5, S.1138-1152
  14. Ziemba, L.: Information retrieval with concept discovery in digital collections for agriculture and natural resources (2011) 0.04
    0.04029262 = product of:
      0.12087786 = sum of:
        0.016146496 = weight(_text_:web in 4728) [ClassicSimilarity], result of:
          0.016146496 = score(doc=4728,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.14422815 = fieldWeight in 4728, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=4728)
        0.014015876 = weight(_text_:information in 4728) [ClassicSimilarity], result of:
          0.014015876 = score(doc=4728,freq=18.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.23274568 = fieldWeight in 4728, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=4728)
        0.07567694 = weight(_text_:extraction in 4728) [ClassicSimilarity], result of:
          0.07567694 = score(doc=4728,freq=4.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.37132156 = fieldWeight in 4728, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03125 = fieldNorm(doc=4728)
        0.015038553 = weight(_text_:system in 4728) [ClassicSimilarity], result of:
          0.015038553 = score(doc=4728,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.13919188 = fieldWeight in 4728, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=4728)
      0.33333334 = coord(4/12)
    
    Abstract
    The amount and complexity of information available in a digital form is already huge and new information is being produced every day. Retrieving information relevant to address a particular need becomes a significant issue. This work utilizes knowledge organization systems (KOS), such as thesauri and ontologies and applies information extraction (IE) and computational linguistics (CL) techniques to organize, manage and retrieve information stored in digital collections in the agricultural domain. Two real world applications of the approach have been developed and are available and actively used by the public. An ontology is used to manage the Water Conservation Digital Library holding a dynamic collection of various types of digital resources in the domain of urban water conservation in Florida, USA. The ontology based back-end powers a fully operational web interface, available at http://library.conservefloridawater.org. The system has demonstrated numerous benefits of the ontology application, including accurate retrieval of resources, information sharing and reuse, and has proved to effectively facilitate information management. The major difficulty encountered with the approach is that large and dynamic number of concepts makes it difficult to keep the ontology consistent and to accurately catalog resources manually. To address the aforementioned issues, a combination of IE and CL techniques, such as Vector Space Model and probabilistic parsing, with the use of Agricultural Thesaurus were adapted to automatically extract concepts important for each of the texts in the Best Management Practices (BMP) Publication Library--a collection of documents in the domain of agricultural BMPs in Florida available at http://lyra.ifas.ufl.edu/LIB. A new approach of domain-specific concept discovery with the use of Internet search engine was developed. Initial evaluation of the results indicates significant improvement in precision of information extraction. The approach presented in this work focuses on problems unique to agriculture and natural resources domain, such as domain specific concepts and vocabularies, but should be applicable to any collection of texts in digital format. It may be of potential interest for anyone who needs to effectively manage a collection of digital resources.
  15. Jiang, X.; Tan, A.-H.: CRCTOL: a semantic-based domain ontology learning system (2009) 0.04
    0.03983244 = product of:
      0.119497314 = sum of:
        0.02018312 = weight(_text_:web in 3320) [ClassicSimilarity], result of:
          0.02018312 = score(doc=3320,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.18028519 = fieldWeight in 3320, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3320)
        0.0058399485 = weight(_text_:information in 3320) [ClassicSimilarity], result of:
          0.0058399485 = score(doc=3320,freq=2.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.09697737 = fieldWeight in 3320, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3320)
        0.06688959 = weight(_text_:extraction in 3320) [ClassicSimilarity], result of:
          0.06688959 = score(doc=3320,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.32820496 = fieldWeight in 3320, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3320)
        0.026584659 = weight(_text_:system in 3320) [ClassicSimilarity], result of:
          0.026584659 = score(doc=3320,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24605882 = fieldWeight in 3320, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3320)
      0.33333334 = coord(4/12)
    
    Abstract
    Domain ontologies play an important role in supporting knowledge-based applications in the Semantic Web. To facilitate the building of ontologies, text mining techniques have been used to perform ontology learning from texts. However, traditional systems employ shallow natural language processing techniques and focus only on concept and taxonomic relation extraction. In this paper we present a system, known as Concept-Relation-Concept Tuple-based Ontology Learning (CRCTOL), for mining ontologies automatically from domain-specific documents. Specifically, CRCTOL adopts a full text parsing technique and employs a combination of statistical and lexico-syntactic methods, including a statistical algorithm that extracts key concepts from a document collection, a word sense disambiguation algorithm that disambiguates words in the key concepts, a rule-based algorithm that extracts relations between the key concepts, and a modified generalized association rule mining algorithm that prunes unimportant relations for ontology learning. As a result, the ontologies learned by CRCTOL are more concise and contain a richer semantics in terms of the range and number of semantic relations compared with alternative systems. We present two case studies where CRCTOL is used to build a terrorism domain ontology and a sport event domain ontology. At the component level, quantitative evaluation by comparing with Text-To-Onto and its successor Text2Onto has shown that CRCTOL is able to extract concepts and semantic relations with a significantly higher level of accuracy. At the ontology level, the quality of the learned ontologies is evaluated by either employing a set of quantitative and qualitative methods including analyzing the graph structural property, comparison to WordNet, and expert rating, or directly comparing with a human-edited benchmark ontology, demonstrating the high quality of the ontologies learned.
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.1, S.150-168
  16. Conde, A.; Larrañaga, M.; Arruarte, A.; Elorriaga, J.A.; Roth, D.: litewi: a combined term extraction and entity linking method for eliciting educational ontologies from textbooks (2016) 0.04
    0.03887838 = product of:
      0.15551353 = sum of:
        0.010115089 = weight(_text_:information in 2645) [ClassicSimilarity], result of:
          0.010115089 = score(doc=2645,freq=6.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.16796975 = fieldWeight in 2645, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2645)
        0.13377918 = weight(_text_:extraction in 2645) [ClassicSimilarity], result of:
          0.13377918 = score(doc=2645,freq=8.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.6564099 = fieldWeight in 2645, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2645)
        0.011619256 = product of:
          0.023238512 = sum of:
            0.023238512 = weight(_text_:22 in 2645) [ClassicSimilarity], result of:
              0.023238512 = score(doc=2645,freq=2.0), product of:
                0.120126344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03430388 = queryNorm
                0.19345059 = fieldWeight in 2645, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2645)
          0.5 = coord(1/2)
      0.25 = coord(3/12)
    
    Abstract
    Major efforts have been conducted on ontology learning, that is, semiautomatic processes for the construction of domain ontologies from diverse sources of information. In the past few years, a research trend has focused on the construction of educational ontologies, that is, ontologies to be used for educational purposes. The identification of the terminology is crucial to build ontologies. Term extraction techniques allow the identification of the domain-related terms from electronic resources. This paper presents LiTeWi, a novel method that combines current unsupervised term extraction approaches for creating educational ontologies for technology supported learning systems from electronic textbooks. LiTeWi uses Wikipedia as an additional information source. Wikipedia contains more than 30 million articles covering the terminology of nearly every domain in 288 languages, which makes it an appropriate generic corpus for term extraction. Furthermore, given that its content is available in several languages, it promotes both domain and language independence. LiTeWi is aimed at being used by teachers, who usually develop their didactic material from textbooks. To evaluate its performance, LiTeWi was tuned up using a textbook on object oriented programming and then tested with two textbooks of different domains-astronomy and molecular biology.
    Date
    22. 1.2016 12:38:14
    Source
    Journal of the Association for Information Science and Technology. 67(2016) no.2, S.380-399
  17. Urs, S.R.; Angrosh, M.A.: Ontology-based knowledge organization systems in digital libraries : a comparison of experiments in OWL and KAON ontologies (2006 (?)) 0.04
    0.03684422 = product of:
      0.11053266 = sum of:
        0.027966553 = weight(_text_:web in 2799) [ClassicSimilarity], result of:
          0.027966553 = score(doc=2799,freq=6.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.24981049 = fieldWeight in 2799, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=2799)
        0.014015876 = weight(_text_:information in 2799) [ClassicSimilarity], result of:
          0.014015876 = score(doc=2799,freq=18.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.23274568 = fieldWeight in 2799, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=2799)
        0.053511675 = weight(_text_:extraction in 2799) [ClassicSimilarity], result of:
          0.053511675 = score(doc=2799,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.26256397 = fieldWeight in 2799, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03125 = fieldNorm(doc=2799)
        0.015038553 = weight(_text_:system in 2799) [ClassicSimilarity], result of:
          0.015038553 = score(doc=2799,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.13919188 = fieldWeight in 2799, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=2799)
      0.33333334 = coord(4/12)
    
    Abstract
    Grounded on a strong belief that ontologies enhance the performance of information retrieval systems, there has been an upsurge of interest in ontologies. Its importance is identified in diverse research fields such as knowledge engineering, knowledge representation, qualitative modeling, language engineering, database design, information integration, object-oriented analysis, information retrieval and extraction, knowledge management and agent-based systems design (Guarino, 1998). While the role-played by ontologies, automatically lends a place of legitimacy for these tools, research in this area gains greater significance in the wake of various challenges faced in the contemporary digital environment. With the objective of overcoming various pitfalls associated with current search mechanisms, ontologies are increasingly used for developing efficient information retrieval systems. An indicator of research interest in the area of ontology is the Swoogle, a search engine for Semantic Web documents, terms and data found on the Web (Ding, Li et al, 2004). Given the complex nature of the digital content archived in digital libraries, ontologies can be employed for designing efficient forms of information retrieval in digital libraries. Knowledge representation assumes greater significance due to its crucial role in ontology development. These systems aid in developing intelligent information systems, wherein the notion of intelligence implies the ability of the system to find implicit consequences of its explicitly represented knowledge (Baader and Nutt, 2003). Knowledge representation formalisms such as 'Description Logics' are used to obtain explicit knowledge representation of the subject domain. These representations are developed into ontologies, which are used for developing intelligent information systems. Against this backdrop, the paper examines the use of Description Logics for conceptually modeling a chosen domain, which would be utilized for developing domain ontologies. The knowledge representation languages identified for this purpose are Web Ontology Language (OWL) and KArlsruhe ONtology (KAON) language. Drawing upon the various technical constructs in developing ontology-based information systems, the paper explains the working of the prototypes and also presents a comparative study of the two prototypes.
    Theme
    Information Gateway
  18. Reasoning Web : Semantic Interoperability on the Web, 13th International Summer School 2017, London, UK, July 7-11, 2017, Tutorial Lectures (2017) 0.04
    0.035986133 = product of:
      0.14394453 = sum of:
        0.066939846 = weight(_text_:web in 3934) [ClassicSimilarity], result of:
          0.066939846 = score(doc=3934,freq=22.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.59793836 = fieldWeight in 3934, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3934)
        0.010115089 = weight(_text_:information in 3934) [ClassicSimilarity], result of:
          0.010115089 = score(doc=3934,freq=6.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.16796975 = fieldWeight in 3934, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3934)
        0.06688959 = weight(_text_:extraction in 3934) [ClassicSimilarity], result of:
          0.06688959 = score(doc=3934,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.32820496 = fieldWeight in 3934, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3934)
      0.25 = coord(3/12)
    
    Abstract
    This volume contains the lecture notes of the 13th Reasoning Web Summer School, RW 2017, held in London, UK, in July 2017. In 2017, the theme of the school was "Semantic Interoperability on the Web", which encompasses subjects such as data integration, open data management, reasoning over linked data, database to ontology mapping, query answering over ontologies, hybrid reasoning with rules and ontologies, and ontology-based dynamic systems. The papers of this volume focus on these topics and also address foundational reasoning techniques used in answer set programming and ontologies.
    Content
    Neumaier, Sebastian (et al.): Data Integration for Open Data on the Web - Stamou, Giorgos (et al.): Ontological Query Answering over Semantic Data - Calì, Andrea: Ontology Querying: Datalog Strikes Back - Sequeda, Juan F.: Integrating Relational Databases with the Semantic Web: A Reflection - Rousset, Marie-Christine (et al.): Datalog Revisited for Reasoning in Linked Data - Kaminski, Roland (et al.): A Tutorial on Hybrid Answer Set Solving with clingo - Eiter, Thomas (et al.): Answer Set Programming with External Source Access - Lukasiewicz, Thomas: Uncertainty Reasoning for the Semantic Web - Calvanese, Diego (et al.): OBDA for Log Extraction in Process Mining
    LCSH
    Information storage and retrieval
    RSWK
    Ontologie <Wissensverarbeitung> / Semantic Web
    Series
    Lecture Notes in Computer Scienc;10370 )(Information Systems and Applications, incl. Internet/Web, and HCI
    Subject
    Ontologie <Wissensverarbeitung> / Semantic Web
    Information storage and retrieval
    Theme
    Semantic Web
  19. Hüsken, P.: Informationssuche im Semantic Web : Methoden des Information Retrieval für die Wissensrepräsentation (2006) 0.04
    0.035234455 = product of:
      0.14093782 = sum of:
        0.06850378 = weight(_text_:web in 4332) [ClassicSimilarity], result of:
          0.06850378 = score(doc=4332,freq=16.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.6119082 = fieldWeight in 4332, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4332)
        0.015670227 = weight(_text_:information in 4332) [ClassicSimilarity], result of:
          0.015670227 = score(doc=4332,freq=10.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.2602176 = fieldWeight in 4332, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=4332)
        0.056763813 = weight(_text_:suche in 4332) [ClassicSimilarity], result of:
          0.056763813 = score(doc=4332,freq=2.0), product of:
            0.17138755 = queryWeight, product of:
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.03430388 = queryNorm
            0.3312015 = fieldWeight in 4332, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.046875 = fieldNorm(doc=4332)
      0.25 = coord(3/12)
    
    Abstract
    Das Semantic Web bezeichnet ein erweitertes World Wide Web (WWW), das die Bedeutung von präsentierten Inhalten in neuen standardisierten Sprachen wie RDF Schema und OWL modelliert. Diese Arbeit befasst sich mit dem Aspekt des Information Retrieval, d.h. es wird untersucht, in wie weit Methoden der Informationssuche sich auf modelliertes Wissen übertragen lassen. Die kennzeichnenden Merkmale von IR-Systemen wie vage Anfragen sowie die Unterstützung unsicheren Wissens werden im Kontext des Semantic Web behandelt. Im Fokus steht die Suche nach Fakten innerhalb einer Wissensdomäne, die entweder explizit modelliert sind oder implizit durch die Anwendung von Inferenz abgeleitet werden können. Aufbauend auf der an der Universität Duisburg-Essen entwickelten Retrievalmaschine PIRE wird die Anwendung unsicherer Inferenz mit probabilistischer Prädikatenlogik (pDatalog) implementiert.
    Footnote
    Zugl.: Dortmund, Univ., Dipl.-Arb., 2006 u.d.T.: Hüsken, Peter: Information-Retrieval im Semantic-Web.
    RSWK
    Information Retrieval / Semantic Web
    Subject
    Information Retrieval / Semantic Web
    Theme
    Semantic Web
  20. Khoo, S.G.; Na, J.-C.: Semantic relations in information science (2006) 0.03
    0.034841284 = product of:
      0.10452385 = sum of:
        0.012109872 = weight(_text_:web in 1978) [ClassicSimilarity], result of:
          0.012109872 = score(doc=1978,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.108171105 = fieldWeight in 1978, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1978)
        0.01162135 = weight(_text_:information in 1978) [ClassicSimilarity], result of:
          0.01162135 = score(doc=1978,freq=22.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.19298252 = fieldWeight in 1978, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1978)
        0.069513716 = weight(_text_:extraction in 1978) [ClassicSimilarity], result of:
          0.069513716 = score(doc=1978,freq=6.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.34108067 = fieldWeight in 1978, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1978)
        0.011278915 = weight(_text_:system in 1978) [ClassicSimilarity], result of:
          0.011278915 = score(doc=1978,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.104393914 = fieldWeight in 1978, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1978)
      0.33333334 = coord(4/12)
    
    Abstract
    This chapter examines the nature of semantic relations and their main applications in information science. The nature and types of semantic relations are discussed from the perspectives of linguistics and psychology. An overview of the semantic relations used in knowledge structures such as thesauri and ontologies is provided, as well as the main techniques used in the automatic extraction of semantic relations from text. The chapter then reviews the use of semantic relations in information extraction, information retrieval, question-answering, and automatic text summarization applications. Concepts and relations are the foundation of knowledge and thought. When we look at the world, we perceive not a mass of colors but objects to which we automatically assign category labels. Our perceptual system automatically segments the world into concepts and categories. Concepts are the building blocks of knowledge; relations act as the cement that links concepts into knowledge structures. We spend much of our lives identifying regular associations and relations between objects, events, and processes so that the world has an understandable structure and predictability. Our lives and work depend on the accuracy and richness of this knowledge structure and its web of relations. Relations are needed for reasoning and inferencing. Chaffin and Herrmann (1988b, p. 290) noted that "relations between ideas have long been viewed as basic to thought, language, comprehension, and memory." Aristotle's Metaphysics (Aristotle, 1961; McKeon, expounded on several types of relations. The majority of the 30 entries in a section of the Metaphysics known today as the Philosophical Lexicon referred to relations and attributes, including cause, part-whole, same and opposite, quality (i.e., attribute) and kind-of, and defined different types of each relation. Hume (1955) pointed out that there is a connection between successive ideas in our minds, even in our dreams, and that the introduction of an idea in our mind automatically recalls an associated idea. He argued that all the objects of human reasoning are divided into relations of ideas and matters of fact and that factual reasoning is founded on the cause-effect relation. His Treatise of Human Nature identified seven kinds of relations: resemblance, identity, relations of time and place, proportion in quantity or number, degrees in quality, contrariety, and causation. Mill (1974, pp. 989-1004) discoursed on several types of relations, claiming that all things are either feelings, substances, or attributes, and that attributes can be a quality (which belongs to one object) or a relation to other objects.
    Linguists in the structuralist tradition (e.g., Lyons, 1977; Saussure, 1959) have asserted that concepts cannot be defined on their own but only in relation to other concepts. Semantic relations appear to reflect a logical structure in the fundamental nature of thought (Caplan & Herrmann, 1993). Green, Bean, and Myaeng (2002) noted that semantic relations play a critical role in how we represent knowledge psychologically, linguistically, and computationally, and that many systems of knowledge representation start with a basic distinction between entities and relations. Green (2001, p. 3) said that "relationships are involved as we combine simple entities to form more complex entities, as we compare entities, as we group entities, as one entity performs a process on another entity, and so forth. Indeed, many things that we might initially regard as basic and elemental are revealed upon further examination to involve internal structure, or in other words, internal relationships." Concepts and relations are often expressed in language and text. Language is used not just for communicating concepts and relations, but also for representing, storing, and reasoning with concepts and relations. We shall examine the nature of semantic relations from a linguistic and psychological perspective, with an emphasis on relations expressed in text. The usefulness of semantic relations in information science, especially in ontology construction, information extraction, information retrieval, question-answering, and text summarization is discussed. Research and development in information science have focused on concepts and terms, but the focus will increasingly shift to the identification, processing, and management of relations to achieve greater effectiveness and refinement in information science techniques. Previous chapters in ARIST on natural language processing (Chowdhury, 2003), text mining (Trybula, 1999), information retrieval and the philosophy of language (Blair, 2003), and query expansion (Efthimiadis, 1996) provide a background for this discussion, as semantic relations are an important part of these applications.
    Source
    Annual review of information science and technology. 40(2006), S.157-228

Years

Languages

  • e 398
  • d 92
  • pt 3
  • f 1
  • sp 1
  • More… Less…

Types

  • a 350
  • el 132
  • m 36
  • x 29
  • n 15
  • s 13
  • r 7
  • p 4
  • A 1
  • EL 1
  • More… Less…

Subjects

Classifications