Search (116 results, page 1 of 6)

  • × theme_ss:"Wissensrepräsentation"
  1. Xiong, C.: Knowledge based text representations for information retrieval (2016) 0.32
    0.32317764 = product of:
      0.48476642 = sum of:
        0.03968332 = product of:
          0.11904996 = sum of:
            0.11904996 = weight(_text_:3a in 5820) [ClassicSimilarity], result of:
              0.11904996 = score(doc=5820,freq=2.0), product of:
                0.3177388 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03747799 = queryNorm
                0.3746787 = fieldWeight in 5820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.33333334 = coord(1/3)
        0.16836207 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.16836207 = score(doc=5820,freq=4.0), product of:
            0.3177388 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03747799 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.16836207 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.16836207 = score(doc=5820,freq=4.0), product of:
            0.3177388 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03747799 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.108358964 = weight(_text_:ranking in 5820) [ClassicSimilarity], result of:
          0.108358964 = score(doc=5820,freq=10.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.5345266 = fieldWeight in 5820, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
      0.6666667 = coord(4/6)
    
    Abstract
    The successes of information retrieval (IR) in recent decades were built upon bag-of-words representations. Effective as it is, bag-of-words is only a shallow text understanding; there is a limited amount of information for document ranking in the word space. This dissertation goes beyond words and builds knowledge based text representations, which embed the external and carefully curated information from knowledge bases, and provide richer and structured evidence for more advanced information retrieval systems. This thesis research first builds query representations with entities associated with the query. Entities' descriptions are used by query expansion techniques that enrich the query with explanation terms. Then we present a general framework that represents a query with entities that appear in the query, are retrieved by the query, or frequently show up in the top retrieved documents. A latent space model is developed to jointly learn the connections from query to entities and the ranking of documents, modeling the external evidence from knowledge bases and internal ranking features cooperatively. To further improve the quality of relevant entities, a defining factor of our query representations, we introduce learning to rank to entity search and retrieve better entities from knowledge bases. In the document representation part, this thesis research also moves one step forward with a bag-of-entities model, in which documents are represented by their automatic entity annotations, and the ranking is performed in the entity space.
    This proposal includes plans to improve the quality of relevant entities with a co-learning framework that learns from both entity labels and document labels. We also plan to develop a hybrid ranking system that combines word based and entity based representations together with their uncertainties considered. At last, we plan to enrich the text representations with connections between entities. We propose several ways to infer entity graph representations for texts, and to rank documents using their structure representations. This dissertation overcomes the limitation of word based representations with external and carefully curated information from knowledge bases. We believe this thesis research is a solid start towards the new generation of intelligent, semantic, and structured information retrieval.
    Content
    Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Language and Information Technologies. Vgl.: https%3A%2F%2Fwww.cs.cmu.edu%2F~cx%2Fpapers%2Fknowledge_based_text_representation.pdf&usg=AOvVaw0SaTSvhWLTh__Uz_HtOtl3.
  2. Zeng, Q.; Yu, M.; Yu, W.; Xiong, J.; Shi, Y.; Jiang, M.: Faceted hierarchy : a new graph type to organize scientific concepts and a construction method (2019) 0.21
    0.20833743 = product of:
      0.41667485 = sum of:
        0.05952498 = product of:
          0.17857493 = sum of:
            0.17857493 = weight(_text_:3a in 400) [ClassicSimilarity], result of:
              0.17857493 = score(doc=400,freq=2.0), product of:
                0.3177388 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03747799 = queryNorm
                0.56201804 = fieldWeight in 400, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=400)
          0.33333334 = coord(1/3)
        0.17857493 = weight(_text_:2f in 400) [ClassicSimilarity], result of:
          0.17857493 = score(doc=400,freq=2.0), product of:
            0.3177388 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03747799 = queryNorm
            0.56201804 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
        0.17857493 = weight(_text_:2f in 400) [ClassicSimilarity], result of:
          0.17857493 = score(doc=400,freq=2.0), product of:
            0.3177388 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03747799 = queryNorm
            0.56201804 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
      0.5 = coord(3/6)
    
    Content
    Vgl.: https%3A%2F%2Faclanthology.org%2FD19-5317.pdf&usg=AOvVaw0ZZFyq5wWTtNTvNkrvjlGA.
  3. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.14
    0.13889162 = product of:
      0.27778324 = sum of:
        0.03968332 = product of:
          0.11904996 = sum of:
            0.11904996 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.11904996 = score(doc=701,freq=2.0), product of:
                0.3177388 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03747799 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
        0.11904996 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.11904996 = score(doc=701,freq=2.0), product of:
            0.3177388 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03747799 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.11904996 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.11904996 = score(doc=701,freq=2.0), product of:
            0.3177388 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03747799 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
      0.5 = coord(3/6)
    
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  4. Alvers, M.R.: Semantische wissensbasierte Suche in den Life Sciences am Beispiel von GoPubMed (2010) 0.03
    0.03380426 = product of:
      0.10141277 = sum of:
        0.052953172 = weight(_text_:suchmaschine in 4262) [ClassicSimilarity], result of:
          0.052953172 = score(doc=4262,freq=2.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.2498851 = fieldWeight in 4262, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03125 = fieldNorm(doc=4262)
        0.0484596 = weight(_text_:ranking in 4262) [ClassicSimilarity], result of:
          0.0484596 = score(doc=4262,freq=2.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.23904754 = fieldWeight in 4262, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03125 = fieldNorm(doc=4262)
      0.33333334 = coord(2/6)
    
    Abstract
    Nie zuvor war der Zugriff auf Informationen so einfach und schnell wie heute. Die Suchmaschine Google ist dabei mit einem Marktanteil von 95 Prozent in Deutschland führend. Aber reicht der heutige Status Quo aus? Wir meinen nein - andere meinen ja. Die Verwendung von Stichworten für die Suche ist sehr begrenzt, nicht intelligent und der Algorithmus zum ranking der Suchergebnisse fragwürdig. Wir zeigen neue Wege der semantischen Suche mittels der Verwendung von Hintergrundwissen. Die (semi)automatische Generierung von Ontologien wird ebenfalls als zentraler Bestandteil einer universellen Wissensplattform vorgestellt und gezeigt, wie Anwender mit dieser Technologie signifikant Zeit sparen und deutlich relevantere Informationen finden.
  5. Vallet, D.; Fernández, M.; Castells, P.: ¬An ontology-based information retrieval model (2005) 0.03
    0.027645696 = product of:
      0.082937084 = sum of:
        0.0726894 = weight(_text_:ranking in 4708) [ClassicSimilarity], result of:
          0.0726894 = score(doc=4708,freq=2.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.35857132 = fieldWeight in 4708, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.046875 = fieldNorm(doc=4708)
        0.010247685 = product of:
          0.030743055 = sum of:
            0.030743055 = weight(_text_:29 in 4708) [ClassicSimilarity], result of:
              0.030743055 = score(doc=4708,freq=2.0), product of:
                0.13183585 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03747799 = queryNorm
                0.23319192 = fieldWeight in 4708, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4708)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    Semantic search has been one of the motivations of the Semantic Web since it was envisioned. We propose a model for the exploitation of ontologybased KBs to improve search over large document repositories. Our approach includes an ontology-based scheme for the semi-automatic annotation of documents, and a retrieval system. The retrieval model is based on an adaptation of the classic vector-space model, including an annotation weighting algorithm, and a ranking algorithm. Semantic search is combined with keyword-based search to achieve tolerance to KB incompleteness. Our proposal is illustrated with sample experiments showing improvements with respect to keyword-based search, and providing ground for further research and discussion.
    Source
    The Semantic Web: research and applications ; second European Semantic WebConference, ESWC 2005, Heraklion, Crete, Greece, May 29 - June 1, 2005 ; proceedings. Eds.: A. Gómez-Pérez u. J. Euzenat
  6. Thenmalar, S.; Geetha, T.V.: Enhanced ontology-based indexing and searching (2014) 0.03
    0.026455574 = product of:
      0.07936672 = sum of:
        0.07344268 = weight(_text_:ranking in 1633) [ClassicSimilarity], result of:
          0.07344268 = score(doc=1633,freq=6.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.3622872 = fieldWeight in 1633, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1633)
        0.005924042 = product of:
          0.017772125 = sum of:
            0.017772125 = weight(_text_:22 in 1633) [ClassicSimilarity], result of:
              0.017772125 = score(doc=1633,freq=2.0), product of:
                0.13124153 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03747799 = queryNorm
                0.1354154 = fieldWeight in 1633, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1633)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    Purpose - The purpose of this paper is to improve the conceptual-based search by incorporating structural ontological information such as concepts and relations. Generally, Semantic-based information retrieval aims to identify relevant information based on the meanings of the query terms or on the context of the terms and the performance of semantic information retrieval is carried out through standard measures-precision and recall. Higher precision leads to the (meaningful) relevant documents obtained and lower recall leads to the less coverage of the concepts. Design/methodology/approach - In this paper, the authors enhance the existing ontology-based indexing proposed by Kohler et al., by incorporating sibling information to the index. The index designed by Kohler et al., contains only super and sub-concepts from the ontology. In addition, in our approach, we focus on two tasks; query expansion and ranking of the expanded queries, to improve the efficiency of the ontology-based search. The aforementioned tasks make use of ontological concepts, and relations existing between those concepts so as to obtain semantically more relevant search results for a given query. Findings - The proposed ontology-based indexing technique is investigated by analysing the coverage of concepts that are being populated in the index. Here, we introduce a new measure called index enhancement measure, to estimate the coverage of ontological concepts being indexed. We have evaluated the ontology-based search for the tourism domain with the tourism documents and tourism-specific ontology. The comparison of search results based on the use of ontology "with and without query expansion" is examined to estimate the efficiency of the proposed query expansion task. The ranking is compared with the ORank system to evaluate the performance of our ontology-based search. From these analyses, the ontology-based search results shows better recall when compared to the other concept-based search systems. The mean average precision of the ontology-based search is found to be 0.79 and the recall is found to be 0.65, the ORank system has the mean average precision of 0.62 and the recall is found to be 0.51, while the concept-based search has the mean average precision of 0.56 and the recall is found to be 0.42. Practical implications - When the concept is not present in the domain-specific ontology, the concept cannot be indexed. When the given query term is not available in the ontology then the term-based results are retrieved. Originality/value - In addition to super and sub-concepts, we incorporate the concepts present in same level (siblings) to the ontological index. The structural information from the ontology is determined for the query expansion. The ranking of the documents depends on the type of the query (single concept query, multiple concept queries and concept with relation queries) and the ontological relations that exists in the query and the documents. With this ontological structural information, the search results showed us better coverage of concepts with respect to the query.
    Date
    20. 1.2015 18:30:22
  7. Dirsch-Weigand, A.; Schmidt, I.: ConWeaver : Automatisierte Wissensnetze für die semantische Suche (2006) 0.02
    0.018721774 = product of:
      0.11233064 = sum of:
        0.11233064 = weight(_text_:suchmaschine in 6026) [ClassicSimilarity], result of:
          0.11233064 = score(doc=6026,freq=4.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.53008634 = fieldWeight in 6026, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.046875 = fieldNorm(doc=6026)
      0.16666667 = coord(1/6)
    
    Abstract
    Google ist zum Inbegriff einer Suchmaschine geworden. Doch ist in Fachkreisen klar, dass Volltexsuchtmaschinen wie Google auch deutliche Schwächen aufweisen und deshalb für die effiziente Suche in Fachportalen, Intranets und Enterprise-Content-Management-Systemen oft nicht ausreichen. Weil Volltextsuchmaschinen nur mit dem genauen Wortlaut suchen, finden sie einerseits Informationen nicht, die zwar dem Inhalt, nicht aber den genauen Formulierungen der Suchanfrage entsprechen. Bezeichnungsalternativen, sprachlichen Varianten sowie fremdsprachliche Benennungen werden nicht als bedeutungsgleich erkannt. Andererseits entstehen unpräzise Suchergebnisse, weil gleich lautende Bezeichnungen unterschiedlicher Bedeutung nicht unterschieden werden. Diese Probleme geht die semantische Suchmaschine ConWeaver an, die das Fraunhofer Institut Integrierte Informations- und Publikationssysteme (Fraunhofer IPSI) in Darmstadt entwickelt hat. An Stelle eines Volltextindexes setzt sie ein Wissensnetz als Suchindex ein. Im Unterschied zu den meisten anderen ontologiebasierten Softwareprodukten baut die Software ConWeaver dieses Wissensnetz automatisiert auf.
  8. Lee, J.; Min, J.-K.; Oh, A.; Chung, C.-W.: Effective ranking and search techniques for Web resources considering semantic relationships (2014) 0.02
    0.017486354 = product of:
      0.10491812 = sum of:
        0.10491812 = weight(_text_:ranking in 2670) [ClassicSimilarity], result of:
          0.10491812 = score(doc=2670,freq=6.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.51755315 = fieldWeight in 2670, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2670)
      0.16666667 = coord(1/6)
    
    Abstract
    On the Semantic Web, the types of resources and the semantic relationships between resources are defined in an ontology. By using that information, the accuracy of information retrieval can be improved. In this paper, we present effective ranking and search techniques considering the semantic relationships in an ontology. Our technique retrieves top-k resources which are the most relevant to query keywords through the semantic relationships. To do this, we propose a weighting measure for the semantic relationship. Based on this measure, we propose a novel ranking method which considers the number of meaningful semantic relationships between a resource and keywords as well as the coverage and discriminating power of keywords. In order to improve the efficiency of the search, we prune the unnecessary search space using the length and weight thresholds of the semantic relationship path. In addition, we exploit Threshold Algorithm based on an extended inverted index to answer top-k results efficiently. The experimental results using real data sets demonstrate that our retrieval method using the semantic information generates accurate results efficiently compared to the traditional methods.
  9. Alaya, N.; Yahia, S.B.; Lamolle, M.: Ranking with ties of OWL ontology reasoners based on learned performances (2016) 0.02
    0.017486354 = product of:
      0.10491812 = sum of:
        0.10491812 = weight(_text_:ranking in 3378) [ClassicSimilarity], result of:
          0.10491812 = score(doc=3378,freq=6.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.51755315 = fieldWeight in 3378, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3378)
      0.16666667 = coord(1/6)
    
    Abstract
    Over the last decade, several ontology reasoners have been proposed to overcome the computational complexity of inference tasks on expressive ontology languages such as OWL 2 DL. Nevertheless, it is well-accepted that there is no outstanding reasoner that can outperform in all input ontologies. Thus, deciding the most suitable reasoner for an ontology based application is still a time and effort consuming task. In this paper, we suggest to develop a new system to provide user support when looking for guidance over ontology reasoners. At first, we will be looking at automatically predict a single reasoner empirical performances, in particular its robustness and efficiency, over any given ontology. Later, we aim at ranking a set of candidate reasoners in a most preferred order by taking into account information regarding their predicted performances. We conducted extensive experiments covering over 2500 well selected real-world ontologies and six state-of-the-art of the most performing reasoners. Our primary prediction and ranking results are encouraging and witnessing the potential benefits of our approach.
  10. Baofu, P.: ¬The future of information architecture : conceiving a better way to understand taxonomy, network, and intelligence (2008) 0.02
    0.0156014785 = product of:
      0.09360887 = sum of:
        0.09360887 = weight(_text_:suchmaschine in 2257) [ClassicSimilarity], result of:
          0.09360887 = score(doc=2257,freq=4.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.44173864 = fieldWeight in 2257, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2257)
      0.16666667 = coord(1/6)
    
    RSWK
    Suchmaschine / Information Retrieval
    Subject
    Suchmaschine / Information Retrieval
  11. Widhalm, R.; Mück, T.: Topic maps : Semantische Suche im Internet (2002) 0.02
    0.015286265 = product of:
      0.091717586 = sum of:
        0.091717586 = weight(_text_:suchmaschine in 4731) [ClassicSimilarity], result of:
          0.091717586 = score(doc=4731,freq=6.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.4328137 = fieldWeight in 4731, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03125 = fieldNorm(doc=4731)
      0.16666667 = coord(1/6)
    
    Abstract
    Das Werk behandelt die aktuellen Entwicklungen zur inhaltlichen Erschließung von Informationsquellen im Internet. Topic Maps, semantische Modelle vernetzter Informationsressourcen unter Verwendung von XML bzw. HyTime, bieten alle notwendigen Modellierungskonstrukte, um Dokumente im Internet zu klassifizieren und ein assoziatives, semantisches Netzwerk über diese zu legen. Neben Einführungen in XML, XLink, XPointer sowie HyTime wird anhand von Einsatzszenarien gezeigt, wie diese neuartige Technologie für Content Management und Information Retrieval im Internet funktioniert. Der Entwurf einer Abfragesprache wird ebenso skizziert wie der Prototyp einer intelligenten Suchmaschine. Das Buch zeigt, wie Topic Maps den Weg zu semantisch gesteuerten Suchprozessen im Internet weisen.
    RSWK
    Internet / Navigieren / Suchmaschine / Abfragesprache / Semantisches Netz / ISO-Norm
    Subject
    Internet / Navigieren / Suchmaschine / Abfragesprache / Semantisches Netz / ISO-Norm
  12. Semantic Media Wiki : Autoren sollen Wiki-Inhalte erschließen (2006) 0.01
    0.0133754825 = product of:
      0.08025289 = sum of:
        0.08025289 = weight(_text_:suchmaschine in 6027) [ClassicSimilarity], result of:
          0.08025289 = score(doc=6027,freq=6.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.378712 = fieldWeight in 6027, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.02734375 = fieldNorm(doc=6027)
      0.16666667 = coord(1/6)
    
    Content
    Aus den so festgelegten Beziehungen zwischen den verlinkten Begriffen sollen Computer automatisch sinnvolle Antworten auf komplexere Anfragen generieren können; z.B. eine Liste erzeugen, in der alle Länder von Afrika aufgeführt sind. Die Ländernamen führen als Link zurück zu dem Eintrag, in dem sie stehen - dem Artikel zum Land, für das man sich interessiert. Aus informationswissenschaftlicher Sicht ist das Informationsergebnis, das die neue Technologie produziert, relativ simpel. Aus sozialwissenschaftlicher Sicht steckt darin aber ein riesiges Potential zur Verbesserung der Bereitstellung von enzyklopädischer Information und Wissen für Menschen auf der ganzen Welt. Spannend ist auch die durch Semantic MediaWiki gegebene Möglichkeit der automatischen Zusammenführung von Informationen, die in den verschiedenen Wiki-Einträgen verteilt sind, bei einer hohen Konsistenz der Ergebnisse. Durch die feststehenden Beziehungen zwischen den Links enthält die automatisch erzeugte Liste nach Angaben der Karlsruher Forscher immer die gleichen Daten, egal, von welcher Seite aus man sie abruft. Die Suchmaschine holt sich die Bevölkerungszahl von Ägypten immer vom festgelegten Ägypten-Eintrag, so dass keine unterschiedlichen Zahlen in der Wiki-Landschaft kursieren können. Ein mit Semantic MediaWiki erstellter Testeintrag zu Deutschland kann unter http://ontoworld.org/index.php/Germany eingesehen werden. Die Faktenbox im unteren Teil des Eintrags zeigt an, was der "Eintrag" der Suchmaschine an Wissen über Deutschland anbieten kann. Diese Ergebnisse werden auch in dem Datenbeschreibungsstandard RDF angeboten. Mehr als das, was in der Faktenbox steht, kann der Eintrag nicht an die Suchmaschine abgeben."
  13. Wang, H.; Liu, Q.; Penin, T.; Fu, L.; Zhang, L.; Tran, T.; Yu, Y.; Pan, Y.: Semplore: a scalable IR approach to search the Web of Data (2009) 0.01
    0.0121149 = product of:
      0.0726894 = sum of:
        0.0726894 = weight(_text_:ranking in 1638) [ClassicSimilarity], result of:
          0.0726894 = score(doc=1638,freq=2.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.35857132 = fieldWeight in 1638, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.046875 = fieldNorm(doc=1638)
      0.16666667 = coord(1/6)
    
    Abstract
    The Web of Data keeps growing rapidly. However, the full exploitation of this large amount of structured data faces numerous challenges like usability, scalability, imprecise information needs and data change. We present Semplore, an IR-based system that aims at addressing these issues. Semplore supports intuitive faceted search and complex queries both on text and structured data. It combines imprecise keyword search and precise structured query in a unified ranking scheme. Scalable query processing is supported by leveraging inverted indexes traditionally used in IR systems. This is combined with a novel block-based index structure to support efficient index update when data changes. The experimental results show that Semplore is an efficient and effective system for searching the Web of Data and can be used as a basic infrastructure for Web-scale Semantic Web search engines.
  14. Maheswari, J.U.; Karpagam, G.R.: ¬A conceptual framework for ontology based information retrieval (2010) 0.01
    0.010095751 = product of:
      0.0605745 = sum of:
        0.0605745 = weight(_text_:ranking in 702) [ClassicSimilarity], result of:
          0.0605745 = score(doc=702,freq=2.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.29880944 = fieldWeight in 702, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.0390625 = fieldNorm(doc=702)
      0.16666667 = coord(1/6)
    
    Abstract
    Improving Information retrieval by employing the use of ontologies to overcome the limitations of syntactic search has been one of the inspirations since its emergence. This paper proposes a conceptual framework to exploit ontology based Information retrieval. This framework constitutes of five phases namely Query parsing, word stemming, ontology matching, weight assignment, ranking and Information retrieval. In the first phase, the user query is parsed into sequence of words. The parsed contents are curtailed to identify the significant word by ignoring superfluous terms such as "to", "is","ed", "about" and the like in the stemming phase. The objective of the stemming phase is to throttle feature descriptors to root words, which in turn will increase efficiency; this reduces the time consumed for searching the superfluous terms, which may not significantly influence the effectiveness of the retrieval process. In the third phase ontology matching is carried out by matching the parsed words with the relevant terms in the existing ontology. If the ontology does not exist, it is recommended to generate the required ontology. In the fourth phase the weights are assigned based on the distance between the stemmed words and the terms in the ontology uses improved matchmaking algorithm. The range of weights varies from 0 to 1 based on the level of distance in the ontology (superclass-subclass). The aggregate weights are calculated for the all the combination of stemmed words. The combination with the highest score is ranked as the best and the corresponding information is retrieved. The conceptual workflow is illustrated with an e-governance case study Academic Information System.
  15. Rajasurya, S.; Muralidharan, T.; Devi, S.; Swamynathan, S.: Semantic information retrieval using ontology in university domain (2012) 0.01
    0.010095751 = product of:
      0.0605745 = sum of:
        0.0605745 = weight(_text_:ranking in 2861) [ClassicSimilarity], result of:
          0.0605745 = score(doc=2861,freq=2.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.29880944 = fieldWeight in 2861, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2861)
      0.16666667 = coord(1/6)
    
    Abstract
    Today's conventional search engines hardly do provide the essential content relevant to the user's search query. This is because the context and semantics of the request made by the user is not analyzed to the full extent. So here the need for a semantic web search arises. SWS is upcoming in the area of web search which combines Natural Language Processing and Artificial Intelligence. The objective of the work done here is to design, develop and implement a semantic search engine- SIEU(Semantic Information Extraction in University Domain) confined to the university domain. SIEU uses ontology as a knowledge base for the information retrieval process. It is not just a mere keyword search. It is one layer above what Google or any other search engines retrieve by analyzing just the keywords. Here the query is analyzed both syntactically and semantically. The developed system retrieves the web results more relevant to the user query through keyword expansion. The results obtained here will be accurate enough to satisfy the request made by the user. The level of accuracy will be enhanced since the query is analyzed semantically. The system will be of great use to the developers and researchers who work on web. The Google results are re-ranked and optimized for providing the relevant links. For ranking an algorithm has been applied which fetches more apt results for the user query.
  16. Boteram, F.: Semantische Relationen in Dokumentationssprachen : vom Thesaurus zum semantischen Netz (2010) 0.01
    0.007934572 = product of:
      0.047607433 = sum of:
        0.047607433 = product of:
          0.07141115 = sum of:
            0.035866898 = weight(_text_:29 in 4792) [ClassicSimilarity], result of:
              0.035866898 = score(doc=4792,freq=2.0), product of:
                0.13183585 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03747799 = queryNorm
                0.27205724 = fieldWeight in 4792, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4792)
            0.03554425 = weight(_text_:22 in 4792) [ClassicSimilarity], result of:
              0.03554425 = score(doc=4792,freq=2.0), product of:
                0.13124153 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03747799 = queryNorm
                0.2708308 = fieldWeight in 4792, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4792)
          0.6666667 = coord(2/3)
      0.16666667 = coord(1/6)
    
    Date
    2. 3.2013 12:29:05
    Source
    Wissensspeicher in digitalen Räumen: Nachhaltigkeit - Verfügbarkeit - semantische Interoperabilität. Proceedings der 11. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation, Konstanz, 20. bis 22. Februar 2008. Hrsg.: J. Sieglerschmidt u. H.P.Ohly
  17. Fischer, D.H.: ¬Ein Lehrbeispiel für eine Ontologie : OpenCyc (2004) 0.01
    0.0077223377 = product of:
      0.046334025 = sum of:
        0.046334025 = weight(_text_:suchmaschine in 4568) [ClassicSimilarity], result of:
          0.046334025 = score(doc=4568,freq=2.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.21864946 = fieldWeight in 4568, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4568)
      0.16666667 = coord(1/6)
    
    Content
    Wer über Ontologien und Ontologiesysteme spricht, sollte auch das System OpenCyc kennen. Aber was heißt hier "kennen"? Ich habe mich als Leser und experimentierender Benutzer mit diesem System intensiver befasst und unter einer Reihe von spezielleren Fragen an das System meine Erfahrungen in einigen Arbeitspapieren protokolliert. Sie sind über das Internet zugänglich'. Hier möchte ich der allgemeinen Orientierung über OpenCyc dienende Anmerkungen dazugeben. Bereits eine Recherche mit der Suchmaschine Google, den Suchworten "Cyc OpenCyc" und Beschränkung der Quellen auf Sprache Deutsch oder Herkunft Deutschland erbringt einige der auch hier wiedergegebenen vordergründigen Informationen, sie zeigt aber auch, dass sich Professoren oder Studenten im Jahr 2003 z.B. in Ulm, Heidelberg, Berlin, Bamberg, Bonn und Darmstadt mit dem Thema Cyc und OpenCyc beschäftigt haben entsprechend meiner Titelthese.
  18. Hollink, L.; Assem, M. van: Estimating the relevance of search results in the Culture-Web : a study of semantic distance measures (2010) 0.01
    0.0068010613 = product of:
      0.040806368 = sum of:
        0.040806368 = product of:
          0.061209552 = sum of:
            0.030743055 = weight(_text_:29 in 4649) [ClassicSimilarity], result of:
              0.030743055 = score(doc=4649,freq=2.0), product of:
                0.13183585 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03747799 = queryNorm
                0.23319192 = fieldWeight in 4649, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4649)
            0.030466499 = weight(_text_:22 in 4649) [ClassicSimilarity], result of:
              0.030466499 = score(doc=4649,freq=2.0), product of:
                0.13124153 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03747799 = queryNorm
                0.23214069 = fieldWeight in 4649, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4649)
          0.6666667 = coord(2/3)
      0.16666667 = coord(1/6)
    
    Date
    29. 7.2011 14:44:56
    26.12.2011 13:40:22
  19. Teutsch, K.: ¬Die Welt ist doch eine Scheibe : Google-Herausforderer eyePlorer (2009) 0.01
    0.0055159554 = product of:
      0.033095732 = sum of:
        0.033095732 = weight(_text_:suchmaschine in 2678) [ClassicSimilarity], result of:
          0.033095732 = score(doc=2678,freq=2.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.15617819 = fieldWeight in 2678, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.01953125 = fieldNorm(doc=2678)
      0.16666667 = coord(1/6)
    
    Content
    Wenn die Maschine denkt Zur Hybris des Projekts passt, dass der eyePlorer ursprünglich HAL heißen sollte - wie der außer Rand und Band geratene Bordcomputer aus Kubricks "2001: Odyssee im Weltraum". Wenn man die Buchstaben aber jeweils um eine Alphabetposition nach rechts verrückt, ergibt sich IBM. Was passiert mit unserem Wissen, wenn die Maschine selbst anfängt zu denken? Ralf von Grafenstein macht ein ernstes Gesicht. "Es ist nicht unser Ansinnen, sie alleinzulassen. Es geht bei uns ja nicht nur darum, zu finden, sondern auch mitzumachen. Die Community ist wichtig. Der Dialog ist beiderseitig." Der Lotse soll in Form einer wachsamen Gemeinschaft also an Bord bleiben. Begünstigt wird diese Annahme auch durch die aufkommenden Anfasstechnologien, mit denen das iPhone derzeit so erfolgreich ist: "Allein zehn Prozent der menschlichen Gehirnleistung gehen auf den Pinzettengriff zurück." Martin Hirsch wundert sich, dass diese Erkenntnis von der IT-Branche erst jetzt berücksichtigt wird. Auf berührungssensiblen Bildschirmen sollen die Nutzer mit wenigen Handgriffen bald spielerisch Inhalte schaffen und dem System zur Verfügung stellen. So wird aus der Suchmaschine ein "Sparringspartner" und aus einem Informationsknopf ein "Knowledge Nugget". Wie auch immer man die Erkenntniszutaten des Internetgroßmarkts serviert: Wissen als Zeitwort ist ein länglicher Prozess. Im Moment sei die Maschine noch auf dem Stand eines Zweijährigen, sagen ihre Schöpfer. Sozialisiert werden soll sie demnächst im Internet, ihre Erziehung erfolgt dann durch die Nutzer. Als er Martin Hirsch mit seiner Scheibe zum ersten Mal gesehen habe, dachte Ralf von Grafenstein: "Das ist überfällig! Das wird kommen! Das muss raus!" Jetzt ist es da, klein, unschuldig und unscheinbar. Man findet es bei Google."
  20. Hinkelmann, K.: Ontopia Omnigator : ein Werkzeug zur Einführung in Topic Maps (20xx) 0.00
    0.0039852113 = product of:
      0.023911266 = sum of:
        0.023911266 = product of:
          0.071733795 = sum of:
            0.071733795 = weight(_text_:29 in 3162) [ClassicSimilarity], result of:
              0.071733795 = score(doc=3162,freq=2.0), product of:
                0.13183585 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03747799 = queryNorm
                0.5441145 = fieldWeight in 3162, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3162)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    4. 9.2011 12:29:09

Authors

Years

Languages

  • e 90
  • d 24
  • f 1
  • sp 1
  • More… Less…

Types

  • a 80
  • el 34
  • m 9
  • x 7
  • s 3
  • n 1
  • r 1
  • More… Less…

Subjects