Search (499 results, page 1 of 25)

  • × theme_ss:"Wissensrepräsentation"
  1. Zeng, Q.; Yu, M.; Yu, W.; Xiong, J.; Shi, Y.; Jiang, M.: Faceted hierarchy : a new graph type to organize scientific concepts and a construction method (2019) 0.24
    0.23947479 = product of:
      0.4310546 = sum of:
        0.056410737 = product of:
          0.1692322 = sum of:
            0.1692322 = weight(_text_:3a in 400) [ClassicSimilarity], result of:
              0.1692322 = score(doc=400,freq=2.0), product of:
                0.30111524 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.035517205 = queryNorm
                0.56201804 = fieldWeight in 400, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=400)
          0.33333334 = coord(1/3)
        0.1692322 = weight(_text_:2f in 400) [ClassicSimilarity], result of:
          0.1692322 = score(doc=400,freq=2.0), product of:
            0.30111524 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.035517205 = queryNorm
            0.56201804 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
        0.022076517 = weight(_text_:use in 400) [ClassicSimilarity], result of:
          0.022076517 = score(doc=400,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.20298971 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
        0.01410293 = weight(_text_:of in 400) [ClassicSimilarity], result of:
          0.01410293 = score(doc=400,freq=12.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.25392252 = fieldWeight in 400, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
        0.1692322 = weight(_text_:2f in 400) [ClassicSimilarity], result of:
          0.1692322 = score(doc=400,freq=2.0), product of:
            0.30111524 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.035517205 = queryNorm
            0.56201804 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
      0.5555556 = coord(5/9)
    
    Abstract
    On a scientific concept hierarchy, a parent concept may have a few attributes, each of which has multiple values being a group of child concepts. We call these attributes facets: classification has a few facets such as application (e.g., face recognition), model (e.g., svm, knn), and metric (e.g., precision). In this work, we aim at building faceted concept hierarchies from scientific literature. Hierarchy construction methods heavily rely on hypernym detection, however, the faceted relations are parent-to-child links but the hypernym relation is a multi-hop, i.e., ancestor-to-descendent link with a specific facet "type-of". We use information extraction techniques to find synonyms, sibling concepts, and ancestor-descendent relations from a data science corpus. And we propose a hierarchy growth algorithm to infer the parent-child links from the three types of relationships. It resolves conflicts by maintaining the acyclic structure of a hierarchy.
    Content
    Vgl.: https%3A%2F%2Faclanthology.org%2FD19-5317.pdf&usg=AOvVaw0ZZFyq5wWTtNTvNkrvjlGA.
    Source
    Graph-Based Methods for Natural Language Processing - proceedings of the Thirteenth Workshop (TextGraphs-13): November 4, 2019, Hong Kong : EMNLP-IJCNLP 2019. Ed.: Dmitry Ustalov
  2. Xiong, C.: Knowledge based text representations for information retrieval (2016) 0.22
    0.22239183 = product of:
      0.40030527 = sum of:
        0.037607156 = product of:
          0.11282147 = sum of:
            0.11282147 = weight(_text_:3a in 5820) [ClassicSimilarity], result of:
              0.11282147 = score(doc=5820,freq=2.0), product of:
                0.30111524 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.035517205 = queryNorm
                0.3746787 = fieldWeight in 5820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.33333334 = coord(1/3)
        0.028725008 = weight(_text_:retrieval in 5820) [ClassicSimilarity], result of:
          0.028725008 = score(doc=5820,freq=8.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.26736724 = fieldWeight in 5820, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.15955365 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.15955365 = score(doc=5820,freq=4.0), product of:
            0.30111524 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.035517205 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.014865796 = weight(_text_:of in 5820) [ClassicSimilarity], result of:
          0.014865796 = score(doc=5820,freq=30.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.26765788 = fieldWeight in 5820, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.15955365 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.15955365 = score(doc=5820,freq=4.0), product of:
            0.30111524 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.035517205 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
      0.5555556 = coord(5/9)
    
    Abstract
    The successes of information retrieval (IR) in recent decades were built upon bag-of-words representations. Effective as it is, bag-of-words is only a shallow text understanding; there is a limited amount of information for document ranking in the word space. This dissertation goes beyond words and builds knowledge based text representations, which embed the external and carefully curated information from knowledge bases, and provide richer and structured evidence for more advanced information retrieval systems. This thesis research first builds query representations with entities associated with the query. Entities' descriptions are used by query expansion techniques that enrich the query with explanation terms. Then we present a general framework that represents a query with entities that appear in the query, are retrieved by the query, or frequently show up in the top retrieved documents. A latent space model is developed to jointly learn the connections from query to entities and the ranking of documents, modeling the external evidence from knowledge bases and internal ranking features cooperatively. To further improve the quality of relevant entities, a defining factor of our query representations, we introduce learning to rank to entity search and retrieve better entities from knowledge bases. In the document representation part, this thesis research also moves one step forward with a bag-of-entities model, in which documents are represented by their automatic entity annotations, and the ranking is performed in the entity space.
    This proposal includes plans to improve the quality of relevant entities with a co-learning framework that learns from both entity labels and document labels. We also plan to develop a hybrid ranking system that combines word based and entity based representations together with their uncertainties considered. At last, we plan to enrich the text representations with connections between entities. We propose several ways to infer entity graph representations for texts, and to rank documents using their structure representations. This dissertation overcomes the limitation of word based representations with external and carefully curated information from knowledge bases. We believe this thesis research is a solid start towards the new generation of intelligent, semantic, and structured information retrieval.
    Content
    Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Language and Information Technologies. Vgl.: https%3A%2F%2Fwww.cs.cmu.edu%2F~cx%2Fpapers%2Fknowledge_based_text_representation.pdf&usg=AOvVaw0SaTSvhWLTh__Uz_HtOtl3.
    Imprint
    Pittsburgh, PA : Carnegie Mellon University, School of Computer Science, Language Technologies Institute
  3. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.18
    0.184635 = product of:
      0.33234298 = sum of:
        0.037607156 = product of:
          0.11282147 = sum of:
            0.11282147 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.11282147 = score(doc=701,freq=2.0), product of:
                0.30111524 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.035517205 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
        0.053739566 = weight(_text_:retrieval in 701) [ClassicSimilarity], result of:
          0.053739566 = score(doc=701,freq=28.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.5001983 = fieldWeight in 701, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.11282147 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.11282147 = score(doc=701,freq=2.0), product of:
            0.30111524 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.035517205 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.015353328 = weight(_text_:of in 701) [ClassicSimilarity], result of:
          0.015353328 = score(doc=701,freq=32.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.27643585 = fieldWeight in 701, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.11282147 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.11282147 = score(doc=701,freq=2.0), product of:
            0.30111524 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.035517205 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
      0.5555556 = coord(5/9)
    
    Abstract
    By the explosion of possibilities for a ubiquitous content production, the information overload problem reaches the level of complexity which cannot be managed by traditional modelling approaches anymore. Due to their pure syntactical nature traditional information retrieval approaches did not succeed in treating content itself (i.e. its meaning, and not its representation). This leads to a very low usefulness of the results of a retrieval process for a user's task at hand. In the last ten years ontologies have been emerged from an interesting conceptualisation paradigm to a very promising (semantic) modelling technology, especially in the context of the Semantic Web. From the information retrieval point of view, ontologies enable a machine-understandable form of content description, such that the retrieval process can be driven by the meaning of the content. However, the very ambiguous nature of the retrieval process in which a user, due to the unfamiliarity with the underlying repository and/or query syntax, just approximates his information need in a query, implies a necessity to include the user in the retrieval process more actively in order to close the gap between the meaning of the content and the meaning of a user's query (i.e. his information need). This thesis lays foundation for such an ontology-based interactive retrieval process, in which the retrieval system interacts with a user in order to conceptually interpret the meaning of his query, whereas the underlying domain ontology drives the conceptualisation process. In that way the retrieval process evolves from a query evaluation process into a highly interactive cooperation between a user and the retrieval system, in which the system tries to anticipate the user's information need and to deliver the relevant content proactively. Moreover, the notion of content relevance for a user's query evolves from a content dependent artefact to the multidimensional context-dependent structure, strongly influenced by the user's preferences. This cooperation process is realized as the so-called Librarian Agent Query Refinement Process. In order to clarify the impact of an ontology on the retrieval process (regarding its complexity and quality), a set of methods and tools for different levels of content and query formalisation is developed, ranging from pure ontology-based inferencing to keyword-based querying in which semantics automatically emerges from the results. Our evaluation studies have shown that the possibilities to conceptualize a user's information need in the right manner and to interpret the retrieval results accordingly are key issues for realizing much more meaningful information retrieval systems.
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  4. Rindflesch, T.C.; Aronson, A.R.: Semantic processing in information retrieval (1993) 0.05
    0.053515617 = product of:
      0.12041014 = sum of:
        0.056202184 = weight(_text_:retrieval in 4121) [ClassicSimilarity], result of:
          0.056202184 = score(doc=4121,freq=10.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.5231199 = fieldWeight in 4121, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4121)
        0.0364244 = weight(_text_:use in 4121) [ClassicSimilarity], result of:
          0.0364244 = score(doc=4121,freq=4.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.33491597 = fieldWeight in 4121, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4121)
        0.016453419 = weight(_text_:of in 4121) [ClassicSimilarity], result of:
          0.016453419 = score(doc=4121,freq=12.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.29624295 = fieldWeight in 4121, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4121)
        0.011330134 = product of:
          0.0339904 = sum of:
            0.0339904 = weight(_text_:29 in 4121) [ClassicSimilarity], result of:
              0.0339904 = score(doc=4121,freq=2.0), product of:
                0.12493842 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.035517205 = queryNorm
                0.27205724 = fieldWeight in 4121, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4121)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    Intuition suggests that one way to enhance the information retrieval process would be the use of phrases to characterize the contents of text. A number of researchers, however, have noted that phrases alone do not improve retrieval effectiveness. In this paper we briefly review the use of phrases in information retrieval and then suggest extensions to this paradigm using semantic information. We claim that semantic processing, which can be viewed as expressing relations between the concepts represented by phrases, will in fact enhance retrieval effectiveness. The availability of the UMLS® domain model, which we exploit extensively, significantly contributes to the feasibility of this processing.
    Date
    29. 6.2015 14:51:28
  5. Aitken, S.; Reid, S.: Evaluation of an ontology-based information retrieval tool (2000) 0.04
    0.03573197 = product of:
      0.107195914 = sum of:
        0.057450015 = weight(_text_:retrieval in 2862) [ClassicSimilarity], result of:
          0.057450015 = score(doc=2862,freq=8.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.5347345 = fieldWeight in 2862, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=2862)
        0.029435357 = weight(_text_:use in 2862) [ClassicSimilarity], result of:
          0.029435357 = score(doc=2862,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.27065295 = fieldWeight in 2862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0625 = fieldNorm(doc=2862)
        0.020310543 = weight(_text_:of in 2862) [ClassicSimilarity], result of:
          0.020310543 = score(doc=2862,freq=14.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.36569026 = fieldWeight in 2862, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=2862)
      0.33333334 = coord(3/9)
    
    Abstract
    This paper evaluates the use of an explicit domain ontology in an information retrieval tool. The evaluation compares the performance of ontology-enhanced retrieval with keyword retrieval for a fixed set of queries across several data sets. The robustness of the IR approach is assessed by comparing the performance of the tool on the original data set with that on previously unseen data.
    Content
    Beitrag für: Workshop on the Applications of Ontologies and Problem-Solving Methods, (eds) Gómez-Pérez, A., Benjamins, V.R., Guarino, N., and Uschold, M. European Conference on Artificial Intelligence 2000, Berlin.
  6. Kiren, T.: ¬A clustering based indexing technique of modularized ontologies for information retrieval (2017) 0.03
    0.03400495 = product of:
      0.07651114 = sum of:
        0.032115534 = weight(_text_:retrieval in 4399) [ClassicSimilarity], result of:
          0.032115534 = score(doc=4399,freq=10.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.29892567 = fieldWeight in 4399, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=4399)
        0.020813942 = weight(_text_:use in 4399) [ClassicSimilarity], result of:
          0.020813942 = score(doc=4399,freq=4.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.19138055 = fieldWeight in 4399, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.03125 = fieldNorm(doc=4399)
        0.017165542 = weight(_text_:of in 4399) [ClassicSimilarity], result of:
          0.017165542 = score(doc=4399,freq=40.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.3090647 = fieldWeight in 4399, product of:
              6.3245554 = tf(freq=40.0), with freq of:
                40.0 = termFreq=40.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=4399)
        0.0064161215 = product of:
          0.019248364 = sum of:
            0.019248364 = weight(_text_:22 in 4399) [ClassicSimilarity], result of:
              0.019248364 = score(doc=4399,freq=2.0), product of:
                0.1243752 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035517205 = queryNorm
                0.15476047 = fieldWeight in 4399, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4399)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    Indexing plays a vital role in Information Retrieval. With the availability of huge volume of information, it has become necessary to index the information in such a way to make easier for the end users to find the information they want efficiently and accurately. Keyword-based indexing uses words as indexing terms. It is not capable of capturing the implicit relation among terms or the semantics of the words in the document. To eliminate this limitation, ontology-based indexing came into existence, which allows semantic based indexing to solve complex and indirect user queries. Ontologies are used for document indexing which allows semantic based information retrieval. Existing ontologies or the ones constructed from scratch are used presently for indexing. Constructing ontologies from scratch is a labor-intensive task and requires extensive domain knowledge whereas use of an existing ontology may leave some important concepts in documents un-annotated. Using multiple ontologies can overcome the problem of missing out concepts to a great extent, but it is difficult to manage (changes in ontologies over time by their developers) multiple ontologies and ontology heterogeneity also arises due to ontologies constructed by different ontology developers. One possible solution to managing multiple ontologies and build from scratch is to use modular ontologies for indexing.
    Modular ontologies are built in modular manner by combining modules from multiple relevant ontologies. Ontology heterogeneity also arises during modular ontology construction because multiple ontologies are being dealt with, during this process. Ontologies need to be aligned before using them for modular ontology construction. The existing approaches for ontology alignment compare all the concepts of each ontology to be aligned, hence not optimized in terms of time and search space utilization. A new indexing technique is proposed based on modular ontology. An efficient ontology alignment technique is proposed to solve the heterogeneity problem during the construction of modular ontology. Results are satisfactory as Precision and Recall are improved by (8%) and (10%) respectively. The value of Pearsons Correlation Coefficient for degree of similarity, time, search space requirement, precision and recall are close to 1 which shows that the results are significant. Further research can be carried out for using modular ontology based indexing technique for Multimedia Information Retrieval and Bio-Medical information retrieval.
    Content
    Submitted to the Faculty of the Computer Science and Engineering Department of the University of Engineering and Technology Lahore in partial fulfillment of the requirements for the Degree of Doctor of Philosophy in Computer Science (2009 - 009-PhD-CS-04). Vgl.: http://prr.hec.gov.pk/jspui/bitstream/123456789/8375/1/Taybah_Kiren_Computer_Science_HSR_2017_UET_Lahore_14.12.2017.pdf.
    Date
    20. 1.2015 18:30:22
    Imprint
    Lahore : University of Engineering and Technology / Department of Computer Science and Engineering
  7. Tramullas, J.; Garrido-Picazo, P.; Sánchez-Casabón, A.I.: Use of Wikipedia categories on information retrieval research : a brief review (2020) 0.03
    0.033473544 = product of:
      0.100420624 = sum of:
        0.043087512 = weight(_text_:retrieval in 5365) [ClassicSimilarity], result of:
          0.043087512 = score(doc=5365,freq=8.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.40105087 = fieldWeight in 5365, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=5365)
        0.038237654 = weight(_text_:use in 5365) [ClassicSimilarity], result of:
          0.038237654 = score(doc=5365,freq=6.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.35158852 = fieldWeight in 5365, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=5365)
        0.01909546 = weight(_text_:of in 5365) [ClassicSimilarity], result of:
          0.01909546 = score(doc=5365,freq=22.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.34381276 = fieldWeight in 5365, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=5365)
      0.33333334 = coord(3/9)
    
    Abstract
    Wikipedia categories, a classification scheme built for organizing and describing Wikpedia articles, are being applied in computer science research. This paper adopts a systematic literature review approach, in order to identify different approaches and uses of Wikipedia categories in information retrieval research. Several types of work are identified, depending on the intrinsic study of the categories structure, or its use as a tool for the processing and analysis of other documentary corpus different to Wikipedia. Information retrieval is identified as one of the major areas of use, in particular its application in the refinement and improvement of search expressions, and the construction of textual corpus. However, the set of available works shows that in many cases research approaches applied and results obtained can be integrated into a comprehensive and inclusive concept of information retrieval.
  8. Paralic, J.; Kostial, I.: Ontology-based information retrieval (2003) 0.03
    0.03179743 = product of:
      0.09539229 = sum of:
        0.056202184 = weight(_text_:retrieval in 1153) [ClassicSimilarity], result of:
          0.056202184 = score(doc=1153,freq=10.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.5231199 = fieldWeight in 1153, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1153)
        0.025755936 = weight(_text_:use in 1153) [ClassicSimilarity], result of:
          0.025755936 = score(doc=1153,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.23682132 = fieldWeight in 1153, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1153)
        0.013434161 = weight(_text_:of in 1153) [ClassicSimilarity], result of:
          0.013434161 = score(doc=1153,freq=8.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.24188137 = fieldWeight in 1153, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1153)
      0.33333334 = coord(3/9)
    
    Abstract
    In the proposed article a new, ontology-based approach to information retrieval (IR) is presented. The system is based on a domain knowledge representation schema in form of ontology. New resources registered within the system are linked to concepts from this ontology. In such a way resources may be retrieved based on the associations and not only based on partial or exact term matching as the use of vector model presumes In order to evaluate the quality of this retrieval mechanism, experiments to measure retrieval efficiency have been performed with well-known Cystic Fibrosis collection of medical scientific papers. The ontology-based retrieval mechanism has been compared with traditional full text search based on vector IR model as well as with the Latent Semantic Indexing method.
  9. MacFarlane, A.; Missaoui, S.; Frankowska-Takhari, S.: On machine learning and knowledge organization in multimedia information retrieval (2020) 0.03
    0.030985437 = product of:
      0.09295631 = sum of:
        0.03590626 = weight(_text_:retrieval in 5732) [ClassicSimilarity], result of:
          0.03590626 = score(doc=5732,freq=8.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.33420905 = fieldWeight in 5732, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5732)
        0.041137166 = weight(_text_:use in 5732) [ClassicSimilarity], result of:
          0.041137166 = score(doc=5732,freq=10.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.37824902 = fieldWeight in 5732, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5732)
        0.015912883 = weight(_text_:of in 5732) [ClassicSimilarity], result of:
          0.015912883 = score(doc=5732,freq=22.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.28651062 = fieldWeight in 5732, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5732)
      0.33333334 = coord(3/9)
    
    Abstract
    Recent technological developments have increased the use of machine learning to solve many problems, including many in information retrieval. Multimedia information retrieval as a problem represents a significant challenge to machine learning as a technological solution, but some problems can still be addressed by using appropriate AI techniques. We review the technological developments and provide a perspective on the use of machine learning in conjunction with knowledge organization to address multimedia IR needs. The semantic gap in multimedia IR remains a significant problem in the field, and solutions to them are many years off. However, new technological developments allow the use of knowledge organization and machine learning in multimedia search systems and services. Specifically, we argue that, the improvement of detection of some classes of lowlevel features in images music and video can be used in conjunction with knowledge organization to tag or label multimedia content for better retrieval performance. We provide an overview of the use of knowledge organization schemes in machine learning and make recommendations to information professionals on the use of this technology with knowledge organization techniques to solve multimedia IR problems. We introduce a five-step process model that extracts features from multimedia objects (Step 1) from both knowledge organization (Step 1a) and machine learning (Step 1b), merging them together (Step 2) to create an index of those multimedia objects (Step 3). We also overview further steps in creating an application to utilize the multimedia objects (Step 4) and maintaining and updating the database of features on those objects (Step 5).
  10. Jimeno-Yepes, A.; Berlanga Llavori, R.; Rebholz-Schuhmann, D.: Ontology refinement for improved information retrieval (2010) 0.03
    0.030826038 = product of:
      0.09247811 = sum of:
        0.05026876 = weight(_text_:retrieval in 4234) [ClassicSimilarity], result of:
          0.05026876 = score(doc=4234,freq=8.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.46789268 = fieldWeight in 4234, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4234)
        0.025755936 = weight(_text_:use in 4234) [ClassicSimilarity], result of:
          0.025755936 = score(doc=4234,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.23682132 = fieldWeight in 4234, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4234)
        0.016453419 = weight(_text_:of in 4234) [ClassicSimilarity], result of:
          0.016453419 = score(doc=4234,freq=12.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.29624295 = fieldWeight in 4234, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4234)
      0.33333334 = coord(3/9)
    
    Abstract
    Ontologies are frequently used in information retrieval being their main applications the expansion of queries, semantic indexing of documents and the organization of search results. Ontologies provide lexical items, allow conceptual normalization and provide different types of relations. However, the optimization of an ontology to perform information retrieval tasks is still unclear. In this paper, we use an ontology query model to analyze the usefulness of ontologies in effectively performing document searches. Moreover, we propose an algorithm to refine ontologies for information retrieval tasks with preliminary positive results.
  11. Schreiber, G.; Amin, A.; Assem, M. van; Boer, V. de; Hardman, L.; Hildebrand, M.; Hollink, L.; Huang, Z.; Kersen, J. van; Niet, M. de; Omelayenko, B.; Ossenbruggen, J. van; Siebes, R.; Taekema, J.; Wielemaker, J.; Wielinga, B.: MultimediaN E-Culture demonstrator (2006) 0.03
    0.029970998 = product of:
      0.06743474 = sum of:
        0.021543756 = weight(_text_:retrieval in 4648) [ClassicSimilarity], result of:
          0.021543756 = score(doc=4648,freq=2.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.20052543 = fieldWeight in 4648, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=4648)
        0.022076517 = weight(_text_:use in 4648) [ClassicSimilarity], result of:
          0.022076517 = score(doc=4648,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.20298971 = fieldWeight in 4648, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=4648)
        0.01410293 = weight(_text_:of in 4648) [ClassicSimilarity], result of:
          0.01410293 = score(doc=4648,freq=12.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.25392252 = fieldWeight in 4648, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4648)
        0.009711544 = product of:
          0.029134631 = sum of:
            0.029134631 = weight(_text_:29 in 4648) [ClassicSimilarity], result of:
              0.029134631 = score(doc=4648,freq=2.0), product of:
                0.12493842 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.035517205 = queryNorm
                0.23319192 = fieldWeight in 4648, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4648)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    The main objective of the MultimediaN E-Culture project is to demonstrate how novel semantic-web and presentation technologies can be deployed to provide better indexing and search support within large virtual collections of culturalheritage resources. The architecture is fully based on open web standards in particular XML, SVG, RDF/OWL and SPARQL. One basic hypothesis underlying this work is that the use of explicit background knowledge in the form of ontologies/vocabularies/thesauri is in particular useful in information retrieval in knowledge-rich domains. This paper gives some details about the internals of the demonstrator.
    Content
    The online version of the demonstrator can be found at: http://e-culture.multimedian.nl/demo/search.
    Date
    29. 7.2011 14:44:56
  12. Scheir, P.; Pammer, V.; Lindstaedt, S.N.: Information retrieval on the Semantic Web : does it exist? (2007) 0.03
    0.02981962 = product of:
      0.08945886 = sum of:
        0.05026876 = weight(_text_:retrieval in 4329) [ClassicSimilarity], result of:
          0.05026876 = score(doc=4329,freq=8.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.46789268 = fieldWeight in 4329, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4329)
        0.025755936 = weight(_text_:use in 4329) [ClassicSimilarity], result of:
          0.025755936 = score(doc=4329,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.23682132 = fieldWeight in 4329, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4329)
        0.013434161 = weight(_text_:of in 4329) [ClassicSimilarity], result of:
          0.013434161 = score(doc=4329,freq=8.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.24188137 = fieldWeight in 4329, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4329)
      0.33333334 = coord(3/9)
    
    Abstract
    Plenty of contemporary attempts to search exist that are associated with the area of Semantic Web. But which of them qualify as information retrieval for the Semantic Web? Do such approaches exist? To answer these questions we take a look at the nature of the Semantic Web and Semantic Desktop and at definitions for information and data retrieval. We survey current approaches referred to by their authors as information retrieval for the Semantic Web or that use Semantic Web technology for search.
  13. Hollink, L.; Assem, M. van: Estimating the relevance of search results in the Culture-Web : a study of semantic distance measures (2010) 0.03
    0.02966261 = product of:
      0.08898783 = sum of:
        0.031220913 = weight(_text_:use in 4649) [ClassicSimilarity], result of:
          0.031220913 = score(doc=4649,freq=4.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.2870708 = fieldWeight in 4649, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=4649)
        0.01909546 = weight(_text_:of in 4649) [ClassicSimilarity], result of:
          0.01909546 = score(doc=4649,freq=22.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.34381276 = fieldWeight in 4649, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4649)
        0.038671453 = product of:
          0.058007177 = sum of:
            0.029134631 = weight(_text_:29 in 4649) [ClassicSimilarity], result of:
              0.029134631 = score(doc=4649,freq=2.0), product of:
                0.12493842 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.035517205 = queryNorm
                0.23319192 = fieldWeight in 4649, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4649)
            0.028872546 = weight(_text_:22 in 4649) [ClassicSimilarity], result of:
              0.028872546 = score(doc=4649,freq=2.0), product of:
                0.1243752 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035517205 = queryNorm
                0.23214069 = fieldWeight in 4649, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4649)
          0.6666667 = coord(2/3)
      0.33333334 = coord(3/9)
    
    Abstract
    More and more cultural heritage institutions publish their collections, vocabularies and metadata on the Web. The resulting Web of linked cultural data opens up exciting new possibilities for searching and browsing through these cultural heritage collections. We report on ongoing work in which we investigate the estimation of relevance in this Web of Culture. We study existing measures of semantic distance and how they apply to two use cases. The use cases relate to the structured, multilingual and multimodal nature of the Culture Web. We distinguish between measures using the Web, such as Google distance and PMI, and measures using the Linked Data Web, i.e. the semantic structure of metadata vocabularies. We perform a small study in which we compare these semantic distance measures to human judgements of relevance. Although it is too early to draw any definitive conclusions, the study provides new insights into the applicability of semantic distance measures to the Web of Culture, and clear starting points for further research.
    Date
    29. 7.2011 14:44:56
    26.12.2011 13:40:22
  14. Gladun, A.; Rogushina, J.: Development of domain thesaurus as a set of ontology concepts with use of semantic similarity and elements of combinatorial optimization (2021) 0.03
    0.029475702 = product of:
      0.088427104 = sum of:
        0.02513438 = weight(_text_:retrieval in 572) [ClassicSimilarity], result of:
          0.02513438 = score(doc=572,freq=2.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.23394634 = fieldWeight in 572, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=572)
        0.0364244 = weight(_text_:use in 572) [ClassicSimilarity], result of:
          0.0364244 = score(doc=572,freq=4.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.33491597 = fieldWeight in 572, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0546875 = fieldNorm(doc=572)
        0.026868323 = weight(_text_:of in 572) [ClassicSimilarity], result of:
          0.026868323 = score(doc=572,freq=32.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.48376274 = fieldWeight in 572, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=572)
      0.33333334 = coord(3/9)
    
    Abstract
    We consider use of ontological background knowledge in intelligent information systems and analyze directions of their reduction in compliance with specifics of particular user task. Such reduction is aimed at simplification of knowledge processing without loss of significant information. We propose methods of generation of task thesauri based on domain ontology that contain such subset of ontological concepts and relations that can be used in task solving. Combinatorial optimization is used for minimization of task thesaurus. In this approach, semantic similarity estimates are used for determination of concept significance for user task. Some practical examples of optimized thesauri application for semantic retrieval and competence analysis demonstrate efficiency of proposed approach.
  15. Teskey, F.N.: Enriched knowledge representation for information retrieval (1987) 0.03
    0.029429577 = product of:
      0.08828873 = sum of:
        0.043534026 = weight(_text_:retrieval in 698) [ClassicSimilarity], result of:
          0.043534026 = score(doc=698,freq=6.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.40520695 = fieldWeight in 698, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=698)
        0.025755936 = weight(_text_:use in 698) [ClassicSimilarity], result of:
          0.025755936 = score(doc=698,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.23682132 = fieldWeight in 698, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0546875 = fieldNorm(doc=698)
        0.018998774 = weight(_text_:of in 698) [ClassicSimilarity], result of:
          0.018998774 = score(doc=698,freq=16.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.34207192 = fieldWeight in 698, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=698)
      0.33333334 = coord(3/9)
    
    Abstract
    In this paper we identify the need for a new theory of information. An information model is developed which distinguishes between data, as directly observable facts, information, as structured collections of data, and knowledge as methods of using information. The model is intended to support a wide range of information systems. In the paper we develop the use of the model for a semantic information retrieval system using the concept of semantic categories. The likely benefits of this area discussed, though as yet no detailed evaluation has been conducted.
    Source
    SIGIR'87: Proceedings of the 10th annual international ACM SIGIR conference on Research and development in information retrieval
  16. Baião Salgado Silva, G.; Lima, G.Â. Borém de Oliveira: Using topic maps in establishing compatibility of semantically structured hypertext contents (2012) 0.03
    0.029015115 = product of:
      0.06528401 = sum of:
        0.01795313 = weight(_text_:retrieval in 633) [ClassicSimilarity], result of:
          0.01795313 = score(doc=633,freq=2.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.16710453 = fieldWeight in 633, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=633)
        0.018397098 = weight(_text_:use in 633) [ClassicSimilarity], result of:
          0.018397098 = score(doc=633,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.1691581 = fieldWeight in 633, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0390625 = fieldNorm(doc=633)
        0.020913625 = weight(_text_:of in 633) [ClassicSimilarity], result of:
          0.020913625 = score(doc=633,freq=38.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.37654874 = fieldWeight in 633, product of:
              6.164414 = tf(freq=38.0), with freq of:
                38.0 = termFreq=38.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=633)
        0.008020152 = product of:
          0.024060456 = sum of:
            0.024060456 = weight(_text_:22 in 633) [ClassicSimilarity], result of:
              0.024060456 = score(doc=633,freq=2.0), product of:
                0.1243752 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035517205 = queryNorm
                0.19345059 = fieldWeight in 633, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=633)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    Considering the characteristics of hypertext systems and problems such as cognitive overload and the disorientation of users, this project studies subject hypertext documents that have undergone conceptual structuring using facets for content representation and improvement of information retrieval during navigation. The main objective was to assess the possibility of the application of topic map technology for automating the compatibilization process of these structures. For this purpose, two dissertations from the UFMG Information Science Post-Graduation Program were adopted as samples. Both dissertations had been duly analyzed and structured on the MHTX (Hypertextual Map) prototype database. The faceted structures of both dissertations, which had been represented in conceptual maps, were then converted into topic maps. It was then possible to use the merge property of the topic maps to promote the semantic interrelationship between the maps and, consequently, between the hypertextual information resources proper. The merge results were then analyzed in the light of theories dealing with the compatibilization of languages developed within the realm of information technology and librarianship from the 1960s on. The main goals accomplished were: (a) the detailed conceptualization of the merge process of the topic maps, considering the possible compatibilization levels and the applicability of this technology in the integration of faceted structures; and (b) the production of a detailed sequence of steps that may be used in the implementation of topic maps based on faceted structures.
    Date
    22. 2.2013 11:39:23
  17. Assem, M. van; Malaisé, V.; Miles, A.; Schreiber, G.: ¬A method to convert thesauri to SKOS (2006) 0.03
    0.028135154 = product of:
      0.0633041 = sum of:
        0.021543756 = weight(_text_:retrieval in 4642) [ClassicSimilarity], result of:
          0.021543756 = score(doc=4642,freq=2.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.20052543 = fieldWeight in 4642, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=4642)
        0.022076517 = weight(_text_:use in 4642) [ClassicSimilarity], result of:
          0.022076517 = score(doc=4642,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.20298971 = fieldWeight in 4642, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=4642)
        0.009972278 = weight(_text_:of in 4642) [ClassicSimilarity], result of:
          0.009972278 = score(doc=4642,freq=6.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.17955035 = fieldWeight in 4642, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4642)
        0.009711544 = product of:
          0.029134631 = sum of:
            0.029134631 = weight(_text_:29 in 4642) [ClassicSimilarity], result of:
              0.029134631 = score(doc=4642,freq=2.0), product of:
                0.12493842 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.035517205 = queryNorm
                0.23319192 = fieldWeight in 4642, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4642)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    Thesauri can be useful resources for indexing and retrieval on the Semantic Web, but often they are not published in RDF/OWL. To convert thesauri to RDF for use in Semantic Web applications and to ensure the quality and utility of the conversion a structured method is required. Moreover, if different thesauri are to be interoperable without complicated mappings, a standard schema for thesauri is required. This paper presents a method for conversion of thesauri to the SKOS RDF/OWL schema, which is a proposal for such a standard under development by W3Cs Semantic Web Best Practices Working Group. We apply the method to three thesauri: IPSV, GTAA and MeSH. With these case studies we evaluate our method and the applicability of SKOS for representing thesauri.
    Date
    29. 7.2011 14:44:56
  18. Semantic applications (2018) 0.03
    0.027854664 = product of:
      0.08356399 = sum of:
        0.043976005 = weight(_text_:retrieval in 5204) [ClassicSimilarity], result of:
          0.043976005 = score(doc=5204,freq=12.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.40932083 = fieldWeight in 5204, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
        0.026017427 = weight(_text_:use in 5204) [ClassicSimilarity], result of:
          0.026017427 = score(doc=5204,freq=4.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.23922569 = fieldWeight in 5204, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
        0.013570553 = weight(_text_:of in 5204) [ClassicSimilarity], result of:
          0.013570553 = score(doc=5204,freq=16.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.24433708 = fieldWeight in 5204, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
      0.33333334 = coord(3/9)
    
    Abstract
    This book describes proven methodologies for developing semantic applications: software applications which explicitly or implicitly uses the semantics (i.e., the meaning) of a domain terminology in order to improve usability, correctness, and completeness. An example is semantic search, where synonyms and related terms are used for enriching the results of a simple text-based search. Ontologies, thesauri or controlled vocabularies are the centerpiece of semantic applications. The book includes technological and architectural best practices for corporate use.
    Content
    Introduction.- Ontology Development.- Compliance using Metadata.- Variety Management for Big Data.- Text Mining in Economics.- Generation of Natural Language Texts.- Sentiment Analysis.- Building Concise Text Corpora from Web Contents.- Ontology-Based Modelling of Web Content.- Personalized Clinical Decision Support for Cancer Care.- Applications of Temporal Conceptual Semantic Systems.- Context-Aware Documentation in the Smart Factory.- Knowledge-Based Production Planning for Industry 4.0.- Information Exchange in Jurisdiction.- Supporting Automated License Clearing.- Managing cultural assets: Implementing typical cultural heritage archive's usage scenarios via Semantic Web technologies.- Semantic Applications for Process Management.- Domain-Specific Semantic Search Applications.
    LCSH
    Information storage and retrieval
    Management of Computing and Information Systems
    Information Storage and Retrieval
    RSWK
    Information Retrieval
    Series
    methodology, technology, corporate use
    Subject
    Information Retrieval
    Information storage and retrieval
    Management of Computing and Information Systems
    Information Storage and Retrieval
  19. Thenmalar, S.; Geetha, T.V.: Enhanced ontology-based indexing and searching (2014) 0.03
    0.027104037 = product of:
      0.060984083 = sum of:
        0.021767013 = weight(_text_:retrieval in 1633) [ClassicSimilarity], result of:
          0.021767013 = score(doc=1633,freq=6.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.20260347 = fieldWeight in 1633, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1633)
        0.0182122 = weight(_text_:use in 1633) [ClassicSimilarity], result of:
          0.0182122 = score(doc=1633,freq=4.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.16745798 = fieldWeight in 1633, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1633)
        0.015390764 = weight(_text_:of in 1633) [ClassicSimilarity], result of:
          0.015390764 = score(doc=1633,freq=42.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.2771099 = fieldWeight in 1633, product of:
              6.4807405 = tf(freq=42.0), with freq of:
                42.0 = termFreq=42.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1633)
        0.0056141065 = product of:
          0.016842319 = sum of:
            0.016842319 = weight(_text_:22 in 1633) [ClassicSimilarity], result of:
              0.016842319 = score(doc=1633,freq=2.0), product of:
                0.1243752 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035517205 = queryNorm
                0.1354154 = fieldWeight in 1633, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1633)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    Purpose - The purpose of this paper is to improve the conceptual-based search by incorporating structural ontological information such as concepts and relations. Generally, Semantic-based information retrieval aims to identify relevant information based on the meanings of the query terms or on the context of the terms and the performance of semantic information retrieval is carried out through standard measures-precision and recall. Higher precision leads to the (meaningful) relevant documents obtained and lower recall leads to the less coverage of the concepts. Design/methodology/approach - In this paper, the authors enhance the existing ontology-based indexing proposed by Kohler et al., by incorporating sibling information to the index. The index designed by Kohler et al., contains only super and sub-concepts from the ontology. In addition, in our approach, we focus on two tasks; query expansion and ranking of the expanded queries, to improve the efficiency of the ontology-based search. The aforementioned tasks make use of ontological concepts, and relations existing between those concepts so as to obtain semantically more relevant search results for a given query. Findings - The proposed ontology-based indexing technique is investigated by analysing the coverage of concepts that are being populated in the index. Here, we introduce a new measure called index enhancement measure, to estimate the coverage of ontological concepts being indexed. We have evaluated the ontology-based search for the tourism domain with the tourism documents and tourism-specific ontology. The comparison of search results based on the use of ontology "with and without query expansion" is examined to estimate the efficiency of the proposed query expansion task. The ranking is compared with the ORank system to evaluate the performance of our ontology-based search. From these analyses, the ontology-based search results shows better recall when compared to the other concept-based search systems. The mean average precision of the ontology-based search is found to be 0.79 and the recall is found to be 0.65, the ORank system has the mean average precision of 0.62 and the recall is found to be 0.51, while the concept-based search has the mean average precision of 0.56 and the recall is found to be 0.42. Practical implications - When the concept is not present in the domain-specific ontology, the concept cannot be indexed. When the given query term is not available in the ontology then the term-based results are retrieved. Originality/value - In addition to super and sub-concepts, we incorporate the concepts present in same level (siblings) to the ontological index. The structural information from the ontology is determined for the query expansion. The ranking of the documents depends on the type of the query (single concept query, multiple concept queries and concept with relation queries) and the ontological relations that exists in the query and the documents. With this ontological structural information, the search results showed us better coverage of concepts with respect to the query.
    Date
    20. 1.2015 18:30:22
    Source
    Aslib journal of information management. 66(2014) no.6, S.678-696
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  20. Styltsvig, H.B.: Ontology-based information retrieval (2006) 0.03
    0.026876299 = product of:
      0.080628894 = sum of:
        0.040623292 = weight(_text_:retrieval in 1154) [ClassicSimilarity], result of:
          0.040623292 = score(doc=1154,freq=16.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.37811437 = fieldWeight in 1154, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=1154)
        0.020813942 = weight(_text_:use in 1154) [ClassicSimilarity], result of:
          0.020813942 = score(doc=1154,freq=4.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.19138055 = fieldWeight in 1154, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.03125 = fieldNorm(doc=1154)
        0.01919166 = weight(_text_:of in 1154) [ClassicSimilarity], result of:
          0.01919166 = score(doc=1154,freq=50.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.34554482 = fieldWeight in 1154, product of:
              7.071068 = tf(freq=50.0), with freq of:
                50.0 = termFreq=50.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=1154)
      0.33333334 = coord(3/9)
    
    Abstract
    In this thesis, we will present methods for introducing ontologies in information retrieval. The main hypothesis is that the inclusion of conceptual knowledge such as ontologies in the information retrieval process can contribute to the solution of major problems currently found in information retrieval. This utilization of ontologies has a number of challenges. Our focus is on the use of similarity measures derived from the knowledge about relations between concepts in ontologies, the recognition of semantic information in texts and the mapping of this knowledge into the ontologies in use, as well as how to fuse together the ideas of ontological similarity and ontological indexing into a realistic information retrieval scenario. To achieve the recognition of semantic knowledge in a text, shallow natural language processing is used during indexing that reveals knowledge to the level of noun phrases. Furthermore, we briefly cover the identification of semantic relations inside and between noun phrases, as well as discuss which kind of problems are caused by an increase in compoundness with respect to the structure of concepts in the evaluation of queries. Measuring similarity between concepts based on distances in the structure of the ontology is discussed. In addition, a shared nodes measure is introduced and, based on a set of intuitive similarity properties, compared to a number of different measures. In this comparison the shared nodes measure appears to be superior, though more computationally complex. Some of the major problems of shared nodes which relate to the way relations differ with respect to the degree they bring the concepts they connect closer are discussed. A generalized measure called weighted shared nodes is introduced to deal with these problems. Finally, the utilization of concept similarity in query evaluation is discussed. A semantic expansion approach that incorporates concept similarity is introduced and a generalized fuzzy set retrieval model that applies expansion during query evaluation is presented. While not commonly used in present information retrieval systems, it appears that the fuzzy set model comprises the flexibility needed when generalizing to an ontology-based retrieval model and, with the introduction of a hierarchical fuzzy aggregation principle, compound concepts can be handled in a straightforward and natural manner.
    Content
    A dissertation Presented to the Faculties of Roskilde University in Partial Fulfillment of the Requirement for the Degree of Doctor of Philosophy. Vgl. unter: http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.117.987 oder http://coitweb.uncc.edu/~ras/RS/Onto-Retrieval.pdf.

Years

Languages

  • e 444
  • d 42
  • pt 4
  • f 1
  • sp 1
  • More… Less…

Types

  • a 359
  • el 144
  • m 33
  • x 23
  • n 13
  • s 13
  • p 6
  • r 5
  • A 1
  • EL 1
  • More… Less…

Subjects

Classifications