Search (9466 results, page 1 of 474)

  • × language_ss:"e"
  1. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.34
    0.34273896 = sum of:
      0.049011134 = product of:
        0.1470334 = sum of:
          0.1470334 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
            0.1470334 = score(doc=701,freq=2.0), product of:
              0.39242527 = queryWeight, product of:
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.04628742 = queryNorm
              0.3746787 = fieldWeight in 701, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.03125 = fieldNorm(doc=701)
        0.33333334 = coord(1/3)
      0.070035525 = weight(_text_:retrieval in 701) [ClassicSimilarity], result of:
        0.070035525 = score(doc=701,freq=28.0), product of:
          0.14001551 = queryWeight, product of:
            3.024915 = idf(docFreq=5836, maxDocs=44218)
            0.04628742 = queryNorm
          0.5001983 = fieldWeight in 701, product of:
            5.2915025 = tf(freq=28.0), with freq of:
              28.0 = termFreq=28.0
            3.024915 = idf(docFreq=5836, maxDocs=44218)
            0.03125 = fieldNorm(doc=701)
      0.1470334 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
        0.1470334 = score(doc=701,freq=2.0), product of:
          0.39242527 = queryWeight, product of:
            8.478011 = idf(docFreq=24, maxDocs=44218)
            0.04628742 = queryNorm
          0.3746787 = fieldWeight in 701, product of:
            1.4142135 = tf(freq=2.0), with freq of:
              2.0 = termFreq=2.0
            8.478011 = idf(docFreq=24, maxDocs=44218)
            0.03125 = fieldNorm(doc=701)
      0.06125315 = weight(_text_:semantic in 701) [ClassicSimilarity], result of:
        0.06125315 = score(doc=701,freq=6.0), product of:
          0.19245663 = queryWeight, product of:
            4.1578603 = idf(docFreq=1879, maxDocs=44218)
            0.04628742 = queryNorm
          0.31826988 = fieldWeight in 701, product of:
            2.4494898 = tf(freq=6.0), with freq of:
              6.0 = termFreq=6.0
            4.1578603 = idf(docFreq=1879, maxDocs=44218)
            0.03125 = fieldNorm(doc=701)
      0.0154057555 = product of:
        0.030811511 = sum of:
          0.030811511 = weight(_text_:web in 701) [ClassicSimilarity], result of:
            0.030811511 = score(doc=701,freq=4.0), product of:
              0.15105948 = queryWeight, product of:
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.04628742 = queryNorm
              0.2039694 = fieldWeight in 701, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.03125 = fieldNorm(doc=701)
        0.5 = coord(1/2)
    
    Abstract
    By the explosion of possibilities for a ubiquitous content production, the information overload problem reaches the level of complexity which cannot be managed by traditional modelling approaches anymore. Due to their pure syntactical nature traditional information retrieval approaches did not succeed in treating content itself (i.e. its meaning, and not its representation). This leads to a very low usefulness of the results of a retrieval process for a user's task at hand. In the last ten years ontologies have been emerged from an interesting conceptualisation paradigm to a very promising (semantic) modelling technology, especially in the context of the Semantic Web. From the information retrieval point of view, ontologies enable a machine-understandable form of content description, such that the retrieval process can be driven by the meaning of the content. However, the very ambiguous nature of the retrieval process in which a user, due to the unfamiliarity with the underlying repository and/or query syntax, just approximates his information need in a query, implies a necessity to include the user in the retrieval process more actively in order to close the gap between the meaning of the content and the meaning of a user's query (i.e. his information need). This thesis lays foundation for such an ontology-based interactive retrieval process, in which the retrieval system interacts with a user in order to conceptually interpret the meaning of his query, whereas the underlying domain ontology drives the conceptualisation process. In that way the retrieval process evolves from a query evaluation process into a highly interactive cooperation between a user and the retrieval system, in which the system tries to anticipate the user's information need and to deliver the relevant content proactively. Moreover, the notion of content relevance for a user's query evolves from a content dependent artefact to the multidimensional context-dependent structure, strongly influenced by the user's preferences. This cooperation process is realized as the so-called Librarian Agent Query Refinement Process. In order to clarify the impact of an ontology on the retrieval process (regarding its complexity and quality), a set of methods and tools for different levels of content and query formalisation is developed, ranging from pure ontology-based inferencing to keyword-based querying in which semantics automatically emerges from the results. Our evaluation studies have shown that the possibilities to conceptualize a user's information need in the right manner and to interpret the retrieval results accordingly are key issues for realizing much more meaningful information retrieval systems.
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
    Theme
    Semantic Web
  2. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.33
    0.3339376 = product of:
      0.41742197 = sum of:
        0.0735167 = product of:
          0.22055008 = sum of:
            0.22055008 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.22055008 = score(doc=562,freq=2.0), product of:
                0.39242527 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04628742 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.22055008 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.22055008 = score(doc=562,freq=2.0), product of:
            0.39242527 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.04628742 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.05304678 = weight(_text_:semantic in 562) [ClassicSimilarity], result of:
          0.05304678 = score(doc=562,freq=2.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.2756298 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.07030838 = sum of:
          0.03268054 = weight(_text_:web in 562) [ClassicSimilarity], result of:
            0.03268054 = score(doc=562,freq=2.0), product of:
              0.15105948 = queryWeight, product of:
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.04628742 = queryNorm
              0.21634221 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
          0.03762784 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
            0.03762784 = score(doc=562,freq=2.0), product of:
              0.16209066 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04628742 = queryNorm
              0.23214069 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
      0.8 = coord(4/5)
    
    Abstract
    Document representations for text classification are typically based on the classical Bag-Of-Words paradigm. This approach comes with deficiencies that motivate the integration of features on a higher semantic level than single words. In this paper we propose an enhancement of the classical document representation through concepts extracted from background knowledge. Boosting is used for actual classification. Experimental evaluations on two well known text corpora support our approach through consistent improvement of the results.
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  3. Kleineberg, M.: Context analysis and context indexing : formal pragmatics in knowledge organization (2014) 0.31
    0.3104071 = product of:
      0.51734513 = sum of:
        0.12252783 = product of:
          0.36758348 = sum of:
            0.36758348 = weight(_text_:3a in 1826) [ClassicSimilarity], result of:
              0.36758348 = score(doc=1826,freq=2.0), product of:
                0.39242527 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04628742 = queryNorm
                0.93669677 = fieldWeight in 1826, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1826)
          0.33333334 = coord(1/3)
        0.36758348 = weight(_text_:2f in 1826) [ClassicSimilarity], result of:
          0.36758348 = score(doc=1826,freq=2.0), product of:
            0.39242527 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.04628742 = queryNorm
            0.93669677 = fieldWeight in 1826, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.078125 = fieldNorm(doc=1826)
        0.027233787 = product of:
          0.054467574 = sum of:
            0.054467574 = weight(_text_:web in 1826) [ClassicSimilarity], result of:
              0.054467574 = score(doc=1826,freq=2.0), product of:
                0.15105948 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.04628742 = queryNorm
                0.36057037 = fieldWeight in 1826, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1826)
          0.5 = coord(1/2)
      0.6 = coord(3/5)
    
    Source
    http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=5&ved=0CDQQFjAE&url=http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F3131107&ei=HzFWVYvGMsiNsgGTyoFI&usg=AFQjCNE2FHUeR9oQTQlNC4TPedv4Mo3DaQ&sig2=Rlzpr7a3BLZZkqZCXXN_IA&bvm=bv.93564037,d.bGg&cad=rja
  4. Vetere, G.; Lenzerini, M.: Models for semantic interoperability in service-oriented architectures (2005) 0.29
    0.28887814 = product of:
      0.48146355 = sum of:
        0.08576949 = product of:
          0.25730845 = sum of:
            0.25730845 = weight(_text_:3a in 306) [ClassicSimilarity], result of:
              0.25730845 = score(doc=306,freq=2.0), product of:
                0.39242527 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04628742 = queryNorm
                0.65568775 = fieldWeight in 306, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=306)
          0.33333334 = coord(1/3)
        0.25730845 = weight(_text_:2f in 306) [ClassicSimilarity], result of:
          0.25730845 = score(doc=306,freq=2.0), product of:
            0.39242527 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.04628742 = queryNorm
            0.65568775 = fieldWeight in 306, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
        0.13838558 = weight(_text_:semantic in 306) [ClassicSimilarity], result of:
          0.13838558 = score(doc=306,freq=10.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.71904814 = fieldWeight in 306, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
      0.6 = coord(3/5)
    
    Abstract
    Although service-oriented architectures go a long way toward providing interoperability in distributed, heterogeneous environments, managing semantic differences in such environments remains a challenge. We give an overview of the issue of semantic interoperability (integration), provide a semantic characterization of services, and discuss the role of ontologies. Then we analyze four basic models of semantic interoperability that differ in respect to their mapping between service descriptions and ontologies and in respect to where the evaluation of the integration logic is performed. We also provide some guidelines for selecting one of the possible interoperability models.
    Content
    Vgl.: http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=5386707&url=http%3A%2F%2Fieeexplore.ieee.org%2Fxpls%2Fabs_all.jsp%3Farnumber%3D5386707.
  5. Farazi, M.: Faceted lightweight ontologies : a formalization and some experiments (2010) 0.26
    0.26492563 = product of:
      0.33115703 = sum of:
        0.061263915 = product of:
          0.18379174 = sum of:
            0.18379174 = weight(_text_:3a in 4997) [ClassicSimilarity], result of:
              0.18379174 = score(doc=4997,freq=2.0), product of:
                0.39242527 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04628742 = queryNorm
                0.46834838 = fieldWeight in 4997, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4997)
          0.33333334 = coord(1/3)
        0.18379174 = weight(_text_:2f in 4997) [ClassicSimilarity], result of:
          0.18379174 = score(doc=4997,freq=2.0), product of:
            0.39242527 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.04628742 = queryNorm
            0.46834838 = fieldWeight in 4997, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4997)
        0.062516235 = weight(_text_:semantic in 4997) [ClassicSimilarity], result of:
          0.062516235 = score(doc=4997,freq=4.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.32483283 = fieldWeight in 4997, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4997)
        0.02358515 = product of:
          0.0471703 = sum of:
            0.0471703 = weight(_text_:web in 4997) [ClassicSimilarity], result of:
              0.0471703 = score(doc=4997,freq=6.0), product of:
                0.15105948 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.04628742 = queryNorm
                0.3122631 = fieldWeight in 4997, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4997)
          0.5 = coord(1/2)
      0.8 = coord(4/5)
    
    Abstract
    While classifications are heavily used to categorize web content, the evolution of the web foresees a more formal structure - ontology - which can serve this purpose. Ontologies are core artifacts of the Semantic Web which enable machines to use inference rules to conduct automated reasoning on data. Lightweight ontologies bridge the gap between classifications and ontologies. A lightweight ontology (LO) is an ontology representing a backbone taxonomy where the concept of the child node is more specific than the concept of the parent node. Formal lightweight ontologies can be generated from their informal ones. The key applications of formal lightweight ontologies are document classification, semantic search, and data integration. However, these applications suffer from the following problems: the disambiguation accuracy of the state of the art NLP tools used in generating formal lightweight ontologies from their informal ones; the lack of background knowledge needed for the formal lightweight ontologies; and the limitation of ontology reuse. In this dissertation, we propose a novel solution to these problems in formal lightweight ontologies; namely, faceted lightweight ontology (FLO). FLO is a lightweight ontology in which terms, present in each node label, and their concepts, are available in the background knowledge (BK), which is organized as a set of facets. A facet can be defined as a distinctive property of the groups of concepts that can help in differentiating one group from another. Background knowledge can be defined as a subset of a knowledge base, such as WordNet, and often represents a specific domain.
    Content
    PhD Dissertation at International Doctorate School in Information and Communication Technology. Vgl.: https%3A%2F%2Fcore.ac.uk%2Fdownload%2Fpdf%2F150083013.pdf&usg=AOvVaw2n-qisNagpyT0lli_6QbAQ.
  6. Xiong, C.: Knowledge based text representations for information retrieval (2016) 0.26
    0.26379827 = product of:
      0.32974783 = sum of:
        0.049011134 = product of:
          0.1470334 = sum of:
            0.1470334 = weight(_text_:3a in 5820) [ClassicSimilarity], result of:
              0.1470334 = score(doc=5820,freq=2.0), product of:
                0.39242527 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04628742 = queryNorm
                0.3746787 = fieldWeight in 5820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.33333334 = coord(1/3)
        0.03743556 = weight(_text_:retrieval in 5820) [ClassicSimilarity], result of:
          0.03743556 = score(doc=5820,freq=8.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.26736724 = fieldWeight in 5820, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.20793661 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.20793661 = score(doc=5820,freq=4.0), product of:
            0.39242527 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.04628742 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.03536452 = weight(_text_:semantic in 5820) [ClassicSimilarity], result of:
          0.03536452 = score(doc=5820,freq=2.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.18375319 = fieldWeight in 5820, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
      0.8 = coord(4/5)
    
    Abstract
    The successes of information retrieval (IR) in recent decades were built upon bag-of-words representations. Effective as it is, bag-of-words is only a shallow text understanding; there is a limited amount of information for document ranking in the word space. This dissertation goes beyond words and builds knowledge based text representations, which embed the external and carefully curated information from knowledge bases, and provide richer and structured evidence for more advanced information retrieval systems. This thesis research first builds query representations with entities associated with the query. Entities' descriptions are used by query expansion techniques that enrich the query with explanation terms. Then we present a general framework that represents a query with entities that appear in the query, are retrieved by the query, or frequently show up in the top retrieved documents. A latent space model is developed to jointly learn the connections from query to entities and the ranking of documents, modeling the external evidence from knowledge bases and internal ranking features cooperatively. To further improve the quality of relevant entities, a defining factor of our query representations, we introduce learning to rank to entity search and retrieve better entities from knowledge bases. In the document representation part, this thesis research also moves one step forward with a bag-of-entities model, in which documents are represented by their automatic entity annotations, and the ranking is performed in the entity space.
    This proposal includes plans to improve the quality of relevant entities with a co-learning framework that learns from both entity labels and document labels. We also plan to develop a hybrid ranking system that combines word based and entity based representations together with their uncertainties considered. At last, we plan to enrich the text representations with connections between entities. We propose several ways to infer entity graph representations for texts, and to rank documents using their structure representations. This dissertation overcomes the limitation of word based representations with external and carefully curated information from knowledge bases. We believe this thesis research is a solid start towards the new generation of intelligent, semantic, and structured information retrieval.
    Content
    Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Language and Information Technologies. Vgl.: https%3A%2F%2Fwww.cs.cmu.edu%2F~cx%2Fpapers%2Fknowledge_based_text_representation.pdf&usg=AOvVaw0SaTSvhWLTh__Uz_HtOtl3.
  7. Li, L.; Shang, Y.; Zhang, W.: Improvement of HITS-based algorithms on Web documents 0.25
    0.24511816 = product of:
      0.40853024 = sum of:
        0.0735167 = product of:
          0.22055008 = sum of:
            0.22055008 = weight(_text_:3a in 2514) [ClassicSimilarity], result of:
              0.22055008 = score(doc=2514,freq=2.0), product of:
                0.39242527 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04628742 = queryNorm
                0.56201804 = fieldWeight in 2514, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2514)
          0.33333334 = coord(1/3)
        0.3119049 = weight(_text_:2f in 2514) [ClassicSimilarity], result of:
          0.3119049 = score(doc=2514,freq=4.0), product of:
            0.39242527 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.04628742 = queryNorm
            0.7948135 = fieldWeight in 2514, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2514)
        0.023108633 = product of:
          0.046217266 = sum of:
            0.046217266 = weight(_text_:web in 2514) [ClassicSimilarity], result of:
              0.046217266 = score(doc=2514,freq=4.0), product of:
                0.15105948 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.04628742 = queryNorm
                0.3059541 = fieldWeight in 2514, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2514)
          0.5 = coord(1/2)
      0.6 = coord(3/5)
    
    Content
    Vgl.: http%3A%2F%2Fdelab.csd.auth.gr%2F~dimitris%2Fcourses%2Fir_spring06%2Fpage_rank_computing%2Fp527-li.pdf. Vgl. auch: http://www2002.org/CDROM/refereed/643/.
    Source
    WWW '02: Proceedings of the 11th International Conference on World Wide Web, May 7-11, 2002, Honolulu, Hawaii, USA
  8. Gödert, W.; Hubrich, J.; Nagelschmidt, M.: Semantic knowledge representation for information retrieval (2014) 0.22
    0.22458757 = product of:
      0.3743126 = sum of:
        0.07428389 = weight(_text_:retrieval in 987) [ClassicSimilarity], result of:
          0.07428389 = score(doc=987,freq=14.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.5305404 = fieldWeight in 987, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=987)
        0.17593628 = weight(_text_:semantic in 987) [ClassicSimilarity], result of:
          0.17593628 = score(doc=987,freq=22.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.91416067 = fieldWeight in 987, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.046875 = fieldNorm(doc=987)
        0.12409244 = sum of:
          0.0864646 = weight(_text_:web in 987) [ClassicSimilarity], result of:
            0.0864646 = score(doc=987,freq=14.0), product of:
              0.15105948 = queryWeight, product of:
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.04628742 = queryNorm
              0.57238775 = fieldWeight in 987, product of:
                3.7416575 = tf(freq=14.0), with freq of:
                  14.0 = termFreq=14.0
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.046875 = fieldNorm(doc=987)
          0.03762784 = weight(_text_:22 in 987) [ClassicSimilarity], result of:
            0.03762784 = score(doc=987,freq=2.0), product of:
              0.16209066 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04628742 = queryNorm
              0.23214069 = fieldWeight in 987, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=987)
      0.6 = coord(3/5)
    
    Abstract
    This book covers the basics of semantic web technologies and indexing languages, and describes their contribution to improve languages as a tool for subject queries and knowledge exploration. The book is relevant to information scientists, knowledge workers and indexers. It provides a suitable combination of theoretical foundations and practical applications.
    Content
    Introduction: envisioning semantic information spacesIndexing and knowledge organization -- Semantic technologies for knowledge representation -- Information retrieval and knowledge exploration -- Approaches to handle heterogeneity -- Problems with establishing semantic interoperability -- Formalization in indexing languages -- Typification of semantic relations -- Inferences in retrieval processes -- Semantic interoperability and inferences -- Remaining research questions.
    Date
    23. 7.2017 13:49:22
    LCSH
    Semantic Web
    Information retrieval
    World Wide Web / Subject access
    RSWK
    Semantic Web
    Information Retrieval
    Subject
    Semantic Web
    Information retrieval
    World Wide Web / Subject access
    Semantic Web
    Information Retrieval
  9. Huo, W.: Automatic multi-word term extraction and its application to Web-page summarization (2012) 0.21
    0.2109694 = product of:
      0.35161567 = sum of:
        0.028076671 = weight(_text_:retrieval in 563) [ClassicSimilarity], result of:
          0.028076671 = score(doc=563,freq=2.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.20052543 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.22055008 = weight(_text_:2f in 563) [ClassicSimilarity], result of:
          0.22055008 = score(doc=563,freq=2.0), product of:
            0.39242527 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.04628742 = queryNorm
            0.56201804 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.10298892 = sum of:
          0.06536108 = weight(_text_:web in 563) [ClassicSimilarity], result of:
            0.06536108 = score(doc=563,freq=8.0), product of:
              0.15105948 = queryWeight, product of:
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.04628742 = queryNorm
              0.43268442 = fieldWeight in 563, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.046875 = fieldNorm(doc=563)
          0.03762784 = weight(_text_:22 in 563) [ClassicSimilarity], result of:
            0.03762784 = score(doc=563,freq=2.0), product of:
              0.16209066 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04628742 = queryNorm
              0.23214069 = fieldWeight in 563, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=563)
      0.6 = coord(3/5)
    
    Abstract
    In this thesis we propose three new word association measures for multi-word term extraction. We combine these association measures with LocalMaxs algorithm in our extraction model and compare the results of different multi-word term extraction methods. Our approach is language and domain independent and requires no training data. It can be applied to such tasks as text summarization, information retrieval, and document classification. We further explore the potential of using multi-word terms as an effective representation for general web-page summarization. We extract multi-word terms from human written summaries in a large collection of web-pages, and generate the summaries by aligning document words with these multi-word terms. Our system applies machine translation technology to learn the aligning process from a training set and focuses on selecting high quality multi-word terms from human written summaries to generate suitable results for web-page summarization.
    Content
    A Thesis presented to The University of Guelph In partial fulfilment of requirements for the degree of Master of Science in Computer Science. Vgl. Unter: http://www.inf.ufrgs.br%2F~ceramisch%2Fdownload_files%2Fpublications%2F2009%2Fp01.pdf.
    Date
    10. 1.2013 19:22:47
  10. Mayfield, J.; Finin, T.: Information retrieval on the Semantic Web : integrating inference and retrieval 0.20
    0.2005874 = product of:
      0.33431232 = sum of:
        0.07324491 = weight(_text_:retrieval in 4330) [ClassicSimilarity], result of:
          0.07324491 = score(doc=4330,freq=10.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.5231199 = fieldWeight in 4330, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4330)
        0.12377582 = weight(_text_:semantic in 4330) [ClassicSimilarity], result of:
          0.12377582 = score(doc=4330,freq=8.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.64313614 = fieldWeight in 4330, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4330)
        0.13729158 = sum of:
          0.09339243 = weight(_text_:web in 4330) [ClassicSimilarity], result of:
            0.09339243 = score(doc=4330,freq=12.0), product of:
              0.15105948 = queryWeight, product of:
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.04628742 = queryNorm
              0.6182494 = fieldWeight in 4330, product of:
                3.4641016 = tf(freq=12.0), with freq of:
                  12.0 = termFreq=12.0
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.0546875 = fieldNorm(doc=4330)
          0.043899145 = weight(_text_:22 in 4330) [ClassicSimilarity], result of:
            0.043899145 = score(doc=4330,freq=2.0), product of:
              0.16209066 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04628742 = queryNorm
              0.2708308 = fieldWeight in 4330, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=4330)
      0.6 = coord(3/5)
    
    Abstract
    One vision of the Semantic Web is that it will be much like the Web we know today, except that documents will be enriched by annotations in machine understandable markup. These annotations will provide metadata about the documents as well as machine interpretable statements capturing some of the meaning of document content. We discuss how the information retrieval paradigm might be recast in such an environment. We suggest that retrieval can be tightly bound to inference. Doing so makes today's Web search engines useful to Semantic Web inference engines, and causes improvements in either retrieval or inference to lead directly to improvements in the other.
    Date
    12. 2.2011 17:35:22
    Theme
    Semantic Web
  11. Scheir, P.; Pammer, V.; Lindstaedt, S.N.: Information retrieval on the Semantic Web : does it exist? (2007) 0.17
    0.17459723 = product of:
      0.2909954 = sum of:
        0.06551223 = weight(_text_:retrieval in 4329) [ClassicSimilarity], result of:
          0.06551223 = score(doc=4329,freq=8.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.46789268 = fieldWeight in 4329, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4329)
        0.17504546 = weight(_text_:semantic in 4329) [ClassicSimilarity], result of:
          0.17504546 = score(doc=4329,freq=16.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.90953195 = fieldWeight in 4329, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4329)
        0.05043768 = product of:
          0.10087536 = sum of:
            0.10087536 = weight(_text_:web in 4329) [ClassicSimilarity], result of:
              0.10087536 = score(doc=4329,freq=14.0), product of:
                0.15105948 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.04628742 = queryNorm
                0.6677857 = fieldWeight in 4329, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4329)
          0.5 = coord(1/2)
      0.6 = coord(3/5)
    
    Abstract
    Plenty of contemporary attempts to search exist that are associated with the area of Semantic Web. But which of them qualify as information retrieval for the Semantic Web? Do such approaches exist? To answer these questions we take a look at the nature of the Semantic Web and Semantic Desktop and at definitions for information and data retrieval. We survey current approaches referred to by their authors as information retrieval for the Semantic Web or that use Semantic Web technology for search.
    Theme
    Semantic Web
  12. Sánchez, M.F.: Semantically enhanced Information Retrieval : an ontology-based approach (2006) 0.17
    0.16890861 = product of:
      0.28151435 = sum of:
        0.066177346 = weight(_text_:retrieval in 4327) [ClassicSimilarity], result of:
          0.066177346 = score(doc=4327,freq=4.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.47264296 = fieldWeight in 4327, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.078125 = fieldNorm(doc=4327)
        0.1768226 = weight(_text_:semantic in 4327) [ClassicSimilarity], result of:
          0.1768226 = score(doc=4327,freq=8.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.91876596 = fieldWeight in 4327, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.078125 = fieldNorm(doc=4327)
        0.03851439 = product of:
          0.07702878 = sum of:
            0.07702878 = weight(_text_:web in 4327) [ClassicSimilarity], result of:
              0.07702878 = score(doc=4327,freq=4.0), product of:
                0.15105948 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.04628742 = queryNorm
                0.5099235 = fieldWeight in 4327, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4327)
          0.5 = coord(1/2)
      0.6 = coord(3/5)
    
    Content
    Part I. Analyzing the state of the art - What is semantic search? Part II. The proposal - An ontology-based IR model - Semantic retrieval on the Web Part III. Extensions - Semantic knowledge gateway - Coping with knowledge incompleteness
    Theme
    Semantic Web
  13. ¬The Semantic Web : research and applications ; second European Semantic WebConference, ESWC 2005, Heraklion, Crete, Greece, May 29 - June 1, 2005 ; proceedings (2005) 0.17
    0.1657434 = product of:
      0.27623898 = sum of:
        0.048630223 = weight(_text_:retrieval in 439) [ClassicSimilarity], result of:
          0.048630223 = score(doc=439,freq=6.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.34732026 = fieldWeight in 439, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=439)
        0.17593628 = weight(_text_:semantic in 439) [ClassicSimilarity], result of:
          0.17593628 = score(doc=439,freq=22.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.91416067 = fieldWeight in 439, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.046875 = fieldNorm(doc=439)
        0.051672477 = product of:
          0.103344955 = sum of:
            0.103344955 = weight(_text_:web in 439) [ClassicSimilarity], result of:
              0.103344955 = score(doc=439,freq=20.0), product of:
                0.15105948 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.04628742 = queryNorm
                0.6841342 = fieldWeight in 439, product of:
                  4.472136 = tf(freq=20.0), with freq of:
                    20.0 = termFreq=20.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.046875 = fieldNorm(doc=439)
          0.5 = coord(1/2)
      0.6 = coord(3/5)
    
    Abstract
    This book constitutes the refereed proceedings of the Second European Semantic Web Conference, ESWC 2005, heldin Heraklion, Crete, Greece in May/June 2005. The 48 revised full papers presented were carefully reviewed and selected from 148 submissions. The papers are organized in topical sections on semantic Web services, languages, ontologies, reasoning and querying, search and information retrieval, user and communities, natural language for the semantic Web, annotation tools, and semantic Web applications.
    LCSH
    Information storage and retrieval systems
    RSWK
    Semantic Web / Kongress / Iraklion <2005>
    Semantic Web / Ontologie <Wissensverarbeitung> / Kongress / Iraklion <2005>
    Subject
    Semantic Web / Kongress / Iraklion <2005>
    Semantic Web / Ontologie <Wissensverarbeitung> / Kongress / Iraklion <2005>
    Information storage and retrieval systems
    Theme
    Semantic Web
  14. Popper, K.R.: Three worlds : the Tanner lecture on human values. Deliverd at the University of Michigan, April 7, 1978 (1978) 0.16
    0.15683563 = product of:
      0.39208907 = sum of:
        0.09802227 = product of:
          0.2940668 = sum of:
            0.2940668 = weight(_text_:3a in 230) [ClassicSimilarity], result of:
              0.2940668 = score(doc=230,freq=2.0), product of:
                0.39242527 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04628742 = queryNorm
                0.7493574 = fieldWeight in 230, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0625 = fieldNorm(doc=230)
          0.33333334 = coord(1/3)
        0.2940668 = weight(_text_:2f in 230) [ClassicSimilarity], result of:
          0.2940668 = score(doc=230,freq=2.0), product of:
            0.39242527 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.04628742 = queryNorm
            0.7493574 = fieldWeight in 230, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=230)
      0.4 = coord(2/5)
    
    Source
    https%3A%2F%2Ftannerlectures.utah.edu%2F_documents%2Fa-to-z%2Fp%2Fpopper80.pdf&usg=AOvVaw3f4QRTEH-OEBmoYr2J_c7H
  15. Franklin, R.A.: Re-inventing subject access for the semantic web (2003) 0.15
    0.1464293 = product of:
      0.24404883 = sum of:
        0.028076671 = weight(_text_:retrieval in 2556) [ClassicSimilarity], result of:
          0.028076671 = score(doc=2556,freq=2.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.20052543 = fieldWeight in 2556, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=2556)
        0.091879725 = weight(_text_:semantic in 2556) [ClassicSimilarity], result of:
          0.091879725 = score(doc=2556,freq=6.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.47740483 = fieldWeight in 2556, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.046875 = fieldNorm(doc=2556)
        0.12409244 = sum of:
          0.0864646 = weight(_text_:web in 2556) [ClassicSimilarity], result of:
            0.0864646 = score(doc=2556,freq=14.0), product of:
              0.15105948 = queryWeight, product of:
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.04628742 = queryNorm
              0.57238775 = fieldWeight in 2556, product of:
                3.7416575 = tf(freq=14.0), with freq of:
                  14.0 = termFreq=14.0
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.046875 = fieldNorm(doc=2556)
          0.03762784 = weight(_text_:22 in 2556) [ClassicSimilarity], result of:
            0.03762784 = score(doc=2556,freq=2.0), product of:
              0.16209066 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04628742 = queryNorm
              0.23214069 = fieldWeight in 2556, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2556)
      0.6 = coord(3/5)
    
    Abstract
    First generation scholarly research on the Web lacked a firm system of authority control. Second generation Web research is beginning to model subject access with library science principles of bibliographic control and cataloguing. Harnessing the Web and organising the intellectual content with standards and controlled vocabulary provides precise search and retrieval capability, increasing relevance and efficient use of technology. Dublin Core metadata standards permit a full evaluation and cataloguing of Web resources appropriate to highly specific research needs and discovery. Current research points to a type of structure based on a system of faceted classification. This system allows the semantic and syntactic relationships to be defined. Controlled vocabulary, such as the Library of Congress Subject Headings, can be assigned, not in a hierarchical structure, but rather as descriptive facets of relating concepts. Web design features such as this are adding value to discovery and filtering out data that lack authority. The system design allows for scalability and extensibility, two technical features that are integral to future development of the digital library and resource discovery.
    Date
    30.12.2008 18:22:46
    Theme
    Semantic Web
  16. Semantic applications (2018) 0.14
    0.14236757 = product of:
      0.23727927 = sum of:
        0.057311267 = weight(_text_:retrieval in 5204) [ClassicSimilarity], result of:
          0.057311267 = score(doc=5204,freq=12.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.40932083 = fieldWeight in 5204, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
        0.14661357 = weight(_text_:semantic in 5204) [ClassicSimilarity], result of:
          0.14661357 = score(doc=5204,freq=22.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.7618005 = fieldWeight in 5204, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
        0.03335444 = product of:
          0.06670888 = sum of:
            0.06670888 = weight(_text_:web in 5204) [ClassicSimilarity], result of:
              0.06670888 = score(doc=5204,freq=12.0), product of:
                0.15105948 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.04628742 = queryNorm
                0.4416067 = fieldWeight in 5204, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5204)
          0.5 = coord(1/2)
      0.6 = coord(3/5)
    
    Abstract
    This book describes proven methodologies for developing semantic applications: software applications which explicitly or implicitly uses the semantics (i.e., the meaning) of a domain terminology in order to improve usability, correctness, and completeness. An example is semantic search, where synonyms and related terms are used for enriching the results of a simple text-based search. Ontologies, thesauri or controlled vocabularies are the centerpiece of semantic applications. The book includes technological and architectural best practices for corporate use.
    Content
    Introduction.- Ontology Development.- Compliance using Metadata.- Variety Management for Big Data.- Text Mining in Economics.- Generation of Natural Language Texts.- Sentiment Analysis.- Building Concise Text Corpora from Web Contents.- Ontology-Based Modelling of Web Content.- Personalized Clinical Decision Support for Cancer Care.- Applications of Temporal Conceptual Semantic Systems.- Context-Aware Documentation in the Smart Factory.- Knowledge-Based Production Planning for Industry 4.0.- Information Exchange in Jurisdiction.- Supporting Automated License Clearing.- Managing cultural assets: Implementing typical cultural heritage archive's usage scenarios via Semantic Web technologies.- Semantic Applications for Process Management.- Domain-Specific Semantic Search Applications.
    LCSH
    Information storage and retrieval
    Information Storage and Retrieval
    RSWK
    Information Retrieval
    Semantic Web
    Subject
    Information Retrieval
    Semantic Web
    Information storage and retrieval
    Information Storage and Retrieval
    Theme
    Semantic Web
  17. Semantic digital libraries (2009) 0.14
    0.13900207 = product of:
      0.2316701 = sum of:
        0.01871778 = weight(_text_:retrieval in 3371) [ClassicSimilarity], result of:
          0.01871778 = score(doc=3371,freq=2.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.13368362 = fieldWeight in 3371, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=3371)
        0.17682262 = weight(_text_:semantic in 3371) [ClassicSimilarity], result of:
          0.17682262 = score(doc=3371,freq=50.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.918766 = fieldWeight in 3371, product of:
              7.071068 = tf(freq=50.0), with freq of:
                50.0 = termFreq=50.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.03125 = fieldNorm(doc=3371)
        0.036129702 = product of:
          0.072259404 = sum of:
            0.072259404 = weight(_text_:web in 3371) [ClassicSimilarity], result of:
              0.072259404 = score(doc=3371,freq=22.0), product of:
                0.15105948 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.04628742 = queryNorm
                0.47835067 = fieldWeight in 3371, product of:
                  4.690416 = tf(freq=22.0), with freq of:
                    22.0 = termFreq=22.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3371)
          0.5 = coord(1/2)
      0.6 = coord(3/5)
    
    Abstract
    Libraries have always been an inspiration for the standards and technologies developed by semantic web activities. However, except for the Dublin Core specification, semantic web and social networking technologies have not been widely adopted and further developed by major digital library initiatives and projects. Yet semantic technologies offer a new level of flexibility, interoperability, and relationships for digital repositories. Kruk and McDaniel present semantic web-related aspects of current digital library activities, and introduce their functionality; they show examples ranging from general architectural descriptions to detailed usages of specific ontologies, and thus stimulate the awareness of researchers, engineers, and potential users of those technologies. Their presentation is completed by chapters on existing prototype systems such as JeromeDL, BRICKS, and Greenstone, as well as a look into the possible future of semantic digital libraries. This book is aimed at researchers and graduate students in areas like digital libraries, the semantic web, social networks, and information retrieval. This audience will benefit from detailed descriptions of both today's possibilities and also the shortcomings of applying semantic web technologies to large digital repositories of often unstructured data.
    Content
    Inhalt: Introduction to Digital Libraries and Semantic Web: Introduction / Bill McDaniel and Sebastian Ryszard Kruk - Digital Libraries and Knowledge Organization / Dagobert Soergel - Semantic Web and Ontologies / Marcin Synak, Maciej Dabrowski and Sebastian Ryszard Kruk - Social Semantic Information Spaces / John G. Breslin A Vision of Semantic Digital Libraries: Goals of Semantic Digital Libraries / Sebastian Ryszard Kruk and Bill McDaniel - Architecture of Semantic Digital Libraries / Sebastian Ryszard Kruk, Adam Westerki and Ewelina Kruk - Long-time Preservation / Markus Reis Ontologies for Semantic Digital Libraries: Bibliographic Ontology / Maciej Dabrowski, Macin Synak and Sebastian Ryszard Kruk - Community-aware Ontologies / Slawomir Grzonkowski, Sebastian Ryszard Kruk, Adam Gzella, Jakub Demczuk and Bill McDaniel Prototypes of Semantic Digital Libraries: JeromeDL: The Social Semantic Digital Library / Sebastian Ryszard Kruk, Mariusz Cygan, Adam Gzella, Tomasz Woroniecki and Maciej Dabrowski - The BRICKS Digital Library Infrastructure / Bernhard Haslhofer and Predrag Knezevié - Semantics in Greenstone / Annika Hinze, George Buchanan, David Bainbridge and Ian Witten Building the Future - Semantic Digital Libraries in Use: Hyperbooks / Gilles Falquet, Luka Nerima and Jean-Claude Ziswiler - Semantic Digital Libraries for Archiving / Bill McDaniel - Evaluation of Semantic and Social Technologies for Digital Libraries / Sebastian Ryszard Kruk, Ewelina Kruk and Katarzyna Stankiewicz - Conclusions: The Future of Semantic Digital Libraries / Sebastian Ryszard Kruk and Bill McDaniel
    LCSH
    Semantic Web
    RSWK
    Elektronische Bibliothek / Semantic Web / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Subject
    Elektronische Bibliothek / Semantic Web / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Semantic Web
  18. Atanassova, I.; Bertin, M.: Semantic facets for scientific information retrieval (2014) 0.14
    0.13643564 = product of:
      0.22739272 = sum of:
        0.05673526 = weight(_text_:retrieval in 4471) [ClassicSimilarity], result of:
          0.05673526 = score(doc=4471,freq=6.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.40520695 = fieldWeight in 4471, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4471)
        0.1515938 = weight(_text_:semantic in 4471) [ClassicSimilarity], result of:
          0.1515938 = score(doc=4471,freq=12.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.78767776 = fieldWeight in 4471, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4471)
        0.019063652 = product of:
          0.038127303 = sum of:
            0.038127303 = weight(_text_:web in 4471) [ClassicSimilarity], result of:
              0.038127303 = score(doc=4471,freq=2.0), product of:
                0.15105948 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.04628742 = queryNorm
                0.25239927 = fieldWeight in 4471, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4471)
          0.5 = coord(1/2)
      0.6 = coord(3/5)
    
    Abstract
    We present an Information Retrieval System for scientific publications that provides the possibility to filter results according to semantic facets. We use sentence-level semantic annotations that identify specific semantic relations in texts, such as methods, definitions, hypotheses, that correspond to common information needs related to scientific literature. The semantic annotations are obtained using a rule-based method that identifies linguistic clues organized into a linguistic ontology. The system is implemented using Solr Search Server and offers efficient search and navigation in scientific papers.
    Source
    Semantic Web Evaluation Challenge. SemWebEval 2014 at ESWC 2014, Anissaras, Crete, Greece, May 25-29, 2014, Revised Selected Papers. Eds.: V. Presutti et al
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  19. Multimedia content and the Semantic Web : methods, standards, and tools (2005) 0.14
    0.13561465 = product of:
      0.2260244 = sum of:
        0.035095837 = weight(_text_:retrieval in 150) [ClassicSimilarity], result of:
          0.035095837 = score(doc=150,freq=18.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.25065678 = fieldWeight in 150, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.01953125 = fieldNorm(doc=150)
        0.10600142 = weight(_text_:semantic in 150) [ClassicSimilarity], result of:
          0.10600142 = score(doc=150,freq=46.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.55078083 = fieldWeight in 150, product of:
              6.78233 = tf(freq=46.0), with freq of:
                46.0 = termFreq=46.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.01953125 = fieldNorm(doc=150)
        0.08492714 = sum of:
          0.057771582 = weight(_text_:web in 150) [ClassicSimilarity], result of:
            0.057771582 = score(doc=150,freq=36.0), product of:
              0.15105948 = queryWeight, product of:
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.04628742 = queryNorm
              0.38244262 = fieldWeight in 150, product of:
                6.0 = tf(freq=36.0), with freq of:
                  36.0 = termFreq=36.0
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.01953125 = fieldNorm(doc=150)
          0.027155556 = weight(_text_:22 in 150) [ClassicSimilarity], result of:
            0.027155556 = score(doc=150,freq=6.0), product of:
              0.16209066 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04628742 = queryNorm
              0.16753313 = fieldWeight in 150, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.01953125 = fieldNorm(doc=150)
      0.6 = coord(3/5)
    
    Classification
    006.7 22
    Date
    7. 3.2007 19:30:22
    DDC
    006.7 22
    Footnote
    Rez. in: JASIST 58(2007) no.3, S.457-458 (A.M.A. Ahmad): "The concept of the semantic web has emerged because search engines and text-based searching are no longer adequate, as these approaches involve an extensive information retrieval process. The deployed searching and retrieving descriptors arc naturally subjective and their deployment is often restricted to the specific application domain for which the descriptors were configured. The new era of information technology imposes different kinds of requirements and challenges. Automatic extracted audiovisual features are required, as these features are more objective, domain-independent, and more native to audiovisual content. This book is a useful guide for researchers, experts, students, and practitioners; it is a very valuable reference and can lead them through their exploration and research in multimedia content and the semantic web. The book is well organized, and introduces the concept of the semantic web and multimedia content analysis to the reader through a logical sequence from standards and hypotheses through system examples, presenting relevant tools and methods. But in some chapters readers will need a good technical background to understand some of the details. Readers may attain sufficient knowledge here to start projects or research related to the book's theme; recent results and articles related to the active research area of integrating multimedia with semantic web technologies are included. This book includes full descriptions of approaches to specific problem domains such as content search, indexing, and retrieval. This book will be very useful to researchers in the multimedia content analysis field who wish to explore the benefits of emerging semantic web technologies in applying multimedia content approaches. The first part of the book covers the definition of the two basic terms multimedia content and semantic web. The Moving Picture Experts Group standards MPEG7 and MPEG21 are quoted extensively. In addition, the means of multimedia content description are elaborated upon and schematically drawn. This extensive description is introduced by authors who are actively involved in those standards and have been participating in the work of the International Organization for Standardization (ISO)/MPEG for many years. On the other hand, this results in bias against the ad hoc or nonstandard tools for multimedia description in favor of the standard approaches. This is a general book for multimedia content; more emphasis on the general multimedia description and extraction could be provided.
    Semantic web technologies are explained, and ontology representation is emphasized. There is an excellent summary of the fundamental theory behind applying a knowledge-engineering approach to vision problems. This summary represents the concept of the semantic web and multimedia content analysis. A definition of the fuzzy knowledge representation that can be used for realization in multimedia content applications has been provided, with a comprehensive analysis. The second part of the book introduces the multimedia content analysis approaches and applications. In addition, some examples of methods applicable to multimedia content analysis are presented. Multimedia content analysis is a very diverse field and concerns many other research fields at the same time; this creates strong diversity issues, as everything from low-level features (e.g., colors, DCT coefficients, motion vectors, etc.) up to the very high and semantic level (e.g., Object, Events, Tracks, etc.) are involved. The second part includes topics on structure identification (e.g., shot detection for video sequences), and object-based video indexing. These conventional analysis methods are supplemented by results on semantic multimedia analysis, including three detailed chapters on the development and use of knowledge models for automatic multimedia analysis. Starting from object-based indexing and continuing with machine learning, these three chapters are very logically organized. Because of the diversity of this research field, including several chapters of recent research results is not sufficient to cover the state of the art of multimedia. The editors of the book should write an introductory chapter about multimedia content analysis approaches, basic problems, and technical issues and challenges, and try to survey the state of the art of the field and thus introduce the field to the reader.
    The final part of the book discusses research in multimedia content management systems and the semantic web, and presents examples and applications for semantic multimedia analysis in search and retrieval systems. These chapters describe example systems in which current projects have been implemented, and include extensive results and real demonstrations. For example, real case scenarios such as ECommerce medical applications and Web services have been introduced. Topics in natural language, speech and image processing techniques and their application for multimedia indexing, and content-based retrieval have been elaborated upon with extensive examples and deployment methods. The editors of the book themselves provide the readers with a chapter about their latest research results on knowledge-based multimedia content indexing and retrieval. Some interesting applications for multimedia content and the semantic web are introduced. Applications that have taken advantage of the metadata provided by MPEG7 in order to realize advance-access services for multimedia content have been provided. The applications discussed in the third part of the book provide useful guidance to researchers and practitioners properly planning to implement semantic multimedia analysis techniques in new research and development projects in both academia and industry. A fourth part should be added to this book: performance measurements for integrated approaches of multimedia analysis and the semantic web. Performance of the semantic approach is a very sophisticated issue and requires extensive elaboration and effort. Measuring the semantic search is an ongoing research area; several chapters concerning performance measurement and analysis would be required to adequately cover this area and introduce it to readers."
    LCSH
    Semantic Web
    Information storage and retrieval systems
    RSWK
    Semantic Web / Multimedia / Automatische Indexierung / Information Retrieval
    Subject
    Semantic Web / Multimedia / Automatische Indexierung / Information Retrieval
    Semantic Web
    Information storage and retrieval systems
    Theme
    Semantic Web
  20. Brunetti, J.M.; Roberto García, R.: User-centered design and evaluation of overview components for semantic data exploration (2014) 0.13
    0.13363014 = product of:
      0.22271688 = sum of:
        0.01871778 = weight(_text_:retrieval in 1626) [ClassicSimilarity], result of:
          0.01871778 = score(doc=1626,freq=2.0), product of:
            0.14001551 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04628742 = queryNorm
            0.13368362 = fieldWeight in 1626, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=1626)
        0.117290854 = weight(_text_:semantic in 1626) [ClassicSimilarity], result of:
          0.117290854 = score(doc=1626,freq=22.0), product of:
            0.19245663 = queryWeight, product of:
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.04628742 = queryNorm
            0.60944045 = fieldWeight in 1626, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              4.1578603 = idf(docFreq=1879, maxDocs=44218)
              0.03125 = fieldNorm(doc=1626)
        0.08670825 = sum of:
          0.061623022 = weight(_text_:web in 1626) [ClassicSimilarity], result of:
            0.061623022 = score(doc=1626,freq=16.0), product of:
              0.15105948 = queryWeight, product of:
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.04628742 = queryNorm
              0.4079388 = fieldWeight in 1626, product of:
                4.0 = tf(freq=16.0), with freq of:
                  16.0 = termFreq=16.0
                3.2635105 = idf(docFreq=4597, maxDocs=44218)
                0.03125 = fieldNorm(doc=1626)
          0.025085226 = weight(_text_:22 in 1626) [ClassicSimilarity], result of:
            0.025085226 = score(doc=1626,freq=2.0), product of:
              0.16209066 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04628742 = queryNorm
              0.15476047 = fieldWeight in 1626, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1626)
      0.6 = coord(3/5)
    
    Abstract
    Purpose - The growing volumes of semantic data available in the web result in the need for handling the information overload phenomenon. The potential of this amount of data is enormous but in most cases it is very difficult for users to visualize, explore and use this data, especially for lay-users without experience with Semantic Web technologies. The paper aims to discuss these issues. Design/methodology/approach - The Visual Information-Seeking Mantra "Overview first, zoom and filter, then details-on-demand" proposed by Shneiderman describes how data should be presented in different stages to achieve an effective exploration. The overview is the first user task when dealing with a data set. The objective is that the user is capable of getting an idea about the overall structure of the data set. Different information architecture (IA) components supporting the overview tasks have been developed, so they are automatically generated from semantic data, and evaluated with end-users. Findings - The chosen IA components are well known to web users, as they are present in most web pages: navigation bars, site maps and site indexes. The authors complement them with Treemaps, a visualization technique for displaying hierarchical data. These components have been developed following an iterative User-Centered Design methodology. Evaluations with end-users have shown that they get easily used to them despite the fact that they are generated automatically from structured data, without requiring knowledge about the underlying semantic technologies, and that the different overview components complement each other as they focus on different information search needs. Originality/value - Obtaining semantic data sets overviews cannot be easily done with the current semantic web browsers. Overviews become difficult to achieve with large heterogeneous data sets, which is typical in the Semantic Web, because traditional IA techniques do not easily scale to large data sets. There is little or no support to obtain overview information quickly and easily at the beginning of the exploration of a new data set. This can be a serious limitation when exploring a data set for the first time, especially for lay-users. The proposal is to reuse and adapt existing IA components to provide this overview to users and show that they can be generated automatically from the thesaurus and ontologies that structure semantic data while providing a comparable user experience to traditional web sites.
    Date
    20. 1.2015 18:30:22
    Series
    Special issue: Semantic search
    Theme
    Semantic Web
    Semantisches Umfeld in Indexierung u. Retrieval

Authors

Years

Languages

Types

  • a 8225
  • m 649
  • el 518
  • s 374
  • r 67
  • b 40
  • x 34
  • n 25
  • i 16
  • p 16
  • d 9
  • ? 2
  • A 1
  • EL 1
  • h 1
  • l 1
  • More… Less…

Themes

Subjects

Classifications