Search (165 results, page 1 of 9)

  • × theme_ss:"Wissensrepräsentation"
  1. Chaudhury, S.; Mallik, A.; Ghosh, H.: Multimedia ontology : representation and applications (2016) 0.09
    0.08867782 = product of:
      0.13301674 = sum of:
        0.10910148 = weight(_text_:book in 2801) [ClassicSimilarity], result of:
          0.10910148 = score(doc=2801,freq=8.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.4876966 = fieldWeight in 2801, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2801)
        0.023915261 = product of:
          0.047830522 = sum of:
            0.047830522 = weight(_text_:search in 2801) [ClassicSimilarity], result of:
              0.047830522 = score(doc=2801,freq=4.0), product of:
                0.17614716 = queryWeight, product of:
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.050679956 = queryNorm
                0.27153727 = fieldWeight in 2801, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2801)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The book covers multimedia ontology in heritage preservation with intellectual explorations of various themes of Indian cultural heritage. The result of more than 15 years of collective research, Multimedia Ontology: Representation and Applications provides a theoretical foundation for understanding the nature of media data and the principles involved in its interpretation. The book presents a unified approach to recent advances in multimedia and explains how a multimedia ontology can fill the semantic gap between concepts and the media world. It relays real-life examples of implementations in different domains to illustrate how this gap can be filled. The book contains information that helps with building semantic, content-based search and retrieval engines and also with developing vertical application-specific search applications. It guides you in designing multimedia tools that aid in logical and conceptual organization of large amounts of multimedia data. As a practical demonstration, it showcases multimedia applications in cultural heritage preservation efforts and the creation of virtual museums. The book describes the limitations of existing ontology techniques in semantic multimedia data processing, as well as some open problems in the representations and applications of multimedia ontology. As an antidote, it introduces new ontology representation and reasoning schemes that overcome these limitations. The long, compiled efforts reflected in Multimedia Ontology: Representation and Applications are a signpost for new achievements and developments in efficiency and accessibility in the field.
  2. Gödert, W.; Hubrich, J.; Nagelschmidt, M.: Semantic knowledge representation for information retrieval (2014) 0.08
    0.07545 = product of:
      0.113175 = sum of:
        0.09257569 = weight(_text_:book in 987) [ClassicSimilarity], result of:
          0.09257569 = score(doc=987,freq=4.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.41382432 = fieldWeight in 987, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.046875 = fieldNorm(doc=987)
        0.020599304 = product of:
          0.041198608 = sum of:
            0.041198608 = weight(_text_:22 in 987) [ClassicSimilarity], result of:
              0.041198608 = score(doc=987,freq=2.0), product of:
                0.17747258 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050679956 = queryNorm
                0.23214069 = fieldWeight in 987, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=987)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This book covers the basics of semantic web technologies and indexing languages, and describes their contribution to improve languages as a tool for subject queries and knowledge exploration. The book is relevant to information scientists, knowledge workers and indexers. It provides a suitable combination of theoretical foundations and practical applications.
    Date
    23. 7.2017 13:49:22
  3. Nagao, M.: Knowledge and inference (1990) 0.07
    0.07426354 = product of:
      0.11139531 = sum of:
        0.094484664 = weight(_text_:book in 3304) [ClassicSimilarity], result of:
          0.094484664 = score(doc=3304,freq=6.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.42235768 = fieldWeight in 3304, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3304)
        0.016910642 = product of:
          0.033821285 = sum of:
            0.033821285 = weight(_text_:search in 3304) [ClassicSimilarity], result of:
              0.033821285 = score(doc=3304,freq=2.0), product of:
                0.17614716 = queryWeight, product of:
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.050679956 = queryNorm
                0.19200584 = fieldWeight in 3304, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3304)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Knowledge and Inference discusses an important problem for software systems: How do we treat knowledge and ideas on a computer and how do we use inference to solve problems on a computer? The book talks about the problems of knowledge and inference for the purpose of merging artificial intelligence and library science. The book begins by clarifying the concept of ""knowledge"" from many points of view, followed by a chapter on the current state of library science and the place of artificial intelligence in library science. Subsequent chapters cover central topics in the artificial intelligence: search and problem solving, methods of making proofs, and the use of knowledge in looking for a proof. There is also a discussion of how to use the knowledge system. The final chapter describes a popular expert system. It describes tools for building expert systems using an example based on Expert Systems-A Practical Introduction by P. Sell (Macmillian, 1985). This type of software is called an ""expert system shell."" This book was written as a textbook for undergraduate students covering only the basics but explaining as much detail as possible.
  4. Semantic applications (2018) 0.07
    0.07095766 = product of:
      0.10643649 = sum of:
        0.0771464 = weight(_text_:book in 5204) [ClassicSimilarity], result of:
          0.0771464 = score(doc=5204,freq=4.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.34485358 = fieldWeight in 5204, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
        0.029290091 = product of:
          0.058580182 = sum of:
            0.058580182 = weight(_text_:search in 5204) [ClassicSimilarity], result of:
              0.058580182 = score(doc=5204,freq=6.0), product of:
                0.17614716 = queryWeight, product of:
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.050679956 = queryNorm
                0.33256388 = fieldWeight in 5204, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5204)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This book describes proven methodologies for developing semantic applications: software applications which explicitly or implicitly uses the semantics (i.e., the meaning) of a domain terminology in order to improve usability, correctness, and completeness. An example is semantic search, where synonyms and related terms are used for enriching the results of a simple text-based search. Ontologies, thesauri or controlled vocabularies are the centerpiece of semantic applications. The book includes technological and architectural best practices for corporate use.
    Content
    Introduction.- Ontology Development.- Compliance using Metadata.- Variety Management for Big Data.- Text Mining in Economics.- Generation of Natural Language Texts.- Sentiment Analysis.- Building Concise Text Corpora from Web Contents.- Ontology-Based Modelling of Web Content.- Personalized Clinical Decision Support for Cancer Care.- Applications of Temporal Conceptual Semantic Systems.- Context-Aware Documentation in the Smart Factory.- Knowledge-Based Production Planning for Industry 4.0.- Information Exchange in Jurisdiction.- Supporting Automated License Clearing.- Managing cultural assets: Implementing typical cultural heritage archive's usage scenarios via Semantic Web technologies.- Semantic Applications for Process Management.- Domain-Specific Semantic Search Applications.
  5. ¬The Semantic Web : research and applications ; second European Semantic WebConference, ESWC 2005, Heraklion, Crete, Greece, May 29 - June 1, 2005 ; proceedings (2005) 0.06
    0.05716911 = product of:
      0.085753664 = sum of:
        0.0654609 = weight(_text_:book in 439) [ClassicSimilarity], result of:
          0.0654609 = score(doc=439,freq=2.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.29261798 = fieldWeight in 439, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.046875 = fieldNorm(doc=439)
        0.02029277 = product of:
          0.04058554 = sum of:
            0.04058554 = weight(_text_:search in 439) [ClassicSimilarity], result of:
              0.04058554 = score(doc=439,freq=2.0), product of:
                0.17614716 = queryWeight, product of:
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.050679956 = queryNorm
                0.230407 = fieldWeight in 439, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.046875 = fieldNorm(doc=439)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This book constitutes the refereed proceedings of the Second European Semantic Web Conference, ESWC 2005, heldin Heraklion, Crete, Greece in May/June 2005. The 48 revised full papers presented were carefully reviewed and selected from 148 submissions. The papers are organized in topical sections on semantic Web services, languages, ontologies, reasoning and querying, search and information retrieval, user and communities, natural language for the semantic Web, annotation tools, and semantic Web applications.
  6. King, B.E.; Reinold, K.: Finding the concept, not just the word : a librarian's guide to ontologies and semantics (2008) 0.06
    0.05692608 = product of:
      0.085389115 = sum of:
        0.056690805 = weight(_text_:book in 2863) [ClassicSimilarity], result of:
          0.056690805 = score(doc=2863,freq=6.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.25341463 = fieldWeight in 2863, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.0234375 = fieldNorm(doc=2863)
        0.02869831 = product of:
          0.05739662 = sum of:
            0.05739662 = weight(_text_:search in 2863) [ClassicSimilarity], result of:
              0.05739662 = score(doc=2863,freq=16.0), product of:
                0.17614716 = queryWeight, product of:
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.050679956 = queryNorm
                0.3258447 = fieldWeight in 2863, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=2863)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Aimed at students and professionals within Library and Information Services (LIS), this book is about the power and potential of ontologies to enhance the electronic search process. The book will compare search strategies and results in the current search environment and demonstrate how these could be transformed using ontologies and concept searching. Simple descriptions, visual representations, and examples of ontologies will bring a full understanding of how these concept maps are constructed to enhance retrieval through natural language queries. Readers will gain a sense of how ontologies are currently being used and how they could be applied in the future, encouraging them to think about how their own work and their users' search experiences could be enhanced by the creation of a customized ontology. Key Features Written by a librarian, for librarians (most work on ontologies is written and read by people in computer science and knowledge management) Written by a librarian who has created her own ontology and performed research on its capabilities Written in easily understandable language, with concepts broken down to the basics The Author Ms. King is the Information Specialist at the Center on Media and Child Health at Children's Hospital Boston. She is a graduate of Smith College (B.A.) and Simmons College (M.L.I.S.). She is an active member of the Special Libraries Association, and was the recipient of the 2005 SLA Innovation in Technology Award for the creation of a customized media effects ontology used for semantic searching. Readership The book is aimed at practicing librarians and information professionals as well as graduate students of Library and Information Science. Contents Introduction Part 1: Understanding Ontologies - organising knowledge; what is an ontology? How are ontologies different from other knowledge representations? How are ontologies currently being used? Key concepts Ontologies in semantic search - determining whether a search was successful; what does semantic search have to offer? Semantic techniques; semantic searching behind the scenes; key concepts Creating an ontology - how to create an ontology; key concepts Building an ontology from existing components - choosing components; customizing your knowledge structure; key concepts Part 2: Semantic Technologies Natural language processing - tagging parts of speech; grammar-based NLP; statistical NLP; semantic analysis,; current applications of NLP; key concepts Using metadata to add semantic information - structured languages; metadata tagging; semantic tagging; key concepts Other semantic capabilities - semantic classification; synsets; topic maps; rules and inference; key concepts Part 3: Case Studies: Theory into Practice Biogen Idec: using semantics in drug discovery research - Biogen Idec's solution; the future The Center on Media and Child Health: using an ontology to explore the effects of media - building the ontology; choosing the source; implementing and comparing to Boolean search; the future Partners HealthCare System: semantic technologies to improve clinical decision support - the medical appointment; partners healthcare system's solution; lessons learned; the future MINDSWAP: using ontologies to aid terrorism; intelligence gathering - building, using and maintaining the ontology; sharing information with other experts; future plans Part 4: Advanced Topics Languages for expressing ontologies - XML; RDF; OWL; SKOS; Ontology language features - comparison chart Tools for building ontologies - basic criteria when evaluating ontologies Part 5: Transitions to the Future
  7. Lassalle, E.; Lassalle, E.: Semantic models in information retrieval (2012) 0.05
    0.052310668 = product of:
      0.078466 = sum of:
        0.05455074 = weight(_text_:book in 97) [ClassicSimilarity], result of:
          0.05455074 = score(doc=97,freq=2.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.2438483 = fieldWeight in 97, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.0390625 = fieldNorm(doc=97)
        0.023915261 = product of:
          0.047830522 = sum of:
            0.047830522 = weight(_text_:search in 97) [ClassicSimilarity], result of:
              0.047830522 = score(doc=97,freq=4.0), product of:
                0.17614716 = queryWeight, product of:
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.050679956 = queryNorm
                0.27153727 = fieldWeight in 97, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=97)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Footnote
    Vgl.: http://www.igi-global.com/book/next-generation-search-engines/64424.
    Source
    Next generation search engines: advanced models for information retrieval. Eds.: C. Jouis, u.a
  8. Djioua, B.; Desclés, J.-P.; Alrahabi, M.: Searching and mining with semantic categories (2012) 0.05
    0.052310668 = product of:
      0.078466 = sum of:
        0.05455074 = weight(_text_:book in 99) [ClassicSimilarity], result of:
          0.05455074 = score(doc=99,freq=2.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.2438483 = fieldWeight in 99, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.0390625 = fieldNorm(doc=99)
        0.023915261 = product of:
          0.047830522 = sum of:
            0.047830522 = weight(_text_:search in 99) [ClassicSimilarity], result of:
              0.047830522 = score(doc=99,freq=4.0), product of:
                0.17614716 = queryWeight, product of:
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.050679956 = queryNorm
                0.27153727 = fieldWeight in 99, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=99)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Footnote
    Vgl.: http://www.igi-global.com/book/next-generation-search-engines/64423.
    Source
    Next generation search engines: advanced models for information retrieval. Eds.: C. Jouis, u.a
  9. Bringsjord, S.; Clark, M.; Taylor, J.: Sophisticated knowledge representation and reasoning requires philosophy (2014) 0.05
    0.04781122 = product of:
      0.07171683 = sum of:
        0.05455074 = weight(_text_:book in 3403) [ClassicSimilarity], result of:
          0.05455074 = score(doc=3403,freq=2.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.2438483 = fieldWeight in 3403, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3403)
        0.017166087 = product of:
          0.034332175 = sum of:
            0.034332175 = weight(_text_:22 in 3403) [ClassicSimilarity], result of:
              0.034332175 = score(doc=3403,freq=2.0), product of:
                0.17747258 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050679956 = queryNorm
                0.19345059 = fieldWeight in 3403, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3403)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    What is knowledge representation and reasoning (KR&R)? Alas, a thorough account would require a book, or at least a dedicated, full-length paper, but here we shall have to make do with something simpler. Since most readers are likely to have an intuitive grasp of the essence of KR&R, our simple account should suffice. The interesting thing is that this simple account itself makes reference to some of the foundational distinctions in the field of philosophy. These distinctions also play a central role in artificial intelligence (AI) and computer science. To begin with, the first distinction in KR&R is that we identify knowledge with knowledge that such-and-such holds (possibly to a degree), rather than knowing how. If you ask an expert tennis player how he manages to serve a ball at 130 miles per hour on his first serve, and then serve a safer, topspin serve on his second should the first be out, you may well receive a confession that, if truth be told, this athlete can't really tell you. He just does it; he does something he has been doing since his youth. Yet, there is no denying that he knows how to serve. In contrast, the knowledge in KR&R must be expressible in declarative statements. For example, our tennis player knows that if his first serve lands outside the service box, it's not in play. He thus knows a proposition, conditional in form.
    Date
    9. 2.2017 19:22:14
  10. Xiong, C.: Knowledge based text representations for information retrieval (2016) 0.04
    0.044793766 = product of:
      0.06719065 = sum of:
        0.053662136 = product of:
          0.16098641 = sum of:
            0.16098641 = weight(_text_:3a in 5820) [ClassicSimilarity], result of:
              0.16098641 = score(doc=5820,freq=2.0), product of:
                0.42966524 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050679956 = queryNorm
                0.3746787 = fieldWeight in 5820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.33333334 = coord(1/3)
        0.013528514 = product of:
          0.027057027 = sum of:
            0.027057027 = weight(_text_:search in 5820) [ClassicSimilarity], result of:
              0.027057027 = score(doc=5820,freq=2.0), product of:
                0.17614716 = queryWeight, product of:
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.050679956 = queryNorm
                0.15360467 = fieldWeight in 5820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The successes of information retrieval (IR) in recent decades were built upon bag-of-words representations. Effective as it is, bag-of-words is only a shallow text understanding; there is a limited amount of information for document ranking in the word space. This dissertation goes beyond words and builds knowledge based text representations, which embed the external and carefully curated information from knowledge bases, and provide richer and structured evidence for more advanced information retrieval systems. This thesis research first builds query representations with entities associated with the query. Entities' descriptions are used by query expansion techniques that enrich the query with explanation terms. Then we present a general framework that represents a query with entities that appear in the query, are retrieved by the query, or frequently show up in the top retrieved documents. A latent space model is developed to jointly learn the connections from query to entities and the ranking of documents, modeling the external evidence from knowledge bases and internal ranking features cooperatively. To further improve the quality of relevant entities, a defining factor of our query representations, we introduce learning to rank to entity search and retrieve better entities from knowledge bases. In the document representation part, this thesis research also moves one step forward with a bag-of-entities model, in which documents are represented by their automatic entity annotations, and the ranking is performed in the entity space.
    Content
    Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Language and Information Technologies. Vgl.: https%3A%2F%2Fwww.cs.cmu.edu%2F~cx%2Fpapers%2Fknowledge_based_text_representation.pdf&usg=AOvVaw0SaTSvhWLTh__Uz_HtOtl3.
  11. Reimer, U.; Brockhausen, P.; Lau, T.; Reich, J.R.: Ontology-based knowledge management at work : the Swiss life case studies (2004) 0.04
    0.041848537 = product of:
      0.0627728 = sum of:
        0.043640595 = weight(_text_:book in 4411) [ClassicSimilarity], result of:
          0.043640595 = score(doc=4411,freq=2.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.19507864 = fieldWeight in 4411, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.03125 = fieldNorm(doc=4411)
        0.019132208 = product of:
          0.038264416 = sum of:
            0.038264416 = weight(_text_:search in 4411) [ClassicSimilarity], result of:
              0.038264416 = score(doc=4411,freq=4.0), product of:
                0.17614716 = queryWeight, product of:
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.050679956 = queryNorm
                0.21722981 = fieldWeight in 4411, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.475677 = idf(docFreq=3718, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4411)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This chapter describes two case studies conducted by the Swiss Life insurance group with the objective of proving the practical applicability and superiority of ontology-based knowledge management over classical approaches based on text retrieval technologies. The first case study in the domain of skills management uses manually constructed ontologies about skills, job functions and education. The purpose of the system is to give support for finding employees with certain skills. The ontologies are used to ensure that the user description of skills and the machine-held index of skills and people use the same vocabulary. The use of a shared vocabulary increases the performance of such a system significantly. The second case study aims at improving content-oriented access to passages of a 1000 page document about the International Accounting Standard on the corporate intranet. To this end, an ontology was automatically extracted from the document. It can be used to reformulate queries that turned out not to deliver the intended results. Since the ontology was automatically built, it is of a rather simple structure, consisting of weighted semantic associations between the relevant concepts in the document. We therefore call it a 'lightweight ontology'. The two case studies cover quite different aspects of using ontologies in knowledge management applications. Whereas in the second case study an ontology was automatically derived from a search space to improve information retrieval, in the first skills management case study the ontology itself introduces a structured search space. In one case study we gathered experience in building an ontology manually, while the challenge of the other case study was automatic ontology creation. A number of the novel Semantic Web-based tools described elsewhere in this book were used to build the two systems and both case studies described have led to projects to deploy live systems within Swiss Life.
  12. Giunchiglia, F.; Villafiorita, A.; Walsh, T.: Theories of abstraction (1997) 0.04
    0.036348514 = product of:
      0.109045535 = sum of:
        0.109045535 = sum of:
          0.054114055 = weight(_text_:search in 4476) [ClassicSimilarity], result of:
            0.054114055 = score(doc=4476,freq=2.0), product of:
              0.17614716 = queryWeight, product of:
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.050679956 = queryNorm
              0.30720934 = fieldWeight in 4476, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.0625 = fieldNorm(doc=4476)
          0.054931477 = weight(_text_:22 in 4476) [ClassicSimilarity], result of:
            0.054931477 = score(doc=4476,freq=2.0), product of:
              0.17747258 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050679956 = queryNorm
              0.30952093 = fieldWeight in 4476, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=4476)
      0.33333334 = coord(1/3)
    
    Abstract
    Describes the types of representations used in different theories of abstractions. Shows how the type of mapping between these representations has been increasingly generalised. Discusses desirable properties preserved by such mappings and identifies how these properties are influenced by the mappings and the presentations defined. Surveys programs made in understanding the complexity reduction associated with abstraction. Focuses on formal models of how abstraction reduces the search space. Presents some of the systems that implement abstraction. shows how the efforts in this area have focused on the mechanisation of languages for the declarative representation of abstraction.
    Date
    1.10.2018 14:13:22
  13. Frické, M.: Logic and the organization of information (2012) 0.04
    0.036001656 = product of:
      0.108004965 = sum of:
        0.108004965 = weight(_text_:book in 1782) [ClassicSimilarity], result of:
          0.108004965 = score(doc=1782,freq=16.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.48279503 = fieldWeight in 1782, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1782)
      0.33333334 = coord(1/3)
    
    Abstract
    Logic and the Organization of Information closely examines the historical and contemporary methodologies used to catalogue information objects-books, ebooks, journals, articles, web pages, images, emails, podcasts and more-in the digital era. This book provides an in-depth technical background for digital librarianship, and covers a broad range of theoretical and practical topics including: classification theory, topic annotation, automatic clustering, generalized synonymy and concept indexing, distributed libraries, semantic web ontologies and Simple Knowledge Organization System (SKOS). It also analyzes the challenges facing today's information architects, and outlines a series of techniques for overcoming them. Logic and the Organization of Information is intended for practitioners and professionals working at a design level as a reference book for digital librarianship. Advanced-level students, researchers and academics studying information science, library science, digital libraries and computer science will also find this book invaluable.
    Footnote
    Rez. in: J. Doc. 70(2014) no.4: "Books on the organization of information and knowledge, aimed at a library/information audience, tend to fall into two clear categories. Most are practical and pragmatic, explaining the "how" as much or more than the "why". Some are theoretical, in part or in whole, showing how the practice of classification, indexing, resource description and the like relates to philosophy, logic, and other foundational bases; the books by Langridge (1992) and by Svenonious (2000) are well-known examples this latter kind. To this category certainly belongs a recent book by Martin Frické (2012). The author takes the reader for an extended tour through a variety of aspects of information organization, including classification and taxonomy, alphabetical vocabularies and indexing, cataloguing and FRBR, and aspects of the semantic web. The emphasis throughout is on showing how practice is, or should be, underpinned by formal structures; there is a particular emphasis on first order predicate calculus. The advantages of a greater, and more explicit, use of symbolic logic is a recurring theme of the book. There is a particularly commendable historical dimension, often omitted in texts on this subject. It cannot be said that this book is entirely an easy read, although it is well written with a helpful index, and its arguments are generally well supported by clear and relevant examples. It is thorough and detailed, but thereby seems better geared to the needs of advanced students and researchers than to the practitioners who are suggested as a main market. For graduate students in library/information science and related disciplines, in particular, this will be a valuable resource. I would place it alongside Svenonious' book as the best insight into the theoretical "why" of information organization. It has evoked a good deal of interest, including a set of essay commentaries in Journal of Information Science (Gilchrist et al., 2013). Introducing these, Alan Gilchrist rightly says that Frické deserves a salute for making explicit the fundamental relationship between the ancient discipline of logic and modern information organization. If information science is to continue to develop, and make a contribution to the organization of the information environments of the future, then this book sets the groundwork for the kind of studies which will be needed." (D. Bawden)
  14. Thenmalar, S.; Geetha, T.V.: Enhanced ontology-based indexing and searching (2014) 0.03
    0.034184426 = product of:
      0.10255328 = sum of:
        0.10255328 = sum of:
          0.07852076 = weight(_text_:search in 1633) [ClassicSimilarity], result of:
            0.07852076 = score(doc=1633,freq=22.0), product of:
              0.17614716 = queryWeight, product of:
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.050679956 = queryNorm
              0.44576794 = fieldWeight in 1633, product of:
                4.690416 = tf(freq=22.0), with freq of:
                  22.0 = termFreq=22.0
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.02734375 = fieldNorm(doc=1633)
          0.02403252 = weight(_text_:22 in 1633) [ClassicSimilarity], result of:
            0.02403252 = score(doc=1633,freq=2.0), product of:
              0.17747258 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050679956 = queryNorm
              0.1354154 = fieldWeight in 1633, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.02734375 = fieldNorm(doc=1633)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The purpose of this paper is to improve the conceptual-based search by incorporating structural ontological information such as concepts and relations. Generally, Semantic-based information retrieval aims to identify relevant information based on the meanings of the query terms or on the context of the terms and the performance of semantic information retrieval is carried out through standard measures-precision and recall. Higher precision leads to the (meaningful) relevant documents obtained and lower recall leads to the less coverage of the concepts. Design/methodology/approach - In this paper, the authors enhance the existing ontology-based indexing proposed by Kohler et al., by incorporating sibling information to the index. The index designed by Kohler et al., contains only super and sub-concepts from the ontology. In addition, in our approach, we focus on two tasks; query expansion and ranking of the expanded queries, to improve the efficiency of the ontology-based search. The aforementioned tasks make use of ontological concepts, and relations existing between those concepts so as to obtain semantically more relevant search results for a given query. Findings - The proposed ontology-based indexing technique is investigated by analysing the coverage of concepts that are being populated in the index. Here, we introduce a new measure called index enhancement measure, to estimate the coverage of ontological concepts being indexed. We have evaluated the ontology-based search for the tourism domain with the tourism documents and tourism-specific ontology. The comparison of search results based on the use of ontology "with and without query expansion" is examined to estimate the efficiency of the proposed query expansion task. The ranking is compared with the ORank system to evaluate the performance of our ontology-based search. From these analyses, the ontology-based search results shows better recall when compared to the other concept-based search systems. The mean average precision of the ontology-based search is found to be 0.79 and the recall is found to be 0.65, the ORank system has the mean average precision of 0.62 and the recall is found to be 0.51, while the concept-based search has the mean average precision of 0.56 and the recall is found to be 0.42. Practical implications - When the concept is not present in the domain-specific ontology, the concept cannot be indexed. When the given query term is not available in the ontology then the term-based results are retrieved. Originality/value - In addition to super and sub-concepts, we incorporate the concepts present in same level (siblings) to the ontological index. The structural information from the ontology is determined for the query expansion. The ranking of the documents depends on the type of the query (single concept query, multiple concept queries and concept with relation queries) and the ontological relations that exists in the query and the documents. With this ontological structural information, the search results showed us better coverage of concepts with respect to the query.
    Date
    20. 1.2015 18:30:22
  15. Mayfield, J.; Finin, T.: Information retrieval on the Semantic Web : integrating inference and retrieval 0.03
    0.03180495 = product of:
      0.09541484 = sum of:
        0.09541484 = sum of:
          0.0473498 = weight(_text_:search in 4330) [ClassicSimilarity], result of:
            0.0473498 = score(doc=4330,freq=2.0), product of:
              0.17614716 = queryWeight, product of:
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.050679956 = queryNorm
              0.2688082 = fieldWeight in 4330, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.0546875 = fieldNorm(doc=4330)
          0.04806504 = weight(_text_:22 in 4330) [ClassicSimilarity], result of:
            0.04806504 = score(doc=4330,freq=2.0), product of:
              0.17747258 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050679956 = queryNorm
              0.2708308 = fieldWeight in 4330, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=4330)
      0.33333334 = coord(1/3)
    
    Abstract
    One vision of the Semantic Web is that it will be much like the Web we know today, except that documents will be enriched by annotations in machine understandable markup. These annotations will provide metadata about the documents as well as machine interpretable statements capturing some of the meaning of document content. We discuss how the information retrieval paradigm might be recast in such an environment. We suggest that retrieval can be tightly bound to inference. Doing so makes today's Web search engines useful to Semantic Web inference engines, and causes improvements in either retrieval or inference to lead directly to improvements in the other.
    Date
    12. 2.2011 17:35:22
  16. Kiren, T.; Shoaib, M.: ¬A novel ontology matching approach using key concepts (2016) 0.03
    0.030970788 = product of:
      0.09291236 = sum of:
        0.09291236 = sum of:
          0.058580182 = weight(_text_:search in 2589) [ClassicSimilarity], result of:
            0.058580182 = score(doc=2589,freq=6.0), product of:
              0.17614716 = queryWeight, product of:
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.050679956 = queryNorm
              0.33256388 = fieldWeight in 2589, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2589)
          0.034332175 = weight(_text_:22 in 2589) [ClassicSimilarity], result of:
            0.034332175 = score(doc=2589,freq=2.0), product of:
              0.17747258 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050679956 = queryNorm
              0.19345059 = fieldWeight in 2589, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2589)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose Ontologies are used to formally describe the concepts within a domain in a machine-understandable way. Matching of heterogeneous ontologies is often essential for many applications like semantic annotation, query answering or ontology integration. Some ontologies may include a large number of entities which make the ontology matching process very complex in terms of the search space and execution time requirements. The purpose of this paper is to present a technique for finding degree of similarity between ontologies that trims down the search space by eliminating the ontology concepts that have less likelihood of being matched. Design/methodology/approach Algorithms are written for finding key concepts, concept matching and relationship matching. WordNet is used for solving synonym problems during the matching process. The technique is evaluated using the reference alignments between ontologies from ontology alignment evaluation initiative benchmark in terms of degree of similarity, Pearson's correlation coefficient and IR measures precision, recall and F-measure. Findings Positive correlation between the degree of similarity and degree of similarity (reference alignment) and computed values of precision, recall and F-measure showed that if only key concepts of ontologies are compared, a time and search space efficient ontology matching system can be developed. Originality/value On the basis of the present novel approach for ontology matching, it is concluded that using key concepts for ontology matching gives comparable results in reduced time and space.
    Date
    20. 1.2015 18:30:22
  17. Davies, J.; Fensel, D.; Harmelen, F. van: Conclusions: ontology-driven knowledge management : towards the Semantic Web? (2004) 0.03
    0.029093731 = product of:
      0.08728119 = sum of:
        0.08728119 = weight(_text_:book in 4407) [ClassicSimilarity], result of:
          0.08728119 = score(doc=4407,freq=2.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.39015728 = fieldWeight in 4407, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.0625 = fieldNorm(doc=4407)
      0.33333334 = coord(1/3)
    
    Abstract
    The global economy is rapidly becoming more and more knowledge intensive. Knowledge is now widely recognized as the fourth production factor, on an equal footing with the traditional production factors of labour, capital and materials. Managing knowledge is as important as the traditional management of labour, capital and materials. In this book, we have shown how Semantic Web technology can make an important contribution to knowledge management.
  18. Arp, R.; Smith, B.; Spear, A.D.: Building ontologies with basic formal ontology (2015) 0.03
    0.029093731 = product of:
      0.08728119 = sum of:
        0.08728119 = weight(_text_:book in 3444) [ClassicSimilarity], result of:
          0.08728119 = score(doc=3444,freq=8.0), product of:
            0.2237077 = queryWeight, product of:
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.050679956 = queryNorm
            0.39015728 = fieldWeight in 3444, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              4.414126 = idf(docFreq=1454, maxDocs=44218)
              0.03125 = fieldNorm(doc=3444)
      0.33333334 = coord(1/3)
    
    Abstract
    In the era of "big data," science is increasingly information driven, and the potential for computers to store, manage, and integrate massive amounts of data has given rise to such new disciplinary fields as biomedical informatics. Applied ontology offers a strategy for the organization of scientific information in computer-tractable form, drawing on concepts not only from computer and information science but also from linguistics, logic, and philosophy. This book provides an introduction to the field of applied ontology that is of particular relevance to biomedicine, covering theoretical components of ontologies, best practices for ontology design, and examples of biomedical ontologies in use. After defining an ontology as a representation of the types of entities in a given domain, the book distinguishes between different kinds of ontologies and taxonomies, and shows how applied ontology draws on more traditional ideas from metaphysics. It presents the core features of the Basic Formal Ontology (BFO), now used by over one hundred ontology projects around the world, and offers examples of domain ontologies that utilize BFO. The book also describes Web Ontology Language (OWL), a common framework for Semantic Web technologies. Throughout, the book provides concrete recommendations for the design and construction of domain ontologies.
  19. Renear, A.H.; Wickett, K.M.; Urban, R.J.; Dubin, D.; Shreeves, S.L.: Collection/item metadata relationships (2008) 0.03
    0.027261382 = product of:
      0.081784144 = sum of:
        0.081784144 = sum of:
          0.04058554 = weight(_text_:search in 2623) [ClassicSimilarity], result of:
            0.04058554 = score(doc=2623,freq=2.0), product of:
              0.17614716 = queryWeight, product of:
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.050679956 = queryNorm
              0.230407 = fieldWeight in 2623, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.046875 = fieldNorm(doc=2623)
          0.041198608 = weight(_text_:22 in 2623) [ClassicSimilarity], result of:
            0.041198608 = score(doc=2623,freq=2.0), product of:
              0.17747258 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050679956 = queryNorm
              0.23214069 = fieldWeight in 2623, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2623)
      0.33333334 = coord(1/3)
    
    Abstract
    Contemporary retrieval systems, which search across collections, usually ignore collection-level metadata. Alternative approaches, exploiting collection-level information, will require an understanding of the various kinds of relationships that can obtain between collection-level and item-level metadata. This paper outlines the problem and describes a project that is developing a logic-based framework for classifying collection/item metadata relationships. This framework will support (i) metadata specification developers defining metadata elements, (ii) metadata creators describing objects, and (iii) system designers implementing systems that take advantage of collection-level metadata. We present three examples of collection/item metadata relationship categories, attribute/value-propagation, value-propagation, and value-constraint and show that even in these simple cases a precise formulation requires modal notions in addition to first-order logic. These formulations are related to recent work in information retrieval and ontology evaluation.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  20. Hollink, L.; Assem, M. van: Estimating the relevance of search results in the Culture-Web : a study of semantic distance measures (2010) 0.03
    0.027261382 = product of:
      0.081784144 = sum of:
        0.081784144 = sum of:
          0.04058554 = weight(_text_:search in 4649) [ClassicSimilarity], result of:
            0.04058554 = score(doc=4649,freq=2.0), product of:
              0.17614716 = queryWeight, product of:
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.050679956 = queryNorm
              0.230407 = fieldWeight in 4649, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.475677 = idf(docFreq=3718, maxDocs=44218)
                0.046875 = fieldNorm(doc=4649)
          0.041198608 = weight(_text_:22 in 4649) [ClassicSimilarity], result of:
            0.041198608 = score(doc=4649,freq=2.0), product of:
              0.17747258 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050679956 = queryNorm
              0.23214069 = fieldWeight in 4649, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=4649)
      0.33333334 = coord(1/3)
    
    Date
    26.12.2011 13:40:22

Authors

Years

Languages

  • e 147
  • d 13
  • el 1
  • f 1
  • More… Less…

Types

  • a 110
  • el 47
  • m 22
  • x 10
  • s 7
  • n 3
  • p 1
  • r 1
  • More… Less…

Subjects