Search (80 results, page 1 of 4)

  • × theme_ss:"Wissensrepräsentation"
  • × type_ss:"a"
  1. Kleineberg, M.: ¬The blind men and the elephant : towards an organization of epistemic contexts (2013) 0.08
    0.07873234 = product of:
      0.11809851 = sum of:
        0.08966068 = weight(_text_:systematic in 1074) [ClassicSimilarity], result of:
          0.08966068 = score(doc=1074,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.31573826 = fieldWeight in 1074, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1074)
        0.028437834 = product of:
          0.05687567 = sum of:
            0.05687567 = weight(_text_:indexing in 1074) [ClassicSimilarity], result of:
              0.05687567 = score(doc=1074,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29905218 = fieldWeight in 1074, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1074)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    In the last two decades of knowledge organization (KO) research, there has been an increasing interest in the context-dependent nature of human knowledge. Contextualism maintains that knowledge is not available in a neutral and objective way, but is always interwoven with the process of knowledge production and the prerequisites of the knower. As a first step towards a systematic organization of epistemic contexts, the concept of knowledge will be considered in its ontological (WHAT) and epistemological (WHO) including methodological (HOW) dimensions. In current KO research, however, either the contextualism is not fully implemented (classification-as-ontology) or the ambition for a context-transcending universal KOS seems to have been abandoned (classification-as-epistemology). Based on a combined ontology and epistemology it will be argued that a phenomena-based approach to KO as stipulated by the León Manifesto, for example, requires a revision of the underlying phenomenon concept as a relation between the known object (WHAT) and the knowing subject (WHO), which is constituted by the application of specific methods (HOW). While traditional subject indexing of documents often relies on the organizing principle "levels of being" (WHAT), for a future context indexing, two novel principles are proposed, namely "levels of knowing" (WHO) and "integral methodological pluralism" (HOW).
  2. Broughton, V.: Facet analysis as a tool for modelling subject domains and terminologies (2011) 0.07
    0.07317951 = product of:
      0.10976927 = sum of:
        0.08966068 = weight(_text_:systematic in 4826) [ClassicSimilarity], result of:
          0.08966068 = score(doc=4826,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.31573826 = fieldWeight in 4826, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4826)
        0.020108584 = product of:
          0.04021717 = sum of:
            0.04021717 = weight(_text_:indexing in 4826) [ClassicSimilarity], result of:
              0.04021717 = score(doc=4826,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.21146181 = fieldWeight in 4826, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4826)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Facet analysis is proposed as a general theory of knowledge organization, with an associated methodology that may be applied to the development of terminology tools in a variety of contexts and formats. Faceted classifications originated as a means of representing complexity in semantic content that facilitates logical organization and effective retrieval in a physical environment. This is achieved through meticulous analysis of concepts, their structural and functional status (based on fundamental categories), and their inter-relationships. These features provide an excellent basis for the general conceptual modelling of domains, and for the generation of KOS other than systematic classifications. This is demonstrated by the adoption of a faceted approach to many web search and visualization tools, and by the emergence of a facet based methodology for the construction of thesauri. Current work on the Bliss Bibliographic Classification (Second Edition) is investigating the ways in which the full complexity of faceted structures may be represented through encoded data, capable of generating intellectually and mechanically compatible forms of indexing tools from a single source. It is suggested that a number of research questions relating to the Semantic Web could be tackled through the medium of facet analysis.
  3. Cui, H.: Competency evaluation of plant character ontologies against domain literature (2010) 0.07
    0.07099311 = product of:
      0.10648966 = sum of:
        0.08966068 = weight(_text_:systematic in 3466) [ClassicSimilarity], result of:
          0.08966068 = score(doc=3466,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.31573826 = fieldWeight in 3466, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3466)
        0.016828977 = product of:
          0.033657953 = sum of:
            0.033657953 = weight(_text_:22 in 3466) [ClassicSimilarity], result of:
              0.033657953 = score(doc=3466,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.19345059 = fieldWeight in 3466, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3466)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Specimen identification keys are still the most commonly created tools used by systematic biologists to access biodiversity information. Creating identification keys requires analyzing and synthesizing large amounts of information from specimens and their descriptions and is a very labor-intensive and time-consuming activity. Automating the generation of identification keys from text descriptions becomes a highly attractive text mining application in the biodiversity domain. Fine-grained semantic annotation of morphological descriptions of organisms is a necessary first step in generating keys from text. Machine-readable ontologies are needed in this process because most biological characters are only implied (i.e., not stated) in descriptions. The immediate question to ask is How well do existing ontologies support semantic annotation and automated key generation? With the intention to either select an existing ontology or develop a unified ontology based on existing ones, this paper evaluates the coverage, semantic consistency, and inter-ontology agreement of a biodiversity character ontology and three plant glossaries that may be turned into ontologies. The coverage and semantic consistency of the ontology/glossaries are checked against the authoritative domain literature, namely, Flora of North America and Flora of China. The evaluation results suggest that more work is needed to improve the coverage and interoperability of the ontology/glossaries. More concepts need to be added to the ontology/glossaries and careful work is needed to improve the semantic consistency. The method used in this paper to evaluate the ontology/glossaries can be used to propose new candidate concepts from the domain literature and suggest appropriate definitions.
    Date
    1. 6.2010 9:55:22
  4. Zhitomirsky-Geffet, M.; Avidan, G.: ¬A new framework for systematic analysis and classification of inconsistencies in multi-viewpoint ontologies (2021) 0.05
    0.050719745 = product of:
      0.15215923 = sum of:
        0.15215923 = weight(_text_:systematic in 589) [ClassicSimilarity], result of:
          0.15215923 = score(doc=589,freq=4.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.5358256 = fieldWeight in 589, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.046875 = fieldNorm(doc=589)
      0.33333334 = coord(1/3)
    
    Abstract
    Plurality of beliefs and theories in different knowledge domains calls for modelling multi-viewpoint ontologies and knowledge organization systems (KOS). A generic theoretical approach recently proposed for heterogeneity representation in KOS was linking each ontological statement to a specific validity scope to determine a set of conditions under which the statement is valid. However, the practical applicability of this approach has yet to be empirically assessed. In addition, there is still a need to investigate the types of inconsistencies that might arise in multi-viewpoint ontologies as well as their possible causes. This study proposes a new framework for systematic analysis and classification of inconsistencies in multi-viewpoint ontologies. The framework is based on eight generic logical structures of ontological statements. To test the validity of the proposed framework, two ontologies from different knowledge domains were examined. We found that only three of the eight structures led to inconsistencies in both ontologies, while the other two structures were always present in logically consistent statements. The study has practical implications for building diversified and personalized knowledge systems.
  5. Sinha, P.K.; Dutta, B.: ¬A systematic analysis of flood ontologies : a parametric approach (2020) 0.04
    0.042266455 = product of:
      0.12679936 = sum of:
        0.12679936 = weight(_text_:systematic in 5758) [ClassicSimilarity], result of:
          0.12679936 = score(doc=5758,freq=4.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.44652134 = fieldWeight in 5758, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5758)
      0.33333334 = coord(1/3)
    
    Abstract
    The article identifies the core literature available on flood ontologies and presents a review on these ontologies from various perspectives like its purpose, type, design methodologies, ontologies (re)used, and also their focus on specific flood disaster phases. The study was conducted in two stages: i) literature identification, where the systematic literature review methodology was employed; and, ii) ontological review, where the parametric approach was applied. The study resulted in a set of fourteen papers discussing the flood ontology (FO). The ontological review revealed that most of the flood ontologies were task ontologies, formal, modular, and used web ontology language (OWL) for their representation. The most (re)used ontologies were SWEET, SSN, Time, and Space. METHONTOLOGY was the preferred design methodology, and for evaluation, application-based or data-based approaches were preferred. The majority of the ontologies were built around the response phase of the disaster. The unavailability of the full ontologies somewhat restricted the current study as the structural ontology metrics are missing. But the scientific community, the developers, of flood disaster management systems can refer to this work for their research to see what is available in the literature on flood ontology and the other major domains essential in building the FO.
  6. Park, J.-r.: Evolution of concept networks and implications for knowledge representation (2007) 0.03
    0.029886894 = product of:
      0.08966068 = sum of:
        0.08966068 = weight(_text_:systematic in 847) [ClassicSimilarity], result of:
          0.08966068 = score(doc=847,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.31573826 = fieldWeight in 847, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0390625 = fieldNorm(doc=847)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The purpose of this paper is to present descriptive characteristics of the historical development of concept networks. The linguistic principles, mechanisms and motivations behind the evolution of concept networks are discussed. Implications emanating from the idea of the historical development of concept networks are discussed in relation to knowledge representation and organization schemes. Design/methodology/approach - Natural language data including both speech and text are analyzed by examining discourse contexts in which a linguistic element such as a polysemy or homonym occurs. Linguistic literature on the historical development of concept networks is reviewed and analyzed. Findings - Semantic sense relations in concept networks can be captured in a systematic and regular manner. The mechanism and impetus behind the process of concept network development suggest that semantic senses in concept networks are closely intertwined with pragmatic contexts and discourse structure. The interrelation and permeability of the semantic senses of concept networks are captured on a continuum scale based on three linguistic parameters: concrete shared semantic sense; discourse and text structure; and contextualized pragmatic information. Research limitations/implications - Research findings signify the critical need for linking discourse structure and contextualized pragmatic information to knowledge representation and organization schemes. Originality/value - The idea of linguistic characteristics, principles, motivation and mechanisms underlying the evolution of concept networks provides theoretical ground for developing a model for integrating knowledge representation and organization schemes with discourse structure and contextualized pragmatic information.
  7. Hjoerland, B.: Semantics and knowledge organization (2007) 0.03
    0.029886894 = product of:
      0.08966068 = sum of:
        0.08966068 = weight(_text_:systematic in 1980) [ClassicSimilarity], result of:
          0.08966068 = score(doc=1980,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.31573826 = fieldWeight in 1980, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1980)
      0.33333334 = coord(1/3)
    
    Abstract
    The aim of this chapter is to demonstrate that semantic issues underlie all research questions within Library and Information Science (LIS, or, as hereafter, IS) and, in particular, the subfield known as Knowledge Organization (KO). Further, it seeks to show that semantics is a field influenced by conflicting views and discusses why it is important to argue for the most fruitful one of these. Moreover, the chapter demonstrates that IS has not yet addressed semantic problems in systematic fashion and examines why the field is very fragmented and without a proper theoretical basis. The focus here is on broad interdisciplinary issues and the long-term perspective. The theoretical problems involving semantics and concepts are very complicated. Therefore, this chapter starts by considering tools developed in KO for information retrieval (IR) as basically semantic tools. In this way, it establishes a specific IS focus on the relation between KO and semantics. It is well known that thesauri consist of a selection of concepts supplemented with information about their semantic relations (such as generic relations or "associative relations"). Some words in thesauri are "preferred terms" (descriptors), whereas others are "lead-in terms." The descriptors represent concepts. The difference between "a word" and "a concept" is that different words may have the same meaning and similar words may have different meanings, whereas one concept expresses one meaning.
  8. Broughton, V.: Language related problems in the construction of faceted terminologies and their automatic management (2008) 0.03
    0.029886894 = product of:
      0.08966068 = sum of:
        0.08966068 = weight(_text_:systematic in 2497) [ClassicSimilarity], result of:
          0.08966068 = score(doc=2497,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.31573826 = fieldWeight in 2497, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2497)
      0.33333334 = coord(1/3)
    
    Content
    The paper describes current work on the generation of a thesaurus format from the schedules of the Bliss Bibliographic Classification 2nd edition (BC2). The practical problems that occur in moving from a concept based approach to a terminological approach cluster around issues of vocabulary control that are not fully addressed in a systematic structure. These difficulties can be exacerbated within domains in the humanities because large numbers of culture specific terms may need to be accommodated in any thesaurus. The ways in which these problems can be resolved within the context of a semi-automated approach to the thesaurus generation have consequences for the management of classification data in the source vocabulary. The way in which the vocabulary is marked up for the purpose of machine manipulation is described, and some of the implications for editorial policy are discussed and examples given. The value of the classification notation as a language independent representation and mapping tool should not be sacrificed in such an exercise.
  9. Crystal, D.: Semantic targeting : past, present, and future (2010) 0.03
    0.029886894 = product of:
      0.08966068 = sum of:
        0.08966068 = weight(_text_:systematic in 3938) [ClassicSimilarity], result of:
          0.08966068 = score(doc=3938,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.31573826 = fieldWeight in 3938, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3938)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - This paper seeks to explicate the notion of "semantics", especially as it is being used in the context of the internet in general and advertising in particular. Design/methodology/approach - The conception of semantics as it evolved within linguistics is placed in its historical context. In the field of online advertising, it shows the limitations of keyword-based approaches and those where a limited amount of context is taken into account (contextual advertising). A more sophisticated notion of semantic targeting is explained, in which the whole page is taken into account in arriving at a semantic categorization. This is achieved through a combination of lexicological analysis and a purpose-built semantic taxonomy. Findings - The combination of a lexical analysis (derived from a dictionary) and a taxonomy (derived from a general encyclopedia, and subsequently refined) resulted in the construction of a "sense engine", which was then applied to online advertising, Examples of the application illustrate how relevance and sensitivity (brand protection) of ad placement can be improved. Several areas of potential further application are outlined. Originality/value - This is the first systematic application of linguistics to provide a solution to the problem of inappropriate ad placement online.
  10. Kohne, J.: Ontology, its origins and its meaning in information icience (2014) 0.03
    0.029886894 = product of:
      0.08966068 = sum of:
        0.08966068 = weight(_text_:systematic in 3401) [ClassicSimilarity], result of:
          0.08966068 = score(doc=3401,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.31573826 = fieldWeight in 3401, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3401)
      0.33333334 = coord(1/3)
    
    Abstract
    Ontology - in Aristotelian terms the science of being qua being - as a classical branch of philosophy describes the foundations of being in general. In this context, ontology is general metaphysics: the science of everything. Pursuing ontology means establishing some systematic order among the being, i.e. dividing things into categories or conceptual frameworks. Explaining the reasons why there are things or even anything, however, is part of what is called special metaphysics (theology, cosmology and psychology). If putting things into categories is the key issue of ontology, then general structures are its main level of analysis. To categorize things is to put them into a structural order. Such categorization of things enables one to understand what reality is about. If this is true, and characterizing the general structures of being is a reasonable access for us to reality, then two kinds of analysis of those structures are available: (i) realism and (ii) nominalism. In a realist (Aristotelian) ontology the general structures of being are understood as a kind of mirror reflecting things in their natural order. Those categories, as they are called in realism, then represent or show the structure of being. Ontological realism understands the relation between categories and being as a kind of correspondence or mapping which gives access to reality itself.
  11. Zeng, Q.; Yu, M.; Yu, W.; Xiong, J.; Shi, Y.; Jiang, M.: Faceted hierarchy : a new graph type to organize scientific concepts and a construction method (2019) 0.03
    0.026304156 = product of:
      0.07891247 = sum of:
        0.07891247 = product of:
          0.2367374 = sum of:
            0.2367374 = weight(_text_:3a in 400) [ClassicSimilarity], result of:
              0.2367374 = score(doc=400,freq=2.0), product of:
                0.4212274 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.049684696 = queryNorm
                0.56201804 = fieldWeight in 400, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=400)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Content
    Vgl.: https%3A%2F%2Faclanthology.org%2FD19-5317.pdf&usg=AOvVaw0ZZFyq5wWTtNTvNkrvjlGA.
  12. Thenmalar, S.; Geetha, T.V.: Enhanced ontology-based indexing and searching (2014) 0.02
    0.024107099 = product of:
      0.072321296 = sum of:
        0.072321296 = sum of:
          0.04876073 = weight(_text_:indexing in 1633) [ClassicSimilarity], result of:
            0.04876073 = score(doc=1633,freq=6.0), product of:
              0.19018644 = queryWeight, product of:
                3.8278677 = idf(docFreq=2614, maxDocs=44218)
                0.049684696 = queryNorm
              0.25638384 = fieldWeight in 1633, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                3.8278677 = idf(docFreq=2614, maxDocs=44218)
                0.02734375 = fieldNorm(doc=1633)
          0.023560567 = weight(_text_:22 in 1633) [ClassicSimilarity], result of:
            0.023560567 = score(doc=1633,freq=2.0), product of:
              0.17398734 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.049684696 = queryNorm
              0.1354154 = fieldWeight in 1633, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.02734375 = fieldNorm(doc=1633)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The purpose of this paper is to improve the conceptual-based search by incorporating structural ontological information such as concepts and relations. Generally, Semantic-based information retrieval aims to identify relevant information based on the meanings of the query terms or on the context of the terms and the performance of semantic information retrieval is carried out through standard measures-precision and recall. Higher precision leads to the (meaningful) relevant documents obtained and lower recall leads to the less coverage of the concepts. Design/methodology/approach - In this paper, the authors enhance the existing ontology-based indexing proposed by Kohler et al., by incorporating sibling information to the index. The index designed by Kohler et al., contains only super and sub-concepts from the ontology. In addition, in our approach, we focus on two tasks; query expansion and ranking of the expanded queries, to improve the efficiency of the ontology-based search. The aforementioned tasks make use of ontological concepts, and relations existing between those concepts so as to obtain semantically more relevant search results for a given query. Findings - The proposed ontology-based indexing technique is investigated by analysing the coverage of concepts that are being populated in the index. Here, we introduce a new measure called index enhancement measure, to estimate the coverage of ontological concepts being indexed. We have evaluated the ontology-based search for the tourism domain with the tourism documents and tourism-specific ontology. The comparison of search results based on the use of ontology "with and without query expansion" is examined to estimate the efficiency of the proposed query expansion task. The ranking is compared with the ORank system to evaluate the performance of our ontology-based search. From these analyses, the ontology-based search results shows better recall when compared to the other concept-based search systems. The mean average precision of the ontology-based search is found to be 0.79 and the recall is found to be 0.65, the ORank system has the mean average precision of 0.62 and the recall is found to be 0.51, while the concept-based search has the mean average precision of 0.56 and the recall is found to be 0.42. Practical implications - When the concept is not present in the domain-specific ontology, the concept cannot be indexed. When the given query term is not available in the ontology then the term-based results are retrieved. Originality/value - In addition to super and sub-concepts, we incorporate the concepts present in same level (siblings) to the ontological index. The structural information from the ontology is determined for the query expansion. The ranking of the documents depends on the type of the query (single concept query, multiple concept queries and concept with relation queries) and the ontological relations that exists in the query and the documents. With this ontological structural information, the search results showed us better coverage of concepts with respect to the query.
    Date
    20. 1.2015 18:30:22
  13. Herre, H.: General Formal Ontology (GFO) : a foundational ontology for conceptual modelling (2010) 0.02
    0.023909515 = product of:
      0.07172854 = sum of:
        0.07172854 = weight(_text_:systematic in 771) [ClassicSimilarity], result of:
          0.07172854 = score(doc=771,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.2525906 = fieldWeight in 771, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.03125 = fieldNorm(doc=771)
      0.33333334 = coord(1/3)
    
    Abstract
    Research in ontology has in recent years become widespread in the field of information systems, in distinct areas of sciences, in business, in economy, and in industry. The importance of ontologies is increasingly recognized in fields diverse as in e-commerce, semantic web, enterprise, information integration, qualitative modelling of physical systems, natural language processing, knowledge engineering, and databases. Ontologies provide formal specifications and harmonized definitions of concepts used to represent knowledge of specific domains. An ontology supplies a unifying framework for communication and establishes the basis of the knowledge about a specific domain. The term ontology has two meanings, it denotes, on the one hand, a research area, on the other hand, a system of organized knowledge. A system of knowledge may exhibit various degrees of formality; in the strongest sense it is an axiomatized and formally represented theory. which is denoted throughout this paper by the term axiomatized ontology. We use the term formal ontology to name an area of research which is becoming a science similar as formal or mathematical logic. Formal ontology is an evolving science which is concerned with the systematic development of axiomatic theories describing forms, modes, and views of being of the world at different levels of abstraction and granularity. Formal ontology combines the methods of mathematical logic with principles of philosophy, but also with the methods of artificial intelligence and linguistics. At themost general level of abstraction, formal ontology is concerned with those categories that apply to every area of the world. The application of formal ontology to domains at different levels of generality yields knowledge systems which are called, according to the level of abstraction, Top Level Ontologies or Foundational Ontologies, Core Domain or Domain Ontologies. Top level or foundational ontologies apply to every area of the world, in contrast to the various Generic, Domain Core or Domain Ontologies, which are associated to more restricted fields of interest. A foundational ontology can serve as a unifying framework for representation and integration of knowledge and may support the communication and harmonisation of conceptual systems. The current paper presents an overview about the current stage of the foundational ontology GFO.
  14. Gödert, W.: ¬An ontology-based model for indexing and retrieval (2013) 0.02
    0.02144916 = product of:
      0.064347476 = sum of:
        0.064347476 = product of:
          0.12869495 = sum of:
            0.12869495 = weight(_text_:indexing in 1510) [ClassicSimilarity], result of:
              0.12869495 = score(doc=1510,freq=8.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.6766778 = fieldWeight in 1510, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1510)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Starting from an unsolved problem of information retrieval this paper presents an ontology-based model for indexing and retrieval. The model combines the methods and experiences of cognitive-to-interpret indexing languages with the strengths and possibilities of formal knowledge representation. The core component of the model uses inferences along the paths of typed relations between the entities of a knowledge representation for enabling the determination of hit quantities in the context of retrieval processes. The entities are arranged in aspect-oriented facets to ensure a consistent hierarchical structure. The possible consequences for indexing and retrieval are discussed.
  15. Gödert, W.: ¬An ontology-based model for indexing and retrieval (2016) 0.02
    0.018768014 = product of:
      0.05630404 = sum of:
        0.05630404 = product of:
          0.11260808 = sum of:
            0.11260808 = weight(_text_:indexing in 2777) [ClassicSimilarity], result of:
              0.11260808 = score(doc=2777,freq=8.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.5920931 = fieldWeight in 2777, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2777)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The presented ontology-based model for indexing and retrieval combines the methods and experiences of traditional indexing languages with their cognitively interpreted entities and relationships with the strengths and possibilities of formal knowledge representation. The core component of the model uses inferences along the paths of typed relations between the entities of a knowledge representation for enabling the determination of result sets in the context of retrieval processes. A proposal for a general, but condensed, inventory of typed relations is given. The entities are arranged in aspect-oriented facets to ensure a consistent hierarchical structure. The possible consequences for indexing and retrieval are discussed.
  16. Wright, L.W.; Nardini, H.K.G.; Aronson, A.R.; Rindflesch, T.C.: Hierarchical concept indexing of full-text documents in the Unified Medical Language System Information sources Map (1999) 0.02
    0.016086869 = product of:
      0.048260607 = sum of:
        0.048260607 = product of:
          0.09652121 = sum of:
            0.09652121 = weight(_text_:indexing in 2111) [ClassicSimilarity], result of:
              0.09652121 = score(doc=2111,freq=8.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.5075084 = fieldWeight in 2111, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2111)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Full-text documents are a vital and rapidly growing part of online biomedical information. A single large document can contain as much information as a small database, but normally lacks the tight structure and consistent indexing of a database. Retrieval systems will often miss highly relevant parts of a document if the document as a whole appears irrelevant. Access to full-text information is further complicated by the need to search separately many disparate information resources. This research explores how these problems can be addressed by the combined use of 2 techniques: 1) natural language processing for automatic concept-based indexing of full text, and 2) methods for exploiting the structure and hierarchy of full-text documents. We describe methods for applying these techniques to a large collection of full-text documents drawn from the Health Services / Technology Assessment Text (HSTAT) database at the NLM and examine how this hierarchical concept indexing can assist both document- and source-level retrieval in the context of NLM's Information Source Map project
  17. Starostenko, O.; Rodríguez-Asomoza, J.; Sénchez-López, S.E.; Chévez-Aragón, J.A.: Shape indexing and retrieval : a hybrid approach using ontological description (2008) 0.01
    0.013931636 = product of:
      0.041794907 = sum of:
        0.041794907 = product of:
          0.083589815 = sum of:
            0.083589815 = weight(_text_:indexing in 4318) [ClassicSimilarity], result of:
              0.083589815 = score(doc=4318,freq=6.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.4395151 = fieldWeight in 4318, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4318)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This paper presents a novel hybrid approach for visual information retrieval (VIR) that combines shape analysis of objects in image with their indexing by textual descriptions. The principal goal of presented technique is applying Two Segments Turning Function (2STF) proposed by authors for efficient invariant to spatial variations shape processing and implementation of semantic Web approaches for ontology-based user-oriented annotations of multimedia information. In the proposed approach the user's textual queries are converted to image features, which are used for images searching, indexing, interpretation, and retrieval. A decision about similarity between retrieved image and user's query is taken computing the shape convergence to 2STF combining it with matching the ontological annotations of objects in image and providing in this way automatic definition of the machine-understandable semantics. In order to evaluate the proposed approach the Image Retrieval by Ontological Description of Shapes system has been designed and tested using some standard image domains.
  18. Buizza, G.: Subject analysis and indexing : an "Italian version" of the analytico-synthetic model (2011) 0.01
    0.013931636 = product of:
      0.041794907 = sum of:
        0.041794907 = product of:
          0.083589815 = sum of:
            0.083589815 = weight(_text_:indexing in 1812) [ClassicSimilarity], result of:
              0.083589815 = score(doc=1812,freq=6.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.4395151 = fieldWeight in 1812, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1812)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The paper presents the theoretical foundation of Italian indexing system. A consistent integration of vocabulary control through a thesaurus (semantics) and of role analysis to construct subject strings (syntax) allows to represent the full theme of a work, even if complex, in one string. The conceptual model produces a binary scheme: each aspect (entities, relationships, etc.) consists of a couple of elements, drawing the two lines of semantics and syntax. The meaning of 'concept' and 'theme' is analysed, also in comparison with the FRBR and FRSAD models, with the proposal of an en riched model. A double existence of concepts is suggested: document-independent adn document-dependent.
    Source
    Subject access: preparing for the future. Conference on August 20 - 21, 2009 in Florence, the IFLA Classification and Indexing Section sponsored an IFLA satellite conference entitled "Looking at the Past and Preparing for the Future". Eds.: P. Landry et al
  19. Kara, S.: ¬An ontology-based retrieval system using semantic indexing (2012) 0.01
    0.013931636 = product of:
      0.041794907 = sum of:
        0.041794907 = product of:
          0.083589815 = sum of:
            0.083589815 = weight(_text_:indexing in 3829) [ClassicSimilarity], result of:
              0.083589815 = score(doc=3829,freq=6.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.4395151 = fieldWeight in 3829, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3829)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    In this thesis, we present an ontology-based information extraction and retrieval system and its application to soccer domain. In general, we deal with three issues in semantic search, namely, usability, scalability and retrieval performance. We propose a keyword-based semantic retrieval approach. The performance of the system is improved considerably using domain-specific information extraction, inference and rules. Scalability is achieved by adapting a semantic indexing approach. The system is implemented using the state-of-the-art technologies in SemanticWeb and its performance is evaluated against traditional systems as well as the query expansion methods. Furthermore, a detailed evaluation is provided to observe the performance gain due to domain-specific information extraction and inference. Finally, we show how we use semantic indexing to solve simple structural ambiguities.
  20. Ma, N.; Zheng, H.T.; Xiao, X.: ¬An ontology-based latent semantic indexing approach using long short-term memory networks (2017) 0.01
    0.0134057235 = product of:
      0.04021717 = sum of:
        0.04021717 = product of:
          0.08043434 = sum of:
            0.08043434 = weight(_text_:indexing in 3810) [ClassicSimilarity], result of:
              0.08043434 = score(doc=3810,freq=8.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.42292362 = fieldWeight in 3810, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3810)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Nowadays, online data shows an astonishing increase and the issue of semantic indexing remains an open question. Ontologies and knowledge bases have been widely used to optimize performance. However, researchers are placing increased emphasis on internal relations of ontologies but neglect latent semantic relations between ontologies and documents. They generally annotate instances mentioned in documents, which are related to concepts in ontologies. In this paper, we propose an Ontology-based Latent Semantic Indexing approach utilizing Long Short-Term Memory networks (LSTM-OLSI). We utilize an importance-aware topic model to extract document-level semantic features and leverage ontologies to extract word-level contextual features. Then we encode the above two levels of features and match their embedding vectors utilizing LSTM networks. Finally, the experimental results reveal that LSTM-OLSI outperforms existing techniques and demonstrates deep comprehension of instances and articles.
    Object
    Latent Semantic Indexing

Authors

Years

Languages

  • e 70
  • d 8

Types