Search (14 results, page 1 of 1)

  • × type_ss:"p"
  1. Tramullas, J.; Garrido-Picazo, P.; Sánchez-Casabón, A.I.: Use of Wikipedia categories on information retrieval research : a brief review (2020) 0.03
    0.026916523 = product of:
      0.053833045 = sum of:
        0.04717497 = weight(_text_:processing in 5365) [ClassicSimilarity], result of:
          0.04717497 = score(doc=5365,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.26835677 = fieldWeight in 5365, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.046875 = fieldNorm(doc=5365)
        0.006658075 = product of:
          0.019974224 = sum of:
            0.019974224 = weight(_text_:science in 5365) [ClassicSimilarity], result of:
              0.019974224 = score(doc=5365,freq=2.0), product of:
                0.11438741 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.043425296 = queryNorm
                0.17461908 = fieldWeight in 5365, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5365)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    Wikipedia categories, a classification scheme built for organizing and describing Wikpedia articles, are being applied in computer science research. This paper adopts a systematic literature review approach, in order to identify different approaches and uses of Wikipedia categories in information retrieval research. Several types of work are identified, depending on the intrinsic study of the categories structure, or its use as a tool for the processing and analysis of other documentary corpus different to Wikipedia. Information retrieval is identified as one of the major areas of use, in particular its application in the refinement and improvement of search expressions, and the construction of textual corpus. However, the set of available works shows that in many cases research approaches applied and results obtained can be integrated into a comprehensive and inclusive concept of information retrieval.
  2. Luo, L.; Ju, J.; Li, Y.-F.; Haffari, G.; Xiong, B.; Pan, S.: ChatRule: mining logical rules with large language models for knowledge graph reasoning (2023) 0.02
    0.02455918 = product of:
      0.04911836 = sum of:
        0.03931248 = weight(_text_:processing in 1171) [ClassicSimilarity], result of:
          0.03931248 = score(doc=1171,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.22363065 = fieldWeight in 1171, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1171)
        0.0098058805 = product of:
          0.029417641 = sum of:
            0.029417641 = weight(_text_:22 in 1171) [ClassicSimilarity], result of:
              0.029417641 = score(doc=1171,freq=2.0), product of:
                0.15206799 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043425296 = queryNorm
                0.19345059 = fieldWeight in 1171, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1171)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    Logical rules are essential for uncovering the logical connections between relations, which could improve the reasoning performance and provide interpretable results on knowledge graphs (KGs). Although there have been many efforts to mine meaningful logical rules over KGs, existing methods suffer from the computationally intensive searches over the rule space and a lack of scalability for large-scale KGs. Besides, they often ignore the semantics of relations which is crucial for uncovering logical connections. Recently, large language models (LLMs) have shown impressive performance in the field of natural language processing and various applications, owing to their emergent ability and generalizability. In this paper, we propose a novel framework, ChatRule, unleashing the power of large language models for mining logical rules over knowledge graphs. Specifically, the framework is initiated with an LLM-based rule generator, leveraging both the semantic and structural information of KGs to prompt LLMs to generate logical rules. To refine the generated rules, a rule ranking module estimates the rule quality by incorporating facts from existing KGs. Last, a rule validator harnesses the reasoning ability of LLMs to validate the logical correctness of ranked rules through chain-of-thought reasoning. ChatRule is evaluated on four large-scale KGs, w.r.t. different rule quality metrics and downstream tasks, showing the effectiveness and scalability of our method.
    Date
    23.11.2023 19:07:22
  3. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.02
    0.01724272 = product of:
      0.06897088 = sum of:
        0.06897088 = product of:
          0.20691264 = sum of:
            0.20691264 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.20691264 = score(doc=862,freq=2.0), product of:
                0.36816013 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.043425296 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  4. Peponakis, M.; Mastora, A.; Kapidakis, S.; Doerr, M.: Expressiveness and machine processability of Knowledge Organization Systems (KOS) : an analysis of concepts and relations (2020) 0.01
    0.00982812 = product of:
      0.03931248 = sum of:
        0.03931248 = weight(_text_:processing in 5787) [ClassicSimilarity], result of:
          0.03931248 = score(doc=5787,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.22363065 = fieldWeight in 5787, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5787)
      0.25 = coord(1/4)
    
    Abstract
    This study considers the expressiveness (that is the expressive power or expressivity) of different types of Knowledge Organization Systems (KOS) and discusses its potential to be machine-processable in the context of the Semantic Web. For this purpose, the theoretical foundations of KOS are reviewed based on conceptualizations introduced by the Functional Requirements for Subject Authority Data (FRSAD) and the Simple Knowledge Organization System (SKOS); natural language processing techniques are also implemented. Applying a comparative analysis, the dataset comprises a thesaurus (Eurovoc), a subject headings system (LCSH) and a classification scheme (DDC). These are compared with an ontology (CIDOC-CRM) by focusing on how they define and handle concepts and relations. It was observed that LCSH and DDC focus on the formalism of character strings (nomens) rather than on the modelling of semantics; their definition of what constitutes a concept is quite fuzzy, and they comprise a large number of complex concepts. By contrast, thesauri have a coherent definition of what constitutes a concept, and apply a systematic approach to the modelling of relations. Ontologies explicitly define diverse types of relations, and are by their nature machine-processable. The paper concludes that the potential of both the expressiveness and machine processability of each KOS is extensively regulated by its structural rules. It is harder to represent subject headings and classification schemes as semantic networks with nodes and arcs, while thesauri are more suitable for such a representation. In addition, a paradigm shift is revealed which focuses on the modelling of relations between concepts, rather than the concepts themselves.
  5. Großjohann, K.: Gathering-, Harvesting-, Suchmaschinen (1996) 0.01
    0.008320566 = product of:
      0.033282265 = sum of:
        0.033282265 = product of:
          0.09984679 = sum of:
            0.09984679 = weight(_text_:22 in 3227) [ClassicSimilarity], result of:
              0.09984679 = score(doc=3227,freq=4.0), product of:
                0.15206799 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043425296 = queryNorm
                0.6565931 = fieldWeight in 3227, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3227)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    7. 2.1996 22:38:41
    Pages
    22 S
  6. Elazar, D.H.: ¬The making of a classification scheme for libraries of Judaica (2000) 0.01
    0.005936935 = product of:
      0.02374774 = sum of:
        0.02374774 = product of:
          0.07124322 = sum of:
            0.07124322 = weight(_text_:29 in 5400) [ClassicSimilarity], result of:
              0.07124322 = score(doc=5400,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.46638384 = fieldWeight in 5400, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=5400)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    1.11.2000 14:55:29
  7. Hobohm, H.-C.: Zensur in der Digitalität - eine Überwindung der Moderne? : Die Rolle der Bibliotheken (2020) 0.01
    0.005936935 = product of:
      0.02374774 = sum of:
        0.02374774 = product of:
          0.07124322 = sum of:
            0.07124322 = weight(_text_:29 in 5371) [ClassicSimilarity], result of:
              0.07124322 = score(doc=5371,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.46638384 = fieldWeight in 5371, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=5371)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Content
    Beitrag zur Tagung: "Nationalsozialismus Digital. Die Verantwortung von Bibliotheken, Archiven und Museen sowie Forschungseinrichtungen und Medien im Umgang mit der NSZeit im Netz." Österreichische Nationalbibliothek, Universität Wien, 27. - 29. November 2019
  8. Wätjen, H.-J.: Mensch oder Maschine? : Auswahl und Erschließung vonm Informationsressourcen im Internet (1996) 0.00
    0.0049029402 = product of:
      0.019611761 = sum of:
        0.019611761 = product of:
          0.058835283 = sum of:
            0.058835283 = weight(_text_:22 in 3161) [ClassicSimilarity], result of:
              0.058835283 = score(doc=3161,freq=2.0), product of:
                0.15206799 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043425296 = queryNorm
                0.38690117 = fieldWeight in 3161, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3161)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    2. 2.1996 15:40:22
  9. Jaenecke, P.: Knowledge organization due to theory formation (1995) 0.00
    0.0039579566 = product of:
      0.015831826 = sum of:
        0.015831826 = product of:
          0.047495477 = sum of:
            0.047495477 = weight(_text_:29 in 3751) [ClassicSimilarity], result of:
              0.047495477 = score(doc=3751,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.31092256 = fieldWeight in 3751, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3751)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    29. 3.1996 17:26:47
  10. Schöneberg, U.; Gödert, W.: Erschließung mathematischer Publikationen mittels linguistischer Verfahren (2012) 0.00
    0.0029684675 = product of:
      0.01187387 = sum of:
        0.01187387 = product of:
          0.03562161 = sum of:
            0.03562161 = weight(_text_:29 in 1055) [ClassicSimilarity], result of:
              0.03562161 = score(doc=1055,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.23319192 = fieldWeight in 1055, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1055)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    12. 9.2013 12:29:05
  11. Isaac, A.; Raemy, J.A.; Meijers, E.; Valk, S. De; Freire, N.: Metadata aggregation via linked data : results of the Europeana Common Culture project (2020) 0.00
    0.0029684675 = product of:
      0.01187387 = sum of:
        0.01187387 = product of:
          0.03562161 = sum of:
            0.03562161 = weight(_text_:29 in 39) [ClassicSimilarity], result of:
              0.03562161 = score(doc=39,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.23319192 = fieldWeight in 39, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=39)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    17.11.2020 11:29:00
  12. Lange, C.; Ion, P.; Dimou, A.; Bratsas, C.; Sperber, W.; Kohlhasel, M.; Antoniou, I.: Getting mathematics towards the Web of Data : the case of the Mathematics Subject Classification (2012) 0.00
    0.0024025259 = product of:
      0.009610103 = sum of:
        0.009610103 = product of:
          0.02883031 = sum of:
            0.02883031 = weight(_text_:science in 111) [ClassicSimilarity], result of:
              0.02883031 = score(doc=111,freq=6.0), product of:
                0.11438741 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.043425296 = queryNorm
                0.25204095 = fieldWeight in 111, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=111)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    The Mathematics Subject Classification (MSC), maintained by the American Mathematical Society's Mathematical Reviews (MR) and FIZ Karlsruhe's Zentralblatt für Mathematik (Zbl), is a scheme for classifying publications in mathematics according to their subjects. While it is widely used, its traditional, idiosyncratic conceptualization and representation requires custom implementations of search, query and annotation support. This did not encourage people to create and explore connections of mathematics to subjects of related domains (e.g. science), and it made the scheme hard to maintain. We have reimplemented the current version of MSC2010 as a Linked Open Dataset using SKOS and our focus is concentrated on turning it into the new MSC authority. This paper explains the motivation, and details of our design considerations and how we realized them in the implementation. We present in-the-field use cases and point out how e-science applications can take advantage of the MSC LOD set. We conclude with a roadmap for bootstrapping the presence of mathematical and mathematics-based science, technology, and engineering knowledge on the Web of Data, where it has been noticeably underrepresented so far, starting from MSC/SKOS as a seed.
  13. Wormell, I.: Multifunctional information work : new demands for training? (1995) 0.00
    0.0022193585 = product of:
      0.008877434 = sum of:
        0.008877434 = product of:
          0.0266323 = sum of:
            0.0266323 = weight(_text_:science in 3371) [ClassicSimilarity], result of:
              0.0266323 = score(doc=3371,freq=2.0), product of:
                0.11438741 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.043425296 = queryNorm
                0.23282544 = fieldWeight in 3371, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3371)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    The paper calls for an integrated approach to information science education where disciplinary interaction is predicated on the forgoing of formal, informal and sustainable links with researchers and pracitioners in other fields. The modern information profession, in order to promote its creativity and to strengthen its development, has to go beyond the traditional roles and functions and should extend the professions' horizons. Thus the LIS education and training programmes must aim to foster professionals who, one day, will create new jobs and not just fill the old ones
  14. Guizzardi, G.; Guarino, N.: Semantics, ontology and explanation (2023) 0.00
    0.0016645187 = product of:
      0.006658075 = sum of:
        0.006658075 = product of:
          0.019974224 = sum of:
            0.019974224 = weight(_text_:science in 976) [ClassicSimilarity], result of:
              0.019974224 = score(doc=976,freq=2.0), product of:
                0.11438741 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.043425296 = queryNorm
                0.17461908 = fieldWeight in 976, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=976)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    The terms 'semantics' and 'ontology' are increasingly appearing together with 'explanation', not only in the scientific literature, but also in organizational communication. However, all of these terms are also being significantly overloaded. In this paper, we discuss their strong relation under particular interpretations. Specifically, we discuss a notion of explanation termed ontological unpacking, which aims at explaining symbolic domain descriptions (conceptual models, knowledge graphs, logical specifications) by revealing their ontological commitment in terms of their assumed truthmakers, i.e., the entities in one's ontology that make the propositions in those descriptions true. To illustrate this idea, we employ an ontological theory of relations to explain (by revealing the hidden semantics of) a very simple symbolic model encoded in the standard modeling language UML. We also discuss the essential role played by ontology-driven conceptual models (resulting from this form of explanation processes) in properly supporting semantic interoperability tasks. Finally, we discuss the relation between ontological unpacking and other forms of explanation in philosophy and science, as well as in the area of Artificial Intelligence.