Search (720 results, page 1 of 36)

  • × theme_ss:"Computerlinguistik"
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.13
    0.13493797 = product of:
      0.20240696 = sum of:
        0.047273763 = product of:
          0.14182128 = sum of:
            0.14182128 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.14182128 = score(doc=562,freq=2.0), product of:
                0.25234294 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.029764405 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.0052465936 = weight(_text_:a in 562) [ClassicSimilarity], result of:
          0.0052465936 = score(doc=562,freq=8.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.15287387 = fieldWeight in 562, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.14182128 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.14182128 = score(doc=562,freq=2.0), product of:
            0.25234294 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.029764405 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.008065332 = product of:
          0.024195995 = sum of:
            0.024195995 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.024195995 = score(doc=562,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
      0.6666667 = coord(4/6)
    
    Abstract
    Document representations for text classification are typically based on the classical Bag-Of-Words paradigm. This approach comes with deficiencies that motivate the integration of features on a higher semantic level than single words. In this paper we propose an enhancement of the classical document representation through concepts extracted from background knowledge. Boosting is used for actual classification. Experimental evaluations on two well known text corpora support our approach through consistent improvement of the results.
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
    Type
    a
  2. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.10
    0.09748046 = product of:
      0.19496092 = sum of:
        0.047273763 = product of:
          0.14182128 = sum of:
            0.14182128 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.14182128 = score(doc=862,freq=2.0), product of:
                0.25234294 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.029764405 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
        0.0058658705 = weight(_text_:a in 862) [ClassicSimilarity], result of:
          0.0058658705 = score(doc=862,freq=10.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.1709182 = fieldWeight in 862, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.14182128 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.14182128 = score(doc=862,freq=2.0), product of:
            0.25234294 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.029764405 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
      0.5 = coord(3/6)
    
    Abstract
    This research revisits the classic Turing test and compares recent large language models such as ChatGPT for their abilities to reproduce human-level comprehension and compelling text generation. Two task challenges- summary and question answering- prompt ChatGPT to produce original content (98-99%) from a single text entry and sequential questions initially posed by Turing in 1950. We score the original and generated content against the OpenAI GPT-2 Output Detector from 2019, and establish multiple cases where the generated content proves original and undetectable (98%). The question of a machine fooling a human judge recedes in this work relative to the question of "how would one prove it?" The original contribution of the work presents a metric and simple grammatical set for understanding the writing mechanics of chatbots in evaluating their readability and statistical clarity, engagement, delivery, overall quality, and plagiarism risks. While Turing's original prose scores at least 14% below the machine-generated output, whether an algorithm displays hints of Turing's true initial thoughts (the "Lovelace 2.0" test) remains unanswerable.
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
    Type
    a
  3. Huo, W.: Automatic multi-word term extraction and its application to Web-page summarization (2012) 0.08
    0.07721514 = product of:
      0.15443029 = sum of:
        0.0045436835 = weight(_text_:a in 563) [ClassicSimilarity], result of:
          0.0045436835 = score(doc=563,freq=6.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.13239266 = fieldWeight in 563, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.14182128 = weight(_text_:2f in 563) [ClassicSimilarity], result of:
          0.14182128 = score(doc=563,freq=2.0), product of:
            0.25234294 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.029764405 = queryNorm
            0.56201804 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.008065332 = product of:
          0.024195995 = sum of:
            0.024195995 = weight(_text_:22 in 563) [ClassicSimilarity], result of:
              0.024195995 = score(doc=563,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.23214069 = fieldWeight in 563, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=563)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    In this thesis we propose three new word association measures for multi-word term extraction. We combine these association measures with LocalMaxs algorithm in our extraction model and compare the results of different multi-word term extraction methods. Our approach is language and domain independent and requires no training data. It can be applied to such tasks as text summarization, information retrieval, and document classification. We further explore the potential of using multi-word terms as an effective representation for general web-page summarization. We extract multi-word terms from human written summaries in a large collection of web-pages, and generate the summaries by aligning document words with these multi-word terms. Our system applies machine translation technology to learn the aligning process from a training set and focuses on selecting high quality multi-word terms from human written summaries to generate suitable results for web-page summarization.
    Content
    A Thesis presented to The University of Guelph In partial fulfilment of requirements for the degree of Master of Science in Computer Science. Vgl. Unter: http://www.inf.ufrgs.br%2F~ceramisch%2Fdownload_files%2Fpublications%2F2009%2Fp01.pdf.
    Date
    10. 1.2013 19:22:47
  4. Ruge, G.: ¬A spreading activation network for automatic generation of thesaurus relationships (1991) 0.02
    0.022829084 = product of:
      0.045658167 = sum of:
        0.010601928 = weight(_text_:a in 4506) [ClassicSimilarity], result of:
          0.010601928 = score(doc=4506,freq=6.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.3089162 = fieldWeight in 4506, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=4506)
        0.01623713 = product of:
          0.06494852 = sum of:
            0.06494852 = weight(_text_:g in 4506) [ClassicSimilarity], result of:
              0.06494852 = score(doc=4506,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.5809685 = fieldWeight in 4506, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4506)
          0.25 = coord(1/4)
        0.018819109 = product of:
          0.056457322 = sum of:
            0.056457322 = weight(_text_:22 in 4506) [ClassicSimilarity], result of:
              0.056457322 = score(doc=4506,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.5416616 = fieldWeight in 4506, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4506)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Date
    8.10.2000 11:52:22
    Source
    Library science with a slant to documentation. 28(1991) no.4, S.125-130
    Type
    a
  5. Boleda, G.; Evert, S.: Multiword expressions : a pain in the neck of lexical semantics (2009) 0.02
    0.017647399 = product of:
      0.035294797 = sum of:
        0.0052465936 = weight(_text_:a in 4888) [ClassicSimilarity], result of:
          0.0052465936 = score(doc=4888,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.15287387 = fieldWeight in 4888, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=4888)
        0.013917539 = product of:
          0.055670157 = sum of:
            0.055670157 = weight(_text_:g in 4888) [ClassicSimilarity], result of:
              0.055670157 = score(doc=4888,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.49797297 = fieldWeight in 4888, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4888)
          0.25 = coord(1/4)
        0.016130663 = product of:
          0.04839199 = sum of:
            0.04839199 = weight(_text_:22 in 4888) [ClassicSimilarity], result of:
              0.04839199 = score(doc=4888,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.46428138 = fieldWeight in 4888, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4888)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Date
    1. 3.2013 14:56:22
  6. Hammwöhner, R.: TransRouter revisited : Decision support in the routing of translation projects (2000) 0.01
    0.011414542 = product of:
      0.022829084 = sum of:
        0.005300964 = weight(_text_:a in 5483) [ClassicSimilarity], result of:
          0.005300964 = score(doc=5483,freq=6.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.1544581 = fieldWeight in 5483, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5483)
        0.008118565 = product of:
          0.03247426 = sum of:
            0.03247426 = weight(_text_:g in 5483) [ClassicSimilarity], result of:
              0.03247426 = score(doc=5483,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.29048425 = fieldWeight in 5483, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5483)
          0.25 = coord(1/4)
        0.009409554 = product of:
          0.028228661 = sum of:
            0.028228661 = weight(_text_:22 in 5483) [ClassicSimilarity], result of:
              0.028228661 = score(doc=5483,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.2708308 = fieldWeight in 5483, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5483)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    This paper gives an outline of the final results of the TransRouter project. In the scope of this project a decision support system for translation managers has been developed, which will support the selection of appropriate routes for translation projects. In this paper emphasis is put on the decision model, which is based on a stepwise refined assessment of translation routes. The workflow of using this system is considered as well
    Date
    10.12.2000 18:22:35
    Source
    Informationskompetenz - Basiskompetenz in der Informationsgesellschaft: Proceedings des 7. Internationalen Symposiums für Informationswissenschaft (ISI 2000), Hrsg.: G. Knorz u. R. Kuhlen
    Type
    a
  7. Rahmstorf, G.: Concept structures for large vocabularies (1998) 0.01
    0.010135347 = product of:
      0.020270694 = sum of:
        0.0052465936 = weight(_text_:a in 75) [ClassicSimilarity], result of:
          0.0052465936 = score(doc=75,freq=8.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.15287387 = fieldWeight in 75, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=75)
        0.0069587696 = product of:
          0.027835079 = sum of:
            0.027835079 = weight(_text_:g in 75) [ClassicSimilarity], result of:
              0.027835079 = score(doc=75,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.24898648 = fieldWeight in 75, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.046875 = fieldNorm(doc=75)
          0.25 = coord(1/4)
        0.008065332 = product of:
          0.024195995 = sum of:
            0.024195995 = weight(_text_:22 in 75) [ClassicSimilarity], result of:
              0.024195995 = score(doc=75,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.23214069 = fieldWeight in 75, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=75)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    A technology is described which supports the acquisition, visualisation and manipulation of large vocabularies with associated structures. It is used for dictionary production, terminology data bases, thesauri, library classification systems etc. Essential features of the technology are a lexicographic user interface, variable word description, unlimited list of word readings, a concept language, automatic transformations of formulas into graphic structures, structure manipulation operations and retransformation into formulas. The concept language includes notations for undefined concepts. The structure of defined concepts can be constructed interactively. The technology supports the generation of large vocabularies with structures representing word senses. Concept structures and ordering systems for indexing and retrieval can be constructed separately and connected by associating relations.
    Date
    30.12.2001 19:01:22
    Type
    a
  8. Bian, G.-W.; Chen, H.-H.: Cross-language information access to multilingual collections on the Internet (2000) 0.01
    0.009783892 = product of:
      0.019567784 = sum of:
        0.0045436835 = weight(_text_:a in 4436) [ClassicSimilarity], result of:
          0.0045436835 = score(doc=4436,freq=6.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.13239266 = fieldWeight in 4436, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=4436)
        0.0069587696 = product of:
          0.027835079 = sum of:
            0.027835079 = weight(_text_:g in 4436) [ClassicSimilarity], result of:
              0.027835079 = score(doc=4436,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.24898648 = fieldWeight in 4436, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4436)
          0.25 = coord(1/4)
        0.008065332 = product of:
          0.024195995 = sum of:
            0.024195995 = weight(_text_:22 in 4436) [ClassicSimilarity], result of:
              0.024195995 = score(doc=4436,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.23214069 = fieldWeight in 4436, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4436)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    Language barrier is the major problem that people face in searching for, retrieving, and understanding multilingual collections on the Internet. This paper deals with query translation and document translation in a Chinese-English information retrieval system called MTIR. Bilingual dictionary and monolingual corpus-based approaches are adopted to select suitable tranlated query terms. A machine transliteration algorithm is introduced to resolve proper name searching. We consider several design issues for document translation, including which material is translated, what roles the HTML tags play in translation, what the tradeoff is between the speed performance and the translation performance, and what from the translated result is presented in. About 100.000 Web pages translated in the last 4 months of 1997 are used for quantitative study of online and real-time Web page translation
    Date
    16. 2.2000 14:22:39
    Type
    a
  9. Warner, A.J.: Natural language processing (1987) 0.01
    0.009501004 = product of:
      0.02850301 = sum of:
        0.006995458 = weight(_text_:a in 337) [ClassicSimilarity], result of:
          0.006995458 = score(doc=337,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.20383182 = fieldWeight in 337, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.125 = fieldNorm(doc=337)
        0.021507552 = product of:
          0.064522654 = sum of:
            0.064522654 = weight(_text_:22 in 337) [ClassicSimilarity], result of:
              0.064522654 = score(doc=337,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.61904186 = fieldWeight in 337, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=337)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Source
    Annual review of information science and technology. 22(1987), S.79-108
    Type
    a
  10. Schwarz, C.: Probleme der syntaktischen Indexierung (1986) 0.01
    0.008517393 = product of:
      0.025552178 = sum of:
        0.006995458 = weight(_text_:a in 8180) [ClassicSimilarity], result of:
          0.006995458 = score(doc=8180,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.20383182 = fieldWeight in 8180, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.125 = fieldNorm(doc=8180)
        0.01855672 = product of:
          0.07422688 = sum of:
            0.07422688 = weight(_text_:g in 8180) [ClassicSimilarity], result of:
              0.07422688 = score(doc=8180,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.663964 = fieldWeight in 8180, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.125 = fieldNorm(doc=8180)
          0.25 = coord(1/4)
      0.33333334 = coord(2/6)
    
    Source
    Informationslinguistische Texterschließung. Hrsg.: C. Schwarz u. G. Thurmair
    Type
    a
  11. Salton, G.: Automated language processing (1968) 0.01
    0.008517393 = product of:
      0.025552178 = sum of:
        0.006995458 = weight(_text_:a in 233) [ClassicSimilarity], result of:
          0.006995458 = score(doc=233,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.20383182 = fieldWeight in 233, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.125 = fieldNorm(doc=233)
        0.01855672 = product of:
          0.07422688 = sum of:
            0.07422688 = weight(_text_:g in 233) [ClassicSimilarity], result of:
              0.07422688 = score(doc=233,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.663964 = fieldWeight in 233, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.125 = fieldNorm(doc=233)
          0.25 = coord(1/4)
      0.33333334 = coord(2/6)
    
    Type
    a
  12. Luo, L.; Ju, J.; Li, Y.-F.; Haffari, G.; Xiong, B.; Pan, S.: ChatRule: mining logical rules with large language models for knowledge graph reasoning (2023) 0.01
    0.008446123 = product of:
      0.016892247 = sum of:
        0.0043721613 = weight(_text_:a in 1171) [ClassicSimilarity], result of:
          0.0043721613 = score(doc=1171,freq=8.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.12739488 = fieldWeight in 1171, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1171)
        0.005798975 = product of:
          0.0231959 = sum of:
            0.0231959 = weight(_text_:g in 1171) [ClassicSimilarity], result of:
              0.0231959 = score(doc=1171,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.20748875 = fieldWeight in 1171, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1171)
          0.25 = coord(1/4)
        0.0067211105 = product of:
          0.020163331 = sum of:
            0.020163331 = weight(_text_:22 in 1171) [ClassicSimilarity], result of:
              0.020163331 = score(doc=1171,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.19345059 = fieldWeight in 1171, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1171)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    Logical rules are essential for uncovering the logical connections between relations, which could improve the reasoning performance and provide interpretable results on knowledge graphs (KGs). Although there have been many efforts to mine meaningful logical rules over KGs, existing methods suffer from the computationally intensive searches over the rule space and a lack of scalability for large-scale KGs. Besides, they often ignore the semantics of relations which is crucial for uncovering logical connections. Recently, large language models (LLMs) have shown impressive performance in the field of natural language processing and various applications, owing to their emergent ability and generalizability. In this paper, we propose a novel framework, ChatRule, unleashing the power of large language models for mining logical rules over knowledge graphs. Specifically, the framework is initiated with an LLM-based rule generator, leveraging both the semantic and structural information of KGs to prompt LLMs to generate logical rules. To refine the generated rules, a rule ranking module estimates the rule quality by incorporating facts from existing KGs. Last, a rule validator harnesses the reasoning ability of LLMs to validate the logical correctness of ranked rules through chain-of-thought reasoning. ChatRule is evaluated on four large-scale KGs, w.r.t. different rule quality metrics and downstream tasks, showing the effectiveness and scalability of our method.
    Date
    23.11.2023 19:07:22
  13. McMahon, J.G.; Smith, F.J.: Improved statistical language model performance with automatic generated word hierarchies (1996) 0.01
    0.008313378 = product of:
      0.024940135 = sum of:
        0.006121026 = weight(_text_:a in 3164) [ClassicSimilarity], result of:
          0.006121026 = score(doc=3164,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.17835285 = fieldWeight in 3164, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=3164)
        0.018819109 = product of:
          0.056457322 = sum of:
            0.056457322 = weight(_text_:22 in 3164) [ClassicSimilarity], result of:
              0.056457322 = score(doc=3164,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.5416616 = fieldWeight in 3164, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3164)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Source
    Computational linguistics. 22(1996) no.2, S.217-248
    Type
    a
  14. Somers, H.: Example-based machine translation : Review article (1999) 0.01
    0.008313378 = product of:
      0.024940135 = sum of:
        0.006121026 = weight(_text_:a in 6672) [ClassicSimilarity], result of:
          0.006121026 = score(doc=6672,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.17835285 = fieldWeight in 6672, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=6672)
        0.018819109 = product of:
          0.056457322 = sum of:
            0.056457322 = weight(_text_:22 in 6672) [ClassicSimilarity], result of:
              0.056457322 = score(doc=6672,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.5416616 = fieldWeight in 6672, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6672)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Date
    31. 7.1996 9:22:19
    Type
    a
  15. New tools for human translators (1997) 0.01
    0.008313378 = product of:
      0.024940135 = sum of:
        0.006121026 = weight(_text_:a in 1179) [ClassicSimilarity], result of:
          0.006121026 = score(doc=1179,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.17835285 = fieldWeight in 1179, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=1179)
        0.018819109 = product of:
          0.056457322 = sum of:
            0.056457322 = weight(_text_:22 in 1179) [ClassicSimilarity], result of:
              0.056457322 = score(doc=1179,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.5416616 = fieldWeight in 1179, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1179)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    A special issue devoted to the theme of new tools for human tranlators
    Date
    31. 7.1996 9:22:19
  16. Baayen, R.H.; Lieber, H.: Word frequency distributions and lexical semantics (1997) 0.01
    0.008313378 = product of:
      0.024940135 = sum of:
        0.006121026 = weight(_text_:a in 3117) [ClassicSimilarity], result of:
          0.006121026 = score(doc=3117,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.17835285 = fieldWeight in 3117, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=3117)
        0.018819109 = product of:
          0.056457322 = sum of:
            0.056457322 = weight(_text_:22 in 3117) [ClassicSimilarity], result of:
              0.056457322 = score(doc=3117,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.5416616 = fieldWeight in 3117, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3117)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Date
    28. 2.1999 10:48:22
    Type
    a
  17. ¬Der Student aus dem Computer (2023) 0.01
    0.008313378 = product of:
      0.024940135 = sum of:
        0.006121026 = weight(_text_:a in 1079) [ClassicSimilarity], result of:
          0.006121026 = score(doc=1079,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.17835285 = fieldWeight in 1079, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=1079)
        0.018819109 = product of:
          0.056457322 = sum of:
            0.056457322 = weight(_text_:22 in 1079) [ClassicSimilarity], result of:
              0.056457322 = score(doc=1079,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.5416616 = fieldWeight in 1079, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1079)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Date
    27. 1.2023 16:22:55
    Type
    a
  18. Schwarz, C.: Natural language and information retrieval : Kommentierte Literaturliste zu Systemen, Verfahren und Tools (1986) 0.01
    0.007452719 = product of:
      0.022358157 = sum of:
        0.006121026 = weight(_text_:a in 408) [ClassicSimilarity], result of:
          0.006121026 = score(doc=408,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.17835285 = fieldWeight in 408, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=408)
        0.01623713 = product of:
          0.06494852 = sum of:
            0.06494852 = weight(_text_:g in 408) [ClassicSimilarity], result of:
              0.06494852 = score(doc=408,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.5809685 = fieldWeight in 408, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.109375 = fieldNorm(doc=408)
          0.25 = coord(1/4)
      0.33333334 = coord(2/6)
    
    Source
    Informationsliguistische Texterschließung. Hrsg.: C. Schwarz u. G. Thurmair
    Type
    a
  19. Salton, G.; Buckley, C.; Smith, M.: On the application of syntactic methodologies in automatic text analysis (1990) 0.01
    0.007452719 = product of:
      0.022358157 = sum of:
        0.006121026 = weight(_text_:a in 7864) [ClassicSimilarity], result of:
          0.006121026 = score(doc=7864,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.17835285 = fieldWeight in 7864, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=7864)
        0.01623713 = product of:
          0.06494852 = sum of:
            0.06494852 = weight(_text_:g in 7864) [ClassicSimilarity], result of:
              0.06494852 = score(doc=7864,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.5809685 = fieldWeight in 7864, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.109375 = fieldNorm(doc=7864)
          0.25 = coord(1/4)
      0.33333334 = coord(2/6)
    
    Type
    a
  20. Rahmstorf, G.: Semantisches Information Retrieval (1994) 0.01
    0.007452719 = product of:
      0.022358157 = sum of:
        0.006121026 = weight(_text_:a in 8879) [ClassicSimilarity], result of:
          0.006121026 = score(doc=8879,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.17835285 = fieldWeight in 8879, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=8879)
        0.01623713 = product of:
          0.06494852 = sum of:
            0.06494852 = weight(_text_:g in 8879) [ClassicSimilarity], result of:
              0.06494852 = score(doc=8879,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.5809685 = fieldWeight in 8879, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.109375 = fieldNorm(doc=8879)
          0.25 = coord(1/4)
      0.33333334 = coord(2/6)
    
    Type
    a

Languages

Types

  • a 629
  • el 77
  • m 47
  • s 26
  • x 9
  • p 7
  • b 1
  • d 1
  • pat 1
  • r 1
  • More… Less…

Subjects

Classifications