Search (125 results, page 1 of 7)

  • × theme_ss:"Computerlinguistik"
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.16
    0.15597759 = product of:
      0.25996265 = sum of:
        0.062332056 = product of:
          0.18699616 = sum of:
            0.18699616 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.18699616 = score(doc=562,freq=2.0), product of:
                0.3327227 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03924537 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.18699616 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.18699616 = score(doc=562,freq=2.0), product of:
            0.3327227 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03924537 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.010634413 = product of:
          0.031903237 = sum of:
            0.031903237 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.031903237 = score(doc=562,freq=2.0), product of:
                0.13743061 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03924537 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
      0.6 = coord(3/5)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  2. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.10
    0.09973129 = product of:
      0.24932823 = sum of:
        0.062332056 = product of:
          0.18699616 = sum of:
            0.18699616 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.18699616 = score(doc=862,freq=2.0), product of:
                0.3327227 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03924537 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
        0.18699616 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.18699616 = score(doc=862,freq=2.0), product of:
            0.3327227 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03924537 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
      0.4 = coord(2/5)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  3. Huo, W.: Automatic multi-word term extraction and its application to Web-page summarization (2012) 0.08
    0.07905223 = product of:
      0.19763057 = sum of:
        0.18699616 = weight(_text_:2f in 563) [ClassicSimilarity], result of:
          0.18699616 = score(doc=563,freq=2.0), product of:
            0.3327227 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03924537 = queryNorm
            0.56201804 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.010634413 = product of:
          0.031903237 = sum of:
            0.031903237 = weight(_text_:22 in 563) [ClassicSimilarity], result of:
              0.031903237 = score(doc=563,freq=2.0), product of:
                0.13743061 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03924537 = queryNorm
                0.23214069 = fieldWeight in 563, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=563)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Content
    A Thesis presented to The University of Guelph In partial fulfilment of requirements for the degree of Master of Science in Computer Science. Vgl. Unter: http://www.inf.ufrgs.br%2F~ceramisch%2Fdownload_files%2Fpublications%2F2009%2Fp01.pdf.
    Date
    10. 1.2013 19:22:47
  4. Baayen, R.H.; Lieber, H.: Word frequency distributions and lexical semantics (1997) 0.08
    0.07640171 = product of:
      0.19100428 = sum of:
        0.16619065 = weight(_text_:relation in 3117) [ClassicSimilarity], result of:
          0.16619065 = score(doc=3117,freq=2.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.8093299 = fieldWeight in 3117, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.109375 = fieldNorm(doc=3117)
        0.02481363 = product of:
          0.07444089 = sum of:
            0.07444089 = weight(_text_:22 in 3117) [ClassicSimilarity], result of:
              0.07444089 = score(doc=3117,freq=2.0), product of:
                0.13743061 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03924537 = queryNorm
                0.5416616 = fieldWeight in 3117, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3117)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Abstract
    Relation between meaning, lexical productivity and frequency of use
    Date
    28. 2.1999 10:48:22
  5. Semantic role universals and argument linking : theoretical, typological, and psycholinguistic perspectives (2006) 0.07
    0.06909622 = product of:
      0.17274055 = sum of:
        0.06715117 = weight(_text_:relation in 3670) [ClassicSimilarity], result of:
          0.06715117 = score(doc=3670,freq=4.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.32701868 = fieldWeight in 3670, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03125 = fieldNorm(doc=3670)
        0.10558939 = weight(_text_:aufsatzsammlung in 3670) [ClassicSimilarity], result of:
          0.10558939 = score(doc=3670,freq=4.0), product of:
            0.25749236 = queryWeight, product of:
              6.5610886 = idf(docFreq=169, maxDocs=44218)
              0.03924537 = queryNorm
            0.41006804 = fieldWeight in 3670, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              6.5610886 = idf(docFreq=169, maxDocs=44218)
              0.03125 = fieldNorm(doc=3670)
      0.4 = coord(2/5)
    
    RSWK
    Thematische Relation / Aufsatzsammlung (BVB)
    Subject
    Thematische Relation / Aufsatzsammlung (BVB)
  6. Informationslinguistische Texterschließung (1986) 0.06
    0.064660035 = product of:
      0.32330015 = sum of:
        0.32330015 = weight(_text_:aufsatzsammlung in 186) [ClassicSimilarity], result of:
          0.32330015 = score(doc=186,freq=24.0), product of:
            0.25749236 = queryWeight, product of:
              6.5610886 = idf(docFreq=169, maxDocs=44218)
              0.03924537 = queryNorm
            1.2555718 = fieldWeight in 186, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              6.5610886 = idf(docFreq=169, maxDocs=44218)
              0.0390625 = fieldNorm(doc=186)
      0.2 = coord(1/5)
    
    RSWK
    Information Retrieval / Aufsatzsammlung (DNB)
    Automatische Sprachanalyse / Morphologie / Aufsatzsammlung (SBB / GBV)
    Automatische Sprachanalyse / Morphologie <Linguistik> / Aufsatzsammlung (DNB)
    Linguistische Datenverarbeitung / Linguistik / Aufsatzsammlung (SWB)
    Linguistik / Information Retrieval / Aufsatzsammlung (SWB / BVB)
    Linguistische Datenverarbeitung / Textanalyse / Aufsatzsammlung (BVB)
    Subject
    Information Retrieval / Aufsatzsammlung (DNB)
    Automatische Sprachanalyse / Morphologie / Aufsatzsammlung (SBB / GBV)
    Automatische Sprachanalyse / Morphologie <Linguistik> / Aufsatzsammlung (DNB)
    Linguistische Datenverarbeitung / Linguistik / Aufsatzsammlung (SWB)
    Linguistik / Information Retrieval / Aufsatzsammlung (SWB / BVB)
    Linguistische Datenverarbeitung / Textanalyse / Aufsatzsammlung (BVB)
  7. Semantik, Lexikographie und Computeranwendungen : Workshop ... (Bonn) : 1995.01.27-28 (1996) 0.06
    0.056339502 = product of:
      0.14084876 = sum of:
        0.13198674 = weight(_text_:aufsatzsammlung in 190) [ClassicSimilarity], result of:
          0.13198674 = score(doc=190,freq=4.0), product of:
            0.25749236 = queryWeight, product of:
              6.5610886 = idf(docFreq=169, maxDocs=44218)
              0.03924537 = queryNorm
            0.51258504 = fieldWeight in 190, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              6.5610886 = idf(docFreq=169, maxDocs=44218)
              0.0390625 = fieldNorm(doc=190)
        0.008862011 = product of:
          0.026586032 = sum of:
            0.026586032 = weight(_text_:22 in 190) [ClassicSimilarity], result of:
              0.026586032 = score(doc=190,freq=2.0), product of:
                0.13743061 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03924537 = queryNorm
                0.19345059 = fieldWeight in 190, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=190)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Date
    14. 4.2007 10:04:22
    RSWK
    Computer / Anwendung / Computerunterstützte Lexikographie / Aufsatzsammlung
    Subject
    Computer / Anwendung / Computerunterstützte Lexikographie / Aufsatzsammlung
  8. Pepper, S.; Arnaud, P.J.L.: Absolutely PHAB : toward a general model of associative relations (2020) 0.05
    0.05106003 = product of:
      0.12765007 = sum of:
        0.11870761 = weight(_text_:relation in 103) [ClassicSimilarity], result of:
          0.11870761 = score(doc=103,freq=8.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.5780928 = fieldWeight in 103, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.0390625 = fieldNorm(doc=103)
        0.008942453 = product of:
          0.02682736 = sum of:
            0.02682736 = weight(_text_:29 in 103) [ClassicSimilarity], result of:
              0.02682736 = score(doc=103,freq=2.0), product of:
                0.13805294 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03924537 = queryNorm
                0.19432661 = fieldWeight in 103, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=103)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Abstract
    There have been many attempts at classifying the semantic modification relations (R) of N + N compounds but this work has not led to the acceptance of a definitive scheme, so that devising a reusable classification is a worthwhile aim. The scope of this undertaking is extended to other binominal lexemes, i.e. units that contain two thing-morphemes without explicitly stating R, like prepositional units, N + relational adjective units, etc. The 25-relation taxonomy of Bourque (2014) was tested against over 15,000 binominal lexemes from 106 languages and extended to a 29-relation scheme ("Bourque2") through the introduction of two new reversible relations. Bourque2 is then mapped onto Hatcher's (1960) four-relation scheme (extended by the addition of a fifth relation, similarity , as "Hatcher2"). This results in a two-tier system usable at different degrees of granularities. On account of its semantic proximity to compounding, metonymy is then taken into account, following Janda's (2011) suggestion that it plays a role in word formation; Peirsman and Geeraerts' (2006) inventory of 23 metonymic patterns is mapped onto Bourque2, confirming the identity of metonymic and binominal modification relations. Finally, Blank's (2003) and Koch's (2001) work on lexical semantics justifies the addition to the scheme of a third, superordinate level which comprises the three Aristotelean principles of similarity, contiguity and contrast.
  9. Collovini de Abreu, S.; Vieira, R.: RelP: Portuguese open relation extraction (2017) 0.04
    0.037538644 = product of:
      0.18769321 = sum of:
        0.18769321 = weight(_text_:relation in 3621) [ClassicSimilarity], result of:
          0.18769321 = score(doc=3621,freq=20.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.914045 = fieldWeight in 3621, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3621)
      0.2 = coord(1/5)
    
    Abstract
    Natural language texts are valuable data sources in many human activities. NLP techniques are being widely used in order to help find the right information to specific needs. In this paper, we present one such technique: relation extraction from texts. This task aims at identifying and classifying semantic relations that occur between entities in a text. For example, the sentence "Roberto Marinho is the founder of Rede Globo" expresses a relation occurring between "Roberto Marinho" and "Rede Globo." This work presents a system for Portuguese Open Relation Extraction, named RelP, which extracts any relation descriptor that describes an explicit relation between named entities in the organisation domain by applying the Conditional Random Fields. For implementing RelP, we define the representation scheme, features based on previous work, and a reference corpus. RelP achieved state of the art results for open relation extraction; the F-measure rate was around 60% between the named entities person, organisation and place. For better understanding of the output, we present a way for organizing the output from the mining of the extracted relation descriptors. This organization can be useful to classify relation types, to cluster the entities involved in a common relation and to populate datasets.
  10. Lund, B.D.; Wang, T.; Mannuru, N.R.; Nie, B.; Shimray, S.; Wang, Z.: ChatGPT and a new academic reality : artificial Intelligence-written research papers and the ethics of the large language models in scholarly publishing (2023) 0.03
    0.03278221 = product of:
      0.081955515 = sum of:
        0.07122457 = weight(_text_:relation in 943) [ClassicSimilarity], result of:
          0.07122457 = score(doc=943,freq=2.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.3468557 = fieldWeight in 943, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.046875 = fieldNorm(doc=943)
        0.010730944 = product of:
          0.03219283 = sum of:
            0.03219283 = weight(_text_:29 in 943) [ClassicSimilarity], result of:
              0.03219283 = score(doc=943,freq=2.0), product of:
                0.13805294 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03924537 = queryNorm
                0.23319192 = fieldWeight in 943, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=943)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Abstract
    This article discusses OpenAI's ChatGPT, a generative pre-trained transformer, which uses natural language processing to fulfill text-based user requests (i.e., a "chatbot"). The history and principles behind ChatGPT and similar models are discussed. This technology is then discussed in relation to its potential impact on academia and scholarly research and publishing. ChatGPT is seen as a potential model for the automated preparation of essays and other types of scholarly manuscripts. Potential ethical issues that could arise with the emergence of large language models like GPT-3, the underlying technology behind ChatGPT, and its usage by academics and researchers, are discussed and situated within the context of broader advancements in artificial intelligence, machine learning, and natural language processing for research and scholarly publishing.
    Date
    19. 4.2023 19:29:44
  11. Dorr, B.J.: Large-scale dictionary construction for foreign language tutoring and interlingual machine translation (1997) 0.03
    0.032743596 = product of:
      0.081858985 = sum of:
        0.07122457 = weight(_text_:relation in 3244) [ClassicSimilarity], result of:
          0.07122457 = score(doc=3244,freq=2.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.3468557 = fieldWeight in 3244, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.046875 = fieldNorm(doc=3244)
        0.010634413 = product of:
          0.031903237 = sum of:
            0.031903237 = weight(_text_:22 in 3244) [ClassicSimilarity], result of:
              0.031903237 = score(doc=3244,freq=2.0), product of:
                0.13743061 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03924537 = queryNorm
                0.23214069 = fieldWeight in 3244, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3244)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Abstract
    Describes techniques for automatic construction of dictionaries for use in large-scale foreign language tutoring (FLT) and interlingual machine translation (MT) systems. The dictionaries are based on a language independent representation called lexical conceptual structure (LCS). Demonstrates that synonymous verb senses share distribution patterns. Shows how the syntax-semantics relation can be used to develop a lexical acquisition approach that contributes both toward the enrichment of existing online resources and toward the development of lexicons containing more complete information than is provided in any of these resources alone. Describes the structure of the LCS and shows how this representation is used in FLT and MT. Focuses on the problem of building LCS dictionaries for large-scale FLT and MT. Describes authoring tools for manual and semi-automatic construction of LCS dictionaries. Presents an approach that uses linguistic techniques for building word definitions automatically. The techniques have been implemented as part of a set of lixicon-development tools used in the MILT FLT project
    Date
    31. 7.1996 9:22:19
  12. Multi-source, multilingual information extraction and summarization (2013) 0.03
    0.026397347 = product of:
      0.13198674 = sum of:
        0.13198674 = weight(_text_:aufsatzsammlung in 978) [ClassicSimilarity], result of:
          0.13198674 = score(doc=978,freq=4.0), product of:
            0.25749236 = queryWeight, product of:
              6.5610886 = idf(docFreq=169, maxDocs=44218)
              0.03924537 = queryNorm
            0.51258504 = fieldWeight in 978, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              6.5610886 = idf(docFreq=169, maxDocs=44218)
              0.0390625 = fieldNorm(doc=978)
      0.2 = coord(1/5)
    
    RSWK
    Natürlichsprachiges System / Information Extraction / Automatische Inhaltsanalyse / Zusammenfassung / Aufsatzsammlung
    Subject
    Natürlichsprachiges System / Information Extraction / Automatische Inhaltsanalyse / Zusammenfassung / Aufsatzsammlung
  13. Information und Sprache : Beiträge zu Informationswissenschaft, Computerlinguistik, Bibliothekswesen und verwandten Fächern. Festschrift für Harald H. Zimmermann. Herausgegeben von Ilse Harms, Heinz-Dirk Luckhardt und Hans W. Giessen (2006) 0.02
    0.02111788 = product of:
      0.10558939 = sum of:
        0.10558939 = weight(_text_:aufsatzsammlung in 91) [ClassicSimilarity], result of:
          0.10558939 = score(doc=91,freq=16.0), product of:
            0.25749236 = queryWeight, product of:
              6.5610886 = idf(docFreq=169, maxDocs=44218)
              0.03924537 = queryNorm
            0.41006804 = fieldWeight in 91, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              6.5610886 = idf(docFreq=169, maxDocs=44218)
              0.015625 = fieldNorm(doc=91)
      0.2 = coord(1/5)
    
    RSWK
    Informations- und Dokumentationswissenschaft / Aufsatzsammlung
    Information Retrieval / Aufsatzsammlung
    Automatische Indexierung / Aufsatzsammlung
    Linguistische Datenverarbeitung / Aufsatzsammlung
    Subject
    Informations- und Dokumentationswissenschaft / Aufsatzsammlung
    Information Retrieval / Aufsatzsammlung
    Automatische Indexierung / Aufsatzsammlung
    Linguistische Datenverarbeitung / Aufsatzsammlung
  14. Grefenstette, G.: Explorations in automatic thesaurus discovery (1994) 0.02
    0.016619066 = product of:
      0.08309533 = sum of:
        0.08309533 = weight(_text_:relation in 170) [ClassicSimilarity], result of:
          0.08309533 = score(doc=170,freq=2.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.40466496 = fieldWeight in 170, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.0546875 = fieldNorm(doc=170)
      0.2 = coord(1/5)
    
    Abstract
    Review of various approaches to automatic thesaurus formation and presentation of the SEXTANT system to analyse text and to determine the basic syntactic contexts for words. Presents an automated method for creating a first-draft thesaurus from raw text. It describes natural processing steps of tokenization, surface syntactic analysis, and syntactic attribute extraction. From these attributes, word and term similarity is calculated and a thesaurus is created showing important common terms and their relation to each other, common verb-noun pairings, common expressions, and word family members
  15. Experimentelles und praktisches Information Retrieval : Festschrift für Gerhard Lustig (1992) 0.01
    0.014244914 = product of:
      0.07122457 = sum of:
        0.07122457 = weight(_text_:relation in 4) [ClassicSimilarity], result of:
          0.07122457 = score(doc=4,freq=2.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.3468557 = fieldWeight in 4, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.046875 = fieldNorm(doc=4)
      0.2 = coord(1/5)
    
    Content
    Enthält die Beiträge: SALTON, G.: Effective text understanding in information retrieval; KRAUSE, J.: Intelligentes Information retrieval; FUHR, N.: Konzepte zur Gestaltung zukünftiger Information-Retrieval-Systeme; HÜTHER, H.: Überlegungen zu einem mathematischen Modell für die Type-Token-, die Grundform-Token und die Grundform-Type-Relation; KNORZ, G.: Automatische Generierung inferentieller Links in und zwischen Hyperdokumenten; KONRAD, E.: Zur Effektivitätsbewertung von Information-Retrieval-Systemen; HENRICHS, N.: Retrievalunterstützung durch automatisch generierte Wortfelder; LÜCK, W., W. RITTBERGER u. M. SCHWANTNER: Der Einsatz des Automatischen Indexierungs- und Retrieval-System (AIR) im Fachinformationszentrum Karlsruhe; REIMER, U.: Verfahren der Automatischen Indexierung. Benötigtes Vorwissen und Ansätze zu seiner automatischen Akquisition: Ein Überblick; ENDRES-NIGGEMEYER, B.: Dokumentrepräsentation: Ein individuelles prozedurales Modell des Abstracting, des Indexierens und Klassifizierens; SEELBACH, D.: Zur Entwicklung von zwei- und mehrsprachigen lexikalischen Datenbanken und Terminologiedatenbanken; ZIMMERMANN, H.: Der Einfluß der Sprachbarrieren in Europa und Möglichkeiten zu ihrer Minderung; LENDERS, W.: Wörter zwischen Welt und Wissen; PANYR, J.: Frames, Thesauri und automatische Klassifikation (Clusteranalyse): HAHN, U.: Forschungsstrategien und Erkenntnisinteressen in der anwendungsorientierten automatischen Sprachverarbeitung. Überlegungen zu einer ingenieurorientierten Computerlinguistik; KUHLEN, R.: Hypertext und Information Retrieval - mehr als Browsing und Suche.
  16. Dork, B.J.; Garman, J.; Weinberg, A.: From syntactic encodings to thematic roles : building lexical entries for interlingual MT (1994/95) 0.01
    0.014244914 = product of:
      0.07122457 = sum of:
        0.07122457 = weight(_text_:relation in 4074) [ClassicSimilarity], result of:
          0.07122457 = score(doc=4074,freq=2.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.3468557 = fieldWeight in 4074, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.046875 = fieldNorm(doc=4074)
      0.2 = coord(1/5)
    
    Abstract
    Aims to construct large scale lexicons for interlingual machine translation of English, Arabic, Korean, and Spanish. Describes techniques that predict salient linguistic features of a non English word using the features of its English gloss in a bilingual dictionary. While not exact, owing to inexact glosses and language to language variations, these techniques can augment an existing dictionary with reasonable accuracy, thus savionf significant time. Conducts 2 experiments that demonstrate the value of these techniques. The 1st tests the feasibility of building a database of thematic grids for over 6500 Arabic verbs based on a mapping between English glosses and the syntactic codes in Longman's Dictionary of Contemporary English. The 2nd experiment tested the automatic classification of verbs into a richer semantic typology from which a more refined set of thematic grids derived. While human intervention will always be necessary for the construction of a semantic classification from LODOCE such intervention is significantly minimized as more knowledge about the syntax semantics relation is introduced
  17. Pepper, S.: ¬The typology and semantics of binominal lexemes : noun-noun compounds and their functional equivalents (2020) 0.01
    0.013430233 = product of:
      0.06715117 = sum of:
        0.06715117 = weight(_text_:relation in 104) [ClassicSimilarity], result of:
          0.06715117 = score(doc=104,freq=4.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.32701868 = fieldWeight in 104, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03125 = fieldNorm(doc=104)
      0.2 = coord(1/5)
    
    Abstract
    The dissertation establishes 'binominal lexeme' as a comparative concept and discusses its cross-linguistic typology and semantics. Informally, a binominal lexeme is a noun-noun compound or functional equivalent; more precisely, it is a lexical item that consists primarily of two thing-morphs between which there exists an unstated semantic relation. Examples of binominals include Mandarin Chinese ?? (tielù) [iron road], French chemin de fer [way of iron] and Russian ???????? ?????? (zeleznaja doroga) [iron:adjz road]. All of these combine a word denoting 'iron' and a word denoting 'road' or 'way' to denote the meaning railway. In each case, the unstated semantic relation is one of composition: a railway is conceptualized as a road that is composed (or made) of iron. However, three different morphosyntactic strategies are employed: compounding, prepositional phrase and relational adjective. This study explores the range of such strategies used by a worldwide sample of 106 languages to express a set of 100 meanings from various semantic domains, resulting in a classification consisting of nine different morphosyntactic types. The semantic relations found in the data are also explored and a classification called the Hatcher-Bourque system is developed that operates at two levels of granularity, together with a tool for classifying binominals, the Bourquifier. The classification is extended to other subfields of language, including metonymy and lexical semantics, and beyond language to the domain of knowledge representation, resulting in a proposal for a general model of associative relations called the PHAB model. The many findings of the research include universals concerning the recruitment of anchoring nominal modification strategies, a method for comparing non-binary typologies, the non-universality (despite its predominance) of compounding, and a scale of frequencies for semantic relations which may provide insights into the associative nature of human thought.
  18. Ballesteros, L.A.: Cross-language retrieval via transitive relation (2000) 0.01
    0.011870761 = product of:
      0.059353806 = sum of:
        0.059353806 = weight(_text_:relation in 30) [ClassicSimilarity], result of:
          0.059353806 = score(doc=30,freq=2.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.2890464 = fieldWeight in 30, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.0390625 = fieldNorm(doc=30)
      0.2 = coord(1/5)
    
  19. ¬The semantics of relationships : an interdisciplinary perspective (2002) 0.01
    0.011870761 = product of:
      0.059353806 = sum of:
        0.059353806 = weight(_text_:relation in 1430) [ClassicSimilarity], result of:
          0.059353806 = score(doc=1430,freq=2.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.2890464 = fieldWeight in 1430, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1430)
      0.2 = coord(1/5)
    
    Content
    Enthält die Beiträge: Pt.1: Types of relationships: CRUDE, D.A.: Hyponymy and its varieties; FELLBAUM, C.: On the semantics of troponymy; PRIBBENOW, S.: Meronymic relationships: from classical mereology to complex part-whole relations; KHOO, C. u.a.: The many facets of cause-effect relation - Pt.2: Relationships in knowledge representation and reasoning: GREEN, R.: Internally-structured conceptual models in cognitive semantics; HOVY, E.: Comparing sets of semantic relations in ontologies; GUARINO, N., C. WELTY: Identity and subsumption; JOUIS; C.: Logic of relationships - Pt.3: Applications of relationships: EVENS, M.: Thesaural relations in information retrieval; KHOO, C., S.H. MYAENG: Identifying semantic relations in text for information retrieval and information extraction; McCRAY, A.T., O. BODENREICHER: A conceptual framework for the biiomedical domain; HETZLER, B.: Visual analysis and exploration of relationships
  20. Stede, M.: Lexicalization in natural language generation (2002) 0.01
    0.011870761 = product of:
      0.059353806 = sum of:
        0.059353806 = weight(_text_:relation in 4245) [ClassicSimilarity], result of:
          0.059353806 = score(doc=4245,freq=2.0), product of:
            0.20534351 = queryWeight, product of:
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.03924537 = queryNorm
            0.2890464 = fieldWeight in 4245, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.232299 = idf(docFreq=641, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4245)
      0.2 = coord(1/5)
    
    Abstract
    Natural language generation (NLG), the automatic production of text by Computers, is commonly seen as a process consisting of the following distinct phases: Obviously, choosing words is a central aspect of generatiog language. In which of the these phases it should take place is not entirely clear, however. The decision depends an various factors: what exactly is seen as an individual lexical item; how the relation between word meaning and background knowledge (concepts) is defined; how one accounts for the interactions between individual lexical choices in the Same sentence; what criteria are employed for choosing between similar words; whether or not output is required in one or more languages. This article surveys these issues and the answers that have been proposed in NLG research. For many applications of natural language processing, large scale lexical resources have become available in recent years, such as the WordNet database. In language generation, however, generic lexicons are not in use yet; rather, almost every generation project develops its own format for lexical representations. The reason is that the entries of a generation lexicon need their specific interfaces to the Input representations processed by the generator; lexical semantics in an NLG lexicon needs to be tailored to the Input. Ort the other hand, the large lexicons used for language analysis typically have only very limited semantic information at all. Yet the syntactic behavior of words remains the same regardless of the particular application; thus, it should be possible to build at least parts of generic NLG lexical entries automatically, which could then be used by different systems.

Years

Languages

  • e 87
  • d 33
  • ru 2
  • m 1
  • More… Less…

Types

  • a 97
  • m 15
  • el 13
  • s 10
  • x 4
  • p 2
  • d 1
  • More… Less…

Subjects

Classifications