Search (57 results, page 1 of 3)

  • × language_ss:"e"
  • × theme_ss:"Computerlinguistik"
  1. Chiba, K.; Kyojima, M.: Document transformation based on syntax-directed free translation (1995) 0.08
    0.08050447 = product of:
      0.1207567 = sum of:
        0.06780831 = weight(_text_:electronic in 4069) [ClassicSimilarity], result of:
          0.06780831 = score(doc=4069,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.34555468 = fieldWeight in 4069, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.0625 = fieldNorm(doc=4069)
        0.052948397 = product of:
          0.10589679 = sum of:
            0.10589679 = weight(_text_:publishing in 4069) [ClassicSimilarity], result of:
              0.10589679 = score(doc=4069,freq=2.0), product of:
                0.24522576 = queryWeight, product of:
                  4.885643 = idf(docFreq=907, maxDocs=44218)
                  0.05019314 = queryNorm
                0.4318339 = fieldWeight in 4069, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.885643 = idf(docFreq=907, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4069)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Electronic publishing. 8(1995) no.1, S.15-29
  2. Hutchins, J.: From first conception to first demonstration : the nascent years of machine translation, 1947-1954. A chronology (1997) 0.08
    0.07917519 = product of:
      0.118762776 = sum of:
        0.08476039 = weight(_text_:electronic in 1463) [ClassicSimilarity], result of:
          0.08476039 = score(doc=1463,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.43194336 = fieldWeight in 1463, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.078125 = fieldNorm(doc=1463)
        0.03400239 = product of:
          0.06800478 = sum of:
            0.06800478 = weight(_text_:22 in 1463) [ClassicSimilarity], result of:
              0.06800478 = score(doc=1463,freq=2.0), product of:
                0.17576782 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05019314 = queryNorm
                0.38690117 = fieldWeight in 1463, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1463)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Chronicles the early history of applying electronic computers to the task of translating natural languages, from the 1st suggestions by Warren Weaver in Mar 1947 to the 1st demonstration of a working, if limited, program in Jan 1954
    Date
    31. 7.1996 9:22:19
  3. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.07
    0.06674763 = product of:
      0.10012144 = sum of:
        0.079720005 = product of:
          0.23916002 = sum of:
            0.23916002 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.23916002 = score(doc=562,freq=2.0), product of:
                0.42553797 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.05019314 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.020401431 = product of:
          0.040802862 = sum of:
            0.040802862 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.040802862 = score(doc=562,freq=2.0), product of:
                0.17576782 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05019314 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  4. Sparck Jones, K.: Synonymy and semantic classification (1986) 0.06
    0.056506928 = product of:
      0.16952078 = sum of:
        0.16952078 = weight(_text_:electronic in 1304) [ClassicSimilarity], result of:
          0.16952078 = score(doc=1304,freq=8.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.8638867 = fieldWeight in 1304, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.078125 = fieldNorm(doc=1304)
      0.33333334 = coord(1/3)
    
    LCSH
    Programming languages (Electronic computers) / Syntax
    Programming languages (Electronic computers) / Semantics
    Subject
    Programming languages (Electronic computers) / Syntax
    Programming languages (Electronic computers) / Semantics
  5. Godby, J.: WordSmith research project bridges gap between tokens and indexes (1998) 0.06
    0.05542263 = product of:
      0.08313394 = sum of:
        0.059332274 = weight(_text_:electronic in 4729) [ClassicSimilarity], result of:
          0.059332274 = score(doc=4729,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.30236036 = fieldWeight in 4729, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4729)
        0.023801671 = product of:
          0.047603343 = sum of:
            0.047603343 = weight(_text_:22 in 4729) [ClassicSimilarity], result of:
              0.047603343 = score(doc=4729,freq=2.0), product of:
                0.17576782 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05019314 = queryNorm
                0.2708308 = fieldWeight in 4729, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4729)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Reports on an OCLC natural language processing research project to develop methods for identifying terminology in unstructured electronic text, especially material associated with new cultural trends and emerging subjects. Current OCLC production software can only identify single words as indexable terms in full text documents, thus a major goal of the WordSmith project is to develop software that can automatically identify and intelligently organize phrases for uses in database indexes. By analyzing user terminology from local newspapers in the USA, the latest cultural trends and technical developments as well as personal and geographic names have been drawm out. Notes that this new vocabulary can also be mapped into reference works
    Source
    OCLC newsletter. 1998, no.234, Jul/Aug, S.22-24
  6. Computational linguistics for the new millennium : divergence or synergy? Proceedings of the International Symposium held at the Ruprecht-Karls Universität Heidelberg, 21-22 July 2000. Festschrift in honour of Peter Hellwig on the occasion of his 60th birthday (2002) 0.04
    0.039587595 = product of:
      0.059381388 = sum of:
        0.042380195 = weight(_text_:electronic in 4900) [ClassicSimilarity], result of:
          0.042380195 = score(doc=4900,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.21597168 = fieldWeight in 4900, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4900)
        0.017001195 = product of:
          0.03400239 = sum of:
            0.03400239 = weight(_text_:22 in 4900) [ClassicSimilarity], result of:
              0.03400239 = score(doc=4900,freq=2.0), product of:
                0.17576782 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05019314 = queryNorm
                0.19345059 = fieldWeight in 4900, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4900)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Content
    Contents: Manfred Klenner / Henriette Visser: Introduction - Khurshid Ahmad: Writing Linguistics: When I use a word it means what I choose it to mean - Jürgen Handke: 2000 and Beyond: The Potential of New Technologies in Linguistics - Jurij Apresjan / Igor Boguslavsky / Leonid Iomdin / Leonid Tsinman: Lexical Functions in NU: Possible Uses - Hubert Lehmann: Practical Machine Translation and Linguistic Theory - Karin Haenelt: A Contextbased Approach towards Content Processing of Electronic Documents - Petr Sgall / Eva Hajicová: Are Linguistic Frameworks Comparable? - Wolfgang Menzel: Theory and Applications in Computational Linguistics - Is there Common Ground? - Robert Porzel / Michael Strube: Towards Context-adaptive Natural Language Processing Systems - Nicoletta Calzolari: Language Resources in a Multilingual Setting: The European Perspective - Piek Vossen: Computational Linguistics for Theory and Practice.
  7. Warner, A.J.: ¬The role of linguistic analysis in full-text retrieval (1994) 0.04
    0.03955485 = product of:
      0.11866455 = sum of:
        0.11866455 = weight(_text_:electronic in 2992) [ClassicSimilarity], result of:
          0.11866455 = score(doc=2992,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.6047207 = fieldWeight in 2992, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.109375 = fieldNorm(doc=2992)
      0.33333334 = coord(1/3)
    
    Source
    Challenges in indexing electronic text and images. Ed.: R. Fidel et al
  8. Priß, U.: ¬The formalization of WordNet by methods of relational concept analysis (1998) 0.03
    0.033904158 = product of:
      0.101712465 = sum of:
        0.101712465 = weight(_text_:electronic in 3079) [ClassicSimilarity], result of:
          0.101712465 = score(doc=3079,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.518332 = fieldWeight in 3079, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.09375 = fieldNorm(doc=3079)
      0.33333334 = coord(1/3)
    
    Source
    WordNet: an electronic lexical database (language, speech and communication). Ed.: C. Fellbaum
  9. ¬The language engineering directory (1993) 0.03
    0.028253464 = product of:
      0.08476039 = sum of:
        0.08476039 = weight(_text_:electronic in 8408) [ClassicSimilarity], result of:
          0.08476039 = score(doc=8408,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.43194336 = fieldWeight in 8408, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.078125 = fieldNorm(doc=8408)
      0.33333334 = coord(1/3)
    
    Abstract
    This is a reference guide to language technology organizations and products around the world. Areas covered in the directory include: Artificial intelligence, Document storage and retrieval, Electronic dictionaries (mono- and multilingual), Expert language systems, Multilingual word processors, Natural language database interfaces, Term databanks, Terminology management, Text content analysis, Thesauri
  10. Zimmermann, H.H.: Language and language technology (1991) 0.03
    0.028253464 = product of:
      0.08476039 = sum of:
        0.08476039 = weight(_text_:electronic in 2568) [ClassicSimilarity], result of:
          0.08476039 = score(doc=2568,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.43194336 = fieldWeight in 2568, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.078125 = fieldNorm(doc=2568)
      0.33333334 = coord(1/3)
    
    Abstract
    Considers aspects of language and linguistic studies that directly affect information handling including: electronic word processing (hyphenation, spelling correction, dictionary-based synonym provision); man-machine communication; machine understanding of spoken language; automatic indexing; and machine translation
  11. WordNet : an electronic lexical database (language, speech and communication) (1998) 0.03
    0.027969502 = product of:
      0.083908506 = sum of:
        0.083908506 = weight(_text_:electronic in 2434) [ClassicSimilarity], result of:
          0.083908506 = score(doc=2434,freq=4.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.4276021 = fieldWeight in 2434, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2434)
      0.33333334 = coord(1/3)
    
    Abstract
    WordNet, an electronic lexical database, is considerd to be the most important resource available to researchers in computational linguistics, text analysis, and many related areas. English nouns, verbs, adjectives, and adverbs are organized into synonym sets, each representing one underlying lexicalized concept. Different relations link the synonym sets. The purpose of this volume is twofold. First, it discusses the design of WordNet and the theoretical motivation behind it. Second, it provides a survey of representative applications, including word sense identification, information retrieval, selectional preferences of verbs, and lexical chains
  12. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.03
    0.026573336 = product of:
      0.079720005 = sum of:
        0.079720005 = product of:
          0.23916002 = sum of:
            0.23916002 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.23916002 = score(doc=862,freq=2.0), product of:
                0.42553797 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.05019314 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  13. Santana Suárez, O.; Carreras Riudavets, F.J.; Hernández Figueroa, Z.; González Cabrera, A.C.: Integration of an XML electronic dictionary with linguistic tools for natural language processing (2007) 0.02
    0.02397386 = product of:
      0.07192158 = sum of:
        0.07192158 = weight(_text_:electronic in 921) [ClassicSimilarity], result of:
          0.07192158 = score(doc=921,freq=4.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.3665161 = fieldWeight in 921, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.046875 = fieldNorm(doc=921)
      0.33333334 = coord(1/3)
    
    Abstract
    This study proposes the codification of lexical information in electronic dictionaries, in accordance with a generic and extendable XML scheme model, and its conjunction with linguistic tools for the processing of natural language. Our approach is different from other similar studies in that we propose XML coding of those items from a dictionary of meanings that are less related to the lexical units. Linguistic information, such as morphology, syllables, phonology, etc., will be included by means of specific linguistic tools. The use of XML as a container for the information allows the use of other XML tools for carrying out searches or for enabling presentation of the information in different resources. This model is particularly important as it combines two parallel paradigms-extendable labelling of documents and computational linguistics-and it is also applicable to other languages. We have included a comparison with the labelling proposal of printed dictionaries carried out by the Text Encoding Initiative (TEI). The proposed design has been validated with a dictionary of more than 145 000 accepted meanings.
  14. Lund, B.D.; Wang, T.; Mannuru, N.R.; Nie, B.; Shimray, S.; Wang, Z.: ChatGPT and a new academic reality : artificial Intelligence-written research papers and the ethics of the large language models in scholarly publishing (2023) 0.02
    0.022927333 = product of:
      0.068781994 = sum of:
        0.068781994 = product of:
          0.13756399 = sum of:
            0.13756399 = weight(_text_:publishing in 943) [ClassicSimilarity], result of:
              0.13756399 = score(doc=943,freq=6.0), product of:
                0.24522576 = queryWeight, product of:
                  4.885643 = idf(docFreq=907, maxDocs=44218)
                  0.05019314 = queryNorm
                0.56096876 = fieldWeight in 943, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.885643 = idf(docFreq=907, maxDocs=44218)
                  0.046875 = fieldNorm(doc=943)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This article discusses OpenAI's ChatGPT, a generative pre-trained transformer, which uses natural language processing to fulfill text-based user requests (i.e., a "chatbot"). The history and principles behind ChatGPT and similar models are discussed. This technology is then discussed in relation to its potential impact on academia and scholarly research and publishing. ChatGPT is seen as a potential model for the automated preparation of essays and other types of scholarly manuscripts. Potential ethical issues that could arise with the emergence of large language models like GPT-3, the underlying technology behind ChatGPT, and its usage by academics and researchers, are discussed and situated within the context of broader advancements in artificial intelligence, machine learning, and natural language processing for research and scholarly publishing.
  15. Renouf, A.: Sticking to the text : a corpus linguist's view of language (1993) 0.02
    0.019777425 = product of:
      0.059332274 = sum of:
        0.059332274 = weight(_text_:electronic in 2314) [ClassicSimilarity], result of:
          0.059332274 = score(doc=2314,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.30236036 = fieldWeight in 2314, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2314)
      0.33333334 = coord(1/3)
    
    Abstract
    Corpus linguistics is the study of large, computer held bodies of text. Some corpus linguists are concerned with language descriptions for its own sake. On the corpus-linguistic continuum, the study of raw ASCII text is situated at one end, and the study of heavily pre-coded text at the other. Discusses the use of word frequency to identify changes in the lexicon; word repetition and word positioning in automatic abstracting and word clusters in automatic text retrieval. Compares the machine extract with manual abstracts. Abstractors and indexers may find themselves taking the original wording of the text more into account as the focus moves towards the electronic medium and away from the hard copy
  16. Mock, K.J.; Vemuri, V.R.: Information filtering via hill climbing, WordNet, and index patterns (1997) 0.02
    0.019777425 = product of:
      0.059332274 = sum of:
        0.059332274 = weight(_text_:electronic in 1517) [ClassicSimilarity], result of:
          0.059332274 = score(doc=1517,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.30236036 = fieldWeight in 1517, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1517)
      0.33333334 = coord(1/3)
    
    Footnote
    Contribution to a special issue devoted to electronic newspapers
  17. Warner, A.J.: Natural language processing (1987) 0.02
    0.018134607 = product of:
      0.05440382 = sum of:
        0.05440382 = product of:
          0.10880764 = sum of:
            0.10880764 = weight(_text_:22 in 337) [ClassicSimilarity], result of:
              0.10880764 = score(doc=337,freq=2.0), product of:
                0.17576782 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05019314 = queryNorm
                0.61904186 = fieldWeight in 337, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=337)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    Annual review of information science and technology. 22(1987), S.79-108
  18. Lund, B.D.: ¬A chat with ChatGPT : how will AI impact scholarly publishing? (2022) 0.02
    0.017649466 = product of:
      0.052948397 = sum of:
        0.052948397 = product of:
          0.10589679 = sum of:
            0.10589679 = weight(_text_:publishing in 850) [ClassicSimilarity], result of:
              0.10589679 = score(doc=850,freq=2.0), product of:
                0.24522576 = queryWeight, product of:
                  4.885643 = idf(docFreq=907, maxDocs=44218)
                  0.05019314 = queryNorm
                0.4318339 = fieldWeight in 850, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.885643 = idf(docFreq=907, maxDocs=44218)
                  0.0625 = fieldNorm(doc=850)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
  19. Chowdhury, G.G.: Natural language processing (2002) 0.02
    0.016952079 = product of:
      0.050856233 = sum of:
        0.050856233 = weight(_text_:electronic in 4284) [ClassicSimilarity], result of:
          0.050856233 = score(doc=4284,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.259166 = fieldWeight in 4284, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.046875 = fieldNorm(doc=4284)
      0.33333334 = coord(1/3)
    
    Abstract
    Natural Language Processing (NLP) is an area of research and application that explores how computers can be used to understand and manipulate natural language text or speech to do useful things. NLP researchers aim to gather knowledge an how human beings understand and use language so that appropriate tools and techniques can be developed to make computer systems understand and manipulate natural languages to perform desired tasks. The foundations of NLP lie in a number of disciplines, namely, computer and information sciences, linguistics, mathematics, electrical and electronic engineering, artificial intelligence and robotics, and psychology. Applications of NLP include a number of fields of study, such as machine translation, natural language text processing and summarization, user interfaces, multilingual and cross-language information retrieval (CLIR), speech recognition, artificial intelligence, and expert systems. One important application area that is relatively new and has not been covered in previous ARIST chapters an NLP relates to the proliferation of the World Wide Web and digital libraries.
  20. Galvez, C.; Moya-Anegón, F. de: ¬An evaluation of conflation accuracy using finite-state transducers (2006) 0.02
    0.016952079 = product of:
      0.050856233 = sum of:
        0.050856233 = weight(_text_:electronic in 5599) [ClassicSimilarity], result of:
          0.050856233 = score(doc=5599,freq=2.0), product of:
            0.19623034 = queryWeight, product of:
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.05019314 = queryNorm
            0.259166 = fieldWeight in 5599, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9095051 = idf(docFreq=2409, maxDocs=44218)
              0.046875 = fieldNorm(doc=5599)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - To evaluate the accuracy of conflation methods based on finite-state transducers (FSTs). Design/methodology/approach - Incorrectly lemmatized and stemmed forms may lead to the retrieval of inappropriate documents. Experimental studies to date have focused on retrieval performance, but very few on conflation performance. The process of normalization we used involved a linguistic toolbox that allowed us to construct, through graphic interfaces, electronic dictionaries represented internally by FSTs. The lexical resources developed were applied to a Spanish test corpus for merging term variants in canonical lemmatized forms. Conflation performance was evaluated in terms of an adaptation of recall and precision measures, based on accuracy and coverage, not actual retrieval. The results were compared with those obtained using a Spanish version of the Porter algorithm. Findings - The conclusion is that the main strength of lemmatization is its accuracy, whereas its main limitation is the underanalysis of variant forms. Originality/value - The report outlines the potential of transducers in their application to normalization processes.

Years

Types

  • a 43
  • m 7
  • p 4
  • el 3
  • s 3
  • x 1
  • More… Less…

Classifications