Search (4 results, page 1 of 1)

  • × theme_ss:"Literaturübersicht"
  • × theme_ss:"Computerlinguistik"
  1. Perez-Carballo, J.; Strzalkowski, T.: Natural language information retrieval : progress report (2000) 0.05
    0.05190724 = product of:
      0.10381448 = sum of:
        0.10381448 = product of:
          0.20762897 = sum of:
            0.20762897 = weight(_text_:2000 in 6421) [ClassicSimilarity], result of:
              0.20762897 = score(doc=6421,freq=5.0), product of:
                0.20949209 = queryWeight, product of:
                  4.0524464 = idf(docFreq=2088, maxDocs=44218)
                  0.051695216 = queryNorm
                0.9911065 = fieldWeight in 6421, product of:
                  2.236068 = tf(freq=5.0), with freq of:
                    5.0 = termFreq=5.0
                  4.0524464 = idf(docFreq=2088, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6421)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Information processing and management. 36(2000) no.1, S.155-205
    Year
    2000
  2. Warner, A.J.: Natural language processing (1987) 0.03
    0.028015953 = product of:
      0.056031905 = sum of:
        0.056031905 = product of:
          0.11206381 = sum of:
            0.11206381 = weight(_text_:22 in 337) [ClassicSimilarity], result of:
              0.11206381 = score(doc=337,freq=2.0), product of:
                0.18102784 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051695216 = queryNorm
                0.61904186 = fieldWeight in 337, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=337)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Annual review of information science and technology. 22(1987), S.79-108
  3. Haas, S.W.: Natural language processing : toward large-scale, robust systems (1996) 0.01
    0.014007976 = product of:
      0.028015953 = sum of:
        0.028015953 = product of:
          0.056031905 = sum of:
            0.056031905 = weight(_text_:22 in 7415) [ClassicSimilarity], result of:
              0.056031905 = score(doc=7415,freq=2.0), product of:
                0.18102784 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051695216 = queryNorm
                0.30952093 = fieldWeight in 7415, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=7415)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    State of the art review of natural language processing updating an earlier review published in ARIST 22(1987). Discusses important developments that have allowed for significant advances in the field of natural language processing: materials and resources; knowledge based systems and statistical approaches; and a strong emphasis on evaluation. Reviews some natural language processing applications and common problems still awaiting solution. Considers closely related applications such as language generation and th egeneration phase of machine translation which face the same problems as natural language processing. Covers natural language methodologies for information retrieval only briefly
  4. Liu, X.; Croft, W.B.: Statistical language modeling for information retrieval (2004) 0.01
    0.011724652 = product of:
      0.023449304 = sum of:
        0.023449304 = product of:
          0.046898607 = sum of:
            0.046898607 = weight(_text_:2000 in 4277) [ClassicSimilarity], result of:
              0.046898607 = score(doc=4277,freq=2.0), product of:
                0.20949209 = queryWeight, product of:
                  4.0524464 = idf(docFreq=2088, maxDocs=44218)
                  0.051695216 = queryNorm
                0.22386816 = fieldWeight in 4277, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.0524464 = idf(docFreq=2088, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4277)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This chapter reviews research and applications in statistical language modeling for information retrieval (IR), which has emerged within the past several years as a new probabilistic framework for describing information retrieval processes. Generally speaking, statistical language modeling, or more simply language modeling (LM), involves estimating a probability distribution that captures statistical regularities of natural language use. Applied to information retrieval, language modeling refers to the problem of estimating the likelihood that a query and a document could have been generated by the same language model, given the language model of the document either with or without a language model of the query. The roots of statistical language modeling date to the beginning of the twentieth century when Markov tried to model letter sequences in works of Russian literature (Manning & Schütze, 1999). Zipf (1929, 1932, 1949, 1965) studied the statistical properties of text and discovered that the frequency of works decays as a Power function of each works rank. However, it was Shannon's (1951) work that inspired later research in this area. In 1951, eager to explore the applications of his newly founded information theory to human language, Shannon used a prediction game involving n-grams to investigate the information content of English text. He evaluated n-gram models' performance by comparing their crossentropy an texts with the true entropy estimated using predictions made by human subjects. For many years, statistical language models have been used primarily for automatic speech recognition. Since 1980, when the first significant language model was proposed (Rosenfeld, 2000), statistical language modeling has become a fundamental component of speech recognition, machine translation, and spelling correction.