Search (230 results, page 2 of 12)

  • × theme_ss:"Computerlinguistik"
  • × year_i:[2000 TO 2010}
  1. Stock, W.G.: Textwortmethode (2000) 0.01
    0.012642393 = product of:
      0.025284786 = sum of:
        0.025284786 = product of:
          0.037927177 = sum of:
            0.0067215143 = weight(_text_:a in 3408) [ClassicSimilarity], result of:
              0.0067215143 = score(doc=3408,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.12739488 = fieldWeight in 3408, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3408)
            0.031205663 = weight(_text_:h in 3408) [ClassicSimilarity], result of:
              0.031205663 = score(doc=3408,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.27449545 = fieldWeight in 3408, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3408)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Auf dem Weg zur Informationskultur: Wa(h)re Information? Festschrift für Norbert Henrichs zum 65. Geburtstag, Hrsg.: T.A. Schröder
    Type
    a
  2. Gombocz, W.L.: Stichwort oder Schlagwort versus Textwort : Grazer und Düsseldorfer Philosophie-Dokumentation und -Information nach bzw. gemäß Norbert Henrichs (2000) 0.01
    0.012642393 = product of:
      0.025284786 = sum of:
        0.025284786 = product of:
          0.037927177 = sum of:
            0.0067215143 = weight(_text_:a in 3413) [ClassicSimilarity], result of:
              0.0067215143 = score(doc=3413,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.12739488 = fieldWeight in 3413, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3413)
            0.031205663 = weight(_text_:h in 3413) [ClassicSimilarity], result of:
              0.031205663 = score(doc=3413,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.27449545 = fieldWeight in 3413, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3413)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Auf dem Weg zur Informationskultur: Wa(h)re Information? Festschrift für Norbert Henrichs zum 65. Geburtstag, Hrsg.: T.A. Schröder
    Type
    a
  3. Perera, P.; Witte, R.: ¬A self-learning context-aware lemmatizer for German (2005) 0.01
    0.012329447 = product of:
      0.024658894 = sum of:
        0.024658894 = product of:
          0.03698834 = sum of:
            0.012023811 = weight(_text_:a in 4638) [ClassicSimilarity], result of:
              0.012023811 = score(doc=4638,freq=10.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.22789092 = fieldWeight in 4638, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4638)
            0.02496453 = weight(_text_:h in 4638) [ClassicSimilarity], result of:
              0.02496453 = score(doc=4638,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 4638, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4638)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Abstract
    Accurate lemmatization of German nouns mandates the use of a lexicon. Comprehensive lexicons, however, are expensive to build and maintain. We present a self-learning lemmatizer capable of automatically creating a full-form lexicon by processing German documents.
    Content
    Vgl. unter: http://acl.ldc.upenn.edu//H/H05/H05-1080.pdf.
    Type
    a
  4. Radev, D.; Fan, W.; Qu, H.; Wu, H.; Grewal, A.: Probabilistic question answering on the Web (2005) 0.01
    0.01211915 = product of:
      0.0242383 = sum of:
        0.0242383 = product of:
          0.036357448 = sum of:
            0.009878568 = weight(_text_:a in 3455) [ClassicSimilarity], result of:
              0.009878568 = score(doc=3455,freq=12.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.18723148 = fieldWeight in 3455, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3455)
            0.026478881 = weight(_text_:h in 3455) [ClassicSimilarity], result of:
              0.026478881 = score(doc=3455,freq=4.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.2329171 = fieldWeight in 3455, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3455)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Abstract
    Web-based search engines such as Google and NorthernLight return documents that are relevant to a user query, not answers to user questions. We have developed an architecture that augments existing search engines so that they support natural language question answering. The process entails five steps: query modulation, document retrieval, passage extraction, phrase extraction, and answer ranking. In this article, we describe some probabilistic approaches to the last three of these stages. We show how our techniques apply to a number of existing search engines, and we also present results contrasting three different methods for question answering. Our algorithm, probabilistic phrase reranking (PPR), uses proximity and question type features and achieves a total reciprocal document rank of .20 an the TREC8 corpus. Our techniques have been implemented as a Web-accessible system, called NSIR.
    Type
    a
  5. Neumann, H.: Inszenierung und Metabotschaften eines periodisch getakteten Fernsehauftritts : Die Neujahrsansprachen der Bundeskanzler Helmut Kohl und Gerhard Schröder im Vergleich (2003) 0.01
    0.011865697 = product of:
      0.023731394 = sum of:
        0.023731394 = product of:
          0.03559709 = sum of:
            0.0047050603 = weight(_text_:a in 1632) [ClassicSimilarity], result of:
              0.0047050603 = score(doc=1632,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.089176424 = fieldWeight in 1632, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1632)
            0.03089203 = weight(_text_:h in 1632) [ClassicSimilarity], result of:
              0.03089203 = score(doc=1632,freq=4.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.27173662 = fieldWeight in 1632, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1632)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Information - Wissenschaft und Praxis. 54(2003) H.5, S.261-272
    Type
    a
  6. Sienel, J.; Weiss, M.; Laube, M.: Sprachtechnologien für die Informationsgesellschaft des 21. Jahrhunderts (2000) 0.01
    0.011452909 = product of:
      0.022905817 = sum of:
        0.022905817 = product of:
          0.034358725 = sum of:
            0.0033607571 = weight(_text_:a in 5557) [ClassicSimilarity], result of:
              0.0033607571 = score(doc=5557,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.06369744 = fieldWeight in 5557, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5557)
            0.030997967 = weight(_text_:22 in 5557) [ClassicSimilarity], result of:
              0.030997967 = score(doc=5557,freq=2.0), product of:
                0.16023713 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045758117 = queryNorm
                0.19345059 = fieldWeight in 5557, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5557)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Date
    26.12.2000 13:22:17
    Type
    a
  7. Pinker, S.: Wörter und Regeln : Die Natur der Sprache (2000) 0.01
    0.011452909 = product of:
      0.022905817 = sum of:
        0.022905817 = product of:
          0.034358725 = sum of:
            0.0033607571 = weight(_text_:a in 734) [ClassicSimilarity], result of:
              0.0033607571 = score(doc=734,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.06369744 = fieldWeight in 734, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=734)
            0.030997967 = weight(_text_:22 in 734) [ClassicSimilarity], result of:
              0.030997967 = score(doc=734,freq=2.0), product of:
                0.16023713 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045758117 = queryNorm
                0.19345059 = fieldWeight in 734, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=734)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Date
    19. 7.2002 14:22:31
    Footnote
    Rez. in: Franfurter Rundschau Nr.43 vom 20.2.2001, S.23 (A. Barthelmy)
  8. Moisl, H.: Artificial neural networks and Natural Language Processing (2009) 0.01
    0.011426046 = product of:
      0.022852091 = sum of:
        0.022852091 = product of:
          0.034278136 = sum of:
            0.009313605 = weight(_text_:a in 3138) [ClassicSimilarity], result of:
              0.009313605 = score(doc=3138,freq=6.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.17652355 = fieldWeight in 3138, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3138)
            0.02496453 = weight(_text_:h in 3138) [ClassicSimilarity], result of:
              0.02496453 = score(doc=3138,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 3138, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3138)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Abstract
    This entry gives an overview of work to date on natural language processing (NLP) using artificial neural networks (ANN). It is in three main parts: the first gives a brief introduction to ANNs, the second outlines some of the main issues in ANN-based NLP, and the third surveys specific application areas. Each part cites a representative selection of research literature that itself contains pointers to further reading.
    Type
    a
  9. Ruiz, M.E.; Srinivasan, P.: Combining machine learning and hierarchical indexing structures for text categorization (2001) 0.01
    0.010418028 = product of:
      0.020836055 = sum of:
        0.020836055 = product of:
          0.031254083 = sum of:
            0.009410121 = weight(_text_:a in 1595) [ClassicSimilarity], result of:
              0.009410121 = score(doc=1595,freq=8.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.17835285 = fieldWeight in 1595, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1595)
            0.021843962 = weight(_text_:h in 1595) [ClassicSimilarity], result of:
              0.021843962 = score(doc=1595,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.19214681 = fieldWeight in 1595, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1595)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Abstract
    This paper presents a method that exploits the hierarchical structure of an indexing vocabulary to guide the development and training of machine learning methods for automatic text categorization. We present the design of a hierarchical classifier based an the divide-and-conquer principle. The method is evaluated using backpropagation neural networks, such as the machine learning algorithm, that leam to assign MeSH categories to a subset of MEDLINE records. Comparisons with traditional Rocchio's algorithm adapted for text categorization, as well as flat neural network classifiers, are provided. The results indicate that the use of hierarchical structures improves Performance significantly.
    Source
    Advances in classification research, vol.10: proceedings of the 10th ASIS SIG/CR Classification Research Workshop. Ed.: Albrechtsen, H. u. J.E. Mai
    Type
    a
  10. Nie wieder nachschlagen? : Hexaglot Blitz Translator (2000) 0.01
    0.010113914 = product of:
      0.020227827 = sum of:
        0.020227827 = product of:
          0.03034174 = sum of:
            0.0053772116 = weight(_text_:a in 7442) [ClassicSimilarity], result of:
              0.0053772116 = score(doc=7442,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.10191591 = fieldWeight in 7442, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=7442)
            0.02496453 = weight(_text_:h in 7442) [ClassicSimilarity], result of:
              0.02496453 = score(doc=7442,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 7442, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=7442)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    CD-Info. 2000, H.11, S.18
    Type
    a
  11. Stock, W.G.: Textwortmethode : Norbert Henrichs zum 65. (3) (2000) 0.01
    0.010113914 = product of:
      0.020227827 = sum of:
        0.020227827 = product of:
          0.03034174 = sum of:
            0.0053772116 = weight(_text_:a in 4891) [ClassicSimilarity], result of:
              0.0053772116 = score(doc=4891,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.10191591 = fieldWeight in 4891, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4891)
            0.02496453 = weight(_text_:h in 4891) [ClassicSimilarity], result of:
              0.02496453 = score(doc=4891,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 4891, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4891)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Password. 2000, H.7/8, S.26-35
    Type
    a
  12. Geisel, S.: ¬Der Traum von der Abschaffung der Wörter : Das Pfingstwunder und die babylonische Sprachverwirrung im Zeitalter des Internets (2001) 0.01
    0.010113914 = product of:
      0.020227827 = sum of:
        0.020227827 = product of:
          0.03034174 = sum of:
            0.0053772116 = weight(_text_:a in 4963) [ClassicSimilarity], result of:
              0.0053772116 = score(doc=4963,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.10191591 = fieldWeight in 4963, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4963)
            0.02496453 = weight(_text_:h in 4963) [ClassicSimilarity], result of:
              0.02496453 = score(doc=4963,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 4963, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4963)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Online Mitteilungen. 2001, Nr.71, S.12-19 [=Mitteilungen VÖB 54(2001) H.4]
    Type
    a
  13. Bager, J.: Teurer Dolmetscher : Forschungsprojekt Verbmobil - Rückblick und Ausblick (2001) 0.01
    0.010113914 = product of:
      0.020227827 = sum of:
        0.020227827 = product of:
          0.03034174 = sum of:
            0.0053772116 = weight(_text_:a in 5659) [ClassicSimilarity], result of:
              0.0053772116 = score(doc=5659,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.10191591 = fieldWeight in 5659, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5659)
            0.02496453 = weight(_text_:h in 5659) [ClassicSimilarity], result of:
              0.02496453 = score(doc=5659,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 5659, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5659)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    c't. 2001, H.26, S.xxx
    Type
    a
  14. Schneider, R.: Question answering : das Retrieval der Zukunft? (2007) 0.01
    0.010113914 = product of:
      0.020227827 = sum of:
        0.020227827 = product of:
          0.03034174 = sum of:
            0.0053772116 = weight(_text_:a in 5953) [ClassicSimilarity], result of:
              0.0053772116 = score(doc=5953,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.10191591 = fieldWeight in 5953, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5953)
            0.02496453 = weight(_text_:h in 5953) [ClassicSimilarity], result of:
              0.02496453 = score(doc=5953,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 5953, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5953)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Zeitschrift für Bibliothekswesen und Bibliographie. 54(2007) H.1, S.3-11
    Type
    a
  15. Hickstein, A.C.: Spracheingabe in ein feldstrukturiertes Datenbanksystem zum Einsatz im Wissensmanagement (2008) 0.01
    0.010113914 = product of:
      0.020227827 = sum of:
        0.020227827 = product of:
          0.03034174 = sum of:
            0.0053772116 = weight(_text_:a in 2532) [ClassicSimilarity], result of:
              0.0053772116 = score(doc=2532,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.10191591 = fieldWeight in 2532, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2532)
            0.02496453 = weight(_text_:h in 2532) [ClassicSimilarity], result of:
              0.02496453 = score(doc=2532,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 2532, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2532)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Information - Wissenschaft und Praxis. 59(2008) H.3, S.171-174
    Type
    a
  16. Kunze, C.: Lexikalisch-semantische Wortnetze in Sprachwissenschaft und Sprachtechnologie (2006) 0.01
    0.010113914 = product of:
      0.020227827 = sum of:
        0.020227827 = product of:
          0.03034174 = sum of:
            0.0053772116 = weight(_text_:a in 6023) [ClassicSimilarity], result of:
              0.0053772116 = score(doc=6023,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.10191591 = fieldWeight in 6023, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6023)
            0.02496453 = weight(_text_:h in 6023) [ClassicSimilarity], result of:
              0.02496453 = score(doc=6023,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 6023, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6023)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Information - Wissenschaft und Praxis. 57(2006) H.6/7, S.309-314
    Type
    a
  17. NUANCE XT9 : Neue Schreibhilfe für Mobiltelefone (2008) 0.01
    0.010113914 = product of:
      0.020227827 = sum of:
        0.020227827 = product of:
          0.03034174 = sum of:
            0.0053772116 = weight(_text_:a in 2268) [ClassicSimilarity], result of:
              0.0053772116 = score(doc=2268,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.10191591 = fieldWeight in 2268, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2268)
            0.02496453 = weight(_text_:h in 2268) [ClassicSimilarity], result of:
              0.02496453 = score(doc=2268,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 2268, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2268)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Chip. 2008, H.4, S.52
    Type
    a
  18. Nhongkai, S.N.; Bentz, H.-J.: Bilinguale Suche mittels Konzeptnetzen (2006) 0.01
    0.010113914 = product of:
      0.020227827 = sum of:
        0.020227827 = product of:
          0.03034174 = sum of:
            0.0053772116 = weight(_text_:a in 3914) [ClassicSimilarity], result of:
              0.0053772116 = score(doc=3914,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.10191591 = fieldWeight in 3914, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3914)
            0.02496453 = weight(_text_:h in 3914) [ClassicSimilarity], result of:
              0.02496453 = score(doc=3914,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 3914, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3914)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Type
    a
  19. Chen, K.-H.: Evaluating Chinese text retrieval with multilingual queries (2002) 0.01
    0.009997789 = product of:
      0.019995578 = sum of:
        0.019995578 = product of:
          0.029993366 = sum of:
            0.008149404 = weight(_text_:a in 1851) [ClassicSimilarity], result of:
              0.008149404 = score(doc=1851,freq=6.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.1544581 = fieldWeight in 1851, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1851)
            0.021843962 = weight(_text_:h in 1851) [ClassicSimilarity], result of:
              0.021843962 = score(doc=1851,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.19214681 = fieldWeight in 1851, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1851)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Abstract
    This paper reports the design of a Chinese test collection with multilingual queries and the application of this test collection to evaluate information retrieval Systems. The effective indexing units, IR models, translation techniques, and query expansion for Chinese text retrieval are identified. The collaboration of East Asian countries for construction of test collections for cross-language multilingual text retrieval is also discussed in this paper. As well, a tool is designed to help assessors judge relevante and gather the events of relevante judgment. The log file created by this tool will be used to analyze the behaviors of assessors in the future.
    Type
    a
  20. Kuo, J.-S.; Li, H.; Yang, Y.-K.: Active learning for constructing transliteration lexicons from the Web (2008) 0.01
    0.009997789 = product of:
      0.019995578 = sum of:
        0.019995578 = product of:
          0.029993366 = sum of:
            0.008149404 = weight(_text_:a in 1345) [ClassicSimilarity], result of:
              0.008149404 = score(doc=1345,freq=6.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.1544581 = fieldWeight in 1345, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1345)
            0.021843962 = weight(_text_:h in 1345) [ClassicSimilarity], result of:
              0.021843962 = score(doc=1345,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.19214681 = fieldWeight in 1345, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1345)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Abstract
    This article presents an adaptive learning framework for Phonetic Similarity Modeling (PSM) that supports the automatic construction of transliteration lexicons. The learning algorithm starts with minimum prior knowledge about machine transliteration and acquires knowledge iteratively from the Web. We study the unsupervised learning and the active learning strategies that minimize human supervision in terms of data labeling. The learning process refines the PSM and constructs a transliteration lexicon at the same time. We evaluate the proposed PSM and its learning algorithm through a series of systematic experiments, which show that the proposed framework is reliably effective on two independent databases.
    Type
    a

Authors

Languages

  • e 162
  • d 61
  • ru 5
  • m 2
  • slv 1
  • More… Less…

Types

  • a 208
  • m 17
  • s 9
  • el 8
  • x 3
  • More… Less…