Search (69 results, page 1 of 4)

  • × theme_ss:"Computerlinguistik"
  • × language_ss:"e"
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.07
    0.07259655 = sum of:
      0.054127328 = product of:
        0.21650931 = sum of:
          0.21650931 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
            0.21650931 = score(doc=562,freq=2.0), product of:
              0.38523552 = queryWeight, product of:
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.045439374 = queryNorm
              0.56201804 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.25 = coord(1/4)
      0.018469224 = product of:
        0.036938448 = sum of:
          0.036938448 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
            0.036938448 = score(doc=562,freq=2.0), product of:
              0.15912095 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.045439374 = queryNorm
              0.23214069 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.5 = coord(1/2)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  2. Schneider, J.W.; Borlund, P.: ¬A bibliometric-based semiautomatic approach to identification of candidate thesaurus terms : parsing and filtering of noun phrases from citation contexts (2005) 0.05
    0.04654435 = product of:
      0.0930887 = sum of:
        0.0930887 = sum of:
          0.049993843 = weight(_text_:i in 156) [ClassicSimilarity], result of:
            0.049993843 = score(doc=156,freq=2.0), product of:
              0.17138503 = queryWeight, product of:
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.045439374 = queryNorm
              0.29170483 = fieldWeight in 156, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.0546875 = fieldNorm(doc=156)
          0.043094855 = weight(_text_:22 in 156) [ClassicSimilarity], result of:
            0.043094855 = score(doc=156,freq=2.0), product of:
              0.15912095 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.045439374 = queryNorm
              0.2708308 = fieldWeight in 156, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=156)
      0.5 = coord(1/2)
    
    Date
    8. 3.2007 19:55:22
    Source
    Context: nature, impact and role. 5th International Conference an Conceptions of Library and Information Sciences, CoLIS 2005 Glasgow, UK, June 2005. Ed. by F. Crestani u. I. Ruthven
  3. Computational linguistics for the new millennium : divergence or synergy? Proceedings of the International Symposium held at the Ruprecht-Karls Universität Heidelberg, 21-22 July 2000. Festschrift in honour of Peter Hellwig on the occasion of his 60th birthday (2002) 0.04
    0.040641725 = product of:
      0.08128345 = sum of:
        0.08128345 = sum of:
          0.050501406 = weight(_text_:i in 4900) [ClassicSimilarity], result of:
            0.050501406 = score(doc=4900,freq=4.0), product of:
              0.17138503 = queryWeight, product of:
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.045439374 = queryNorm
              0.29466638 = fieldWeight in 4900, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4900)
          0.03078204 = weight(_text_:22 in 4900) [ClassicSimilarity], result of:
            0.03078204 = score(doc=4900,freq=2.0), product of:
              0.15912095 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.045439374 = queryNorm
              0.19345059 = fieldWeight in 4900, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4900)
      0.5 = coord(1/2)
    
    Content
    Contents: Manfred Klenner / Henriette Visser: Introduction - Khurshid Ahmad: Writing Linguistics: When I use a word it means what I choose it to mean - Jürgen Handke: 2000 and Beyond: The Potential of New Technologies in Linguistics - Jurij Apresjan / Igor Boguslavsky / Leonid Iomdin / Leonid Tsinman: Lexical Functions in NU: Possible Uses - Hubert Lehmann: Practical Machine Translation and Linguistic Theory - Karin Haenelt: A Contextbased Approach towards Content Processing of Electronic Documents - Petr Sgall / Eva Hajicová: Are Linguistic Frameworks Comparable? - Wolfgang Menzel: Theory and Applications in Computational Linguistics - Is there Common Ground? - Robert Porzel / Michael Strube: Towards Context-adaptive Natural Language Processing Systems - Nicoletta Calzolari: Language Resources in a Multilingual Setting: The European Perspective - Piek Vossen: Computational Linguistics for Theory and Practice.
  4. Solvberg, I.; Nordbo, I.; Aamodt, A.: Knowledge-based information retrieval (1991/92) 0.04
    0.040401127 = product of:
      0.080802254 = sum of:
        0.080802254 = product of:
          0.16160451 = sum of:
            0.16160451 = weight(_text_:i in 546) [ClassicSimilarity], result of:
              0.16160451 = score(doc=546,freq=4.0), product of:
                0.17138503 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.045439374 = queryNorm
                0.9429324 = fieldWeight in 546, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.125 = fieldNorm(doc=546)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  5. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.03
    0.027063664 = product of:
      0.054127328 = sum of:
        0.054127328 = product of:
          0.21650931 = sum of:
            0.21650931 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.21650931 = score(doc=862,freq=2.0), product of:
                0.38523552 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.045439374 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.25 = coord(1/4)
      0.5 = coord(1/2)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  6. Feldman, S.: Find what I mean, not what I say : meaning-based search tools (2000) 0.03
    0.025250703 = product of:
      0.050501406 = sum of:
        0.050501406 = product of:
          0.10100281 = sum of:
            0.10100281 = weight(_text_:i in 4799) [ClassicSimilarity], result of:
              0.10100281 = score(doc=4799,freq=4.0), product of:
                0.17138503 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.045439374 = queryNorm
                0.58933276 = fieldWeight in 4799, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4799)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  7. Warner, A.J.: Natural language processing (1987) 0.02
    0.024625631 = product of:
      0.049251262 = sum of:
        0.049251262 = product of:
          0.098502524 = sum of:
            0.098502524 = weight(_text_:22 in 337) [ClassicSimilarity], result of:
              0.098502524 = score(doc=337,freq=2.0), product of:
                0.15912095 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045439374 = queryNorm
                0.61904186 = fieldWeight in 337, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=337)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Annual review of information science and technology. 22(1987), S.79-108
  8. McMahon, J.G.; Smith, F.J.: Improved statistical language model performance with automatic generated word hierarchies (1996) 0.02
    0.021547427 = product of:
      0.043094855 = sum of:
        0.043094855 = product of:
          0.08618971 = sum of:
            0.08618971 = weight(_text_:22 in 3164) [ClassicSimilarity], result of:
              0.08618971 = score(doc=3164,freq=2.0), product of:
                0.15912095 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045439374 = queryNorm
                0.5416616 = fieldWeight in 3164, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3164)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Computational linguistics. 22(1996) no.2, S.217-248
  9. Ruge, G.: ¬A spreading activation network for automatic generation of thesaurus relationships (1991) 0.02
    0.021547427 = product of:
      0.043094855 = sum of:
        0.043094855 = product of:
          0.08618971 = sum of:
            0.08618971 = weight(_text_:22 in 4506) [ClassicSimilarity], result of:
              0.08618971 = score(doc=4506,freq=2.0), product of:
                0.15912095 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045439374 = queryNorm
                0.5416616 = fieldWeight in 4506, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4506)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    8.10.2000 11:52:22
  10. Somers, H.: Example-based machine translation : Review article (1999) 0.02
    0.021547427 = product of:
      0.043094855 = sum of:
        0.043094855 = product of:
          0.08618971 = sum of:
            0.08618971 = weight(_text_:22 in 6672) [ClassicSimilarity], result of:
              0.08618971 = score(doc=6672,freq=2.0), product of:
                0.15912095 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045439374 = queryNorm
                0.5416616 = fieldWeight in 6672, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6672)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    31. 7.1996 9:22:19
  11. New tools for human translators (1997) 0.02
    0.021547427 = product of:
      0.043094855 = sum of:
        0.043094855 = product of:
          0.08618971 = sum of:
            0.08618971 = weight(_text_:22 in 1179) [ClassicSimilarity], result of:
              0.08618971 = score(doc=1179,freq=2.0), product of:
                0.15912095 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045439374 = queryNorm
                0.5416616 = fieldWeight in 1179, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1179)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    31. 7.1996 9:22:19
  12. Baayen, R.H.; Lieber, H.: Word frequency distributions and lexical semantics (1997) 0.02
    0.021547427 = product of:
      0.043094855 = sum of:
        0.043094855 = product of:
          0.08618971 = sum of:
            0.08618971 = weight(_text_:22 in 3117) [ClassicSimilarity], result of:
              0.08618971 = score(doc=3117,freq=2.0), product of:
                0.15912095 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045439374 = queryNorm
                0.5416616 = fieldWeight in 3117, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3117)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    28. 2.1999 10:48:22
  13. Alonge, A.; Calzolari, N.; Vossen, P.; Bloksma, L.; Castellon, I.; Marti, M.A.; Peters, W.: ¬The linguistic design of the EuroWordNet database (1998) 0.02
    0.021425933 = product of:
      0.042851865 = sum of:
        0.042851865 = product of:
          0.08570373 = sum of:
            0.08570373 = weight(_text_:i in 6440) [ClassicSimilarity], result of:
              0.08570373 = score(doc=6440,freq=2.0), product of:
                0.17138503 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.045439374 = queryNorm
                0.50006545 = fieldWeight in 6440, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6440)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  14. Vossen, P.; Bloksma, L.; Alonge, A.; Marinai, E.; Peters, C.; Catellon, I.; Marti, M.A.; Rigau, G.: Compatibility in interpretation of relations in EuroWordNet (1998) 0.02
    0.021425933 = product of:
      0.042851865 = sum of:
        0.042851865 = product of:
          0.08570373 = sum of:
            0.08570373 = weight(_text_:i in 6442) [ClassicSimilarity], result of:
              0.08570373 = score(doc=6442,freq=2.0), product of:
                0.17138503 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.045439374 = queryNorm
                0.50006545 = fieldWeight in 6442, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6442)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  15. Godby, C.J.: ¬Two Techniques for the Identification of Phrases in Full Text (2001) 0.02
    0.021425933 = product of:
      0.042851865 = sum of:
        0.042851865 = product of:
          0.08570373 = sum of:
            0.08570373 = weight(_text_:i in 1000) [ClassicSimilarity], result of:
              0.08570373 = score(doc=1000,freq=2.0), product of:
                0.17138503 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.045439374 = queryNorm
                0.50006545 = fieldWeight in 1000, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1000)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Footnote
    Teil eines Themenheftes: OCLC and the Internet: An Historical Overview of Research Activities, 1990-1999 - Part I
  16. Manning, C.D.: Part-of-Speech Tagging from 97% to 100% : is it time for some linguistics? (2011) 0.02
    0.019962436 = product of:
      0.03992487 = sum of:
        0.03992487 = product of:
          0.07984974 = sum of:
            0.07984974 = weight(_text_:i in 1121) [ClassicSimilarity], result of:
              0.07984974 = score(doc=1121,freq=10.0), product of:
                0.17138503 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.045439374 = queryNorm
                0.46590847 = fieldWeight in 1121, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1121)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    I examine what would be necessary to move part-of-speech tagging performance from its current level of about 97.3% token accuracy (56% sentence accuracy) to close to 100% accuracy. I suggest that it must still be possible to greatly increase tagging performance and examine some useful improvements that have recently been made to the Stanford Part-of-Speech Tagger. However, an error analysis of some of the remaining errors suggests that there is limited further mileage to be had either from better machine learning or better features in a discriminative sequence classifier. The prospects for further gains from semisupervised learning also seem quite limited. Rather, I suggest and begin to demonstrate that the largest opportunity for further progress comes from improving the taxonomic basis of the linguistic resources from which taggers are trained. That is, from improved descriptive linguistics. However, I conclude by suggesting that there are also limits to this process. The status of some words may not be able to be adequately captured by assigning them to one of a small number of categories. While conventions can be used in such cases to improve tagging consistency, they lack a strong linguistic basis.
    Source
    Computational Linguistics and Intelligent Text Processing, 12th International Conference, CICLing 2011, Proceedings, Part I. Ed.: Alexander Gelbukh
  17. Byrne, C.C.; McCracken, S.A.: ¬An adaptive thesaurus employing semantic distance, relational inheritance and nominal compound interpretation for linguistic support of information retrieval (1999) 0.02
    0.018469224 = product of:
      0.036938448 = sum of:
        0.036938448 = product of:
          0.073876895 = sum of:
            0.073876895 = weight(_text_:22 in 4483) [ClassicSimilarity], result of:
              0.073876895 = score(doc=4483,freq=2.0), product of:
                0.15912095 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045439374 = queryNorm
                0.46428138 = fieldWeight in 4483, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4483)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    15. 3.2000 10:22:37
  18. Boleda, G.; Evert, S.: Multiword expressions : a pain in the neck of lexical semantics (2009) 0.02
    0.018469224 = product of:
      0.036938448 = sum of:
        0.036938448 = product of:
          0.073876895 = sum of:
            0.073876895 = weight(_text_:22 in 4888) [ClassicSimilarity], result of:
              0.073876895 = score(doc=4888,freq=2.0), product of:
                0.15912095 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045439374 = queryNorm
                0.46428138 = fieldWeight in 4888, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4888)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    1. 3.2013 14:56:22
  19. Hutchins, J.: From first conception to first demonstration : the nascent years of machine translation, 1947-1954. A chronology (1997) 0.02
    0.01539102 = product of:
      0.03078204 = sum of:
        0.03078204 = product of:
          0.06156408 = sum of:
            0.06156408 = weight(_text_:22 in 1463) [ClassicSimilarity], result of:
              0.06156408 = score(doc=1463,freq=2.0), product of:
                0.15912095 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045439374 = queryNorm
                0.38690117 = fieldWeight in 1463, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1463)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    31. 7.1996 9:22:19
  20. He, Q.: ¬A study of the strength indexes in co-word analysis (2000) 0.02
    0.015150423 = product of:
      0.030300846 = sum of:
        0.030300846 = product of:
          0.060601693 = sum of:
            0.060601693 = weight(_text_:i in 111) [ClassicSimilarity], result of:
              0.060601693 = score(doc=111,freq=4.0), product of:
                0.17138503 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.045439374 = queryNorm
                0.35359967 = fieldWeight in 111, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.046875 = fieldNorm(doc=111)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Co-word analysis is a technique for detecting the knowledge structure of scientific literature and mapping the dynamics in a research field. It is used to count the co-occurrences of term pairs, compute the strength between term pairs, and map the research field by inserting terms and their linkages into a graphical structure according to the strength values. In previous co-word studies, there are two indexes used to measure the strength between term pairs in order to identify the major areas in a research field - the inclusion index (I) and the equivalence index (E). This study will conduct two co-word analysis experiments using the two indexes, respectively, and compare the results from the two experiments. The results show, due to the difference in their computation, index I is more likely to identify general subject areas in a research field while index E is more likely to identify subject areas at more specific levels

Years

Types

  • a 56
  • el 8
  • s 4
  • m 3
  • p 3
  • x 2
  • More… Less…

Classifications