Search (73 results, page 3 of 4)

  • × theme_ss:"Computerlinguistik"
  • × type_ss:"a"
  1. Godby, J.: WordSmith research project bridges gap between tokens and indexes (1998) 0.01
    0.010792375 = product of:
      0.02158475 = sum of:
        0.02158475 = product of:
          0.0431695 = sum of:
            0.0431695 = weight(_text_:22 in 4729) [ClassicSimilarity], result of:
              0.0431695 = score(doc=4729,freq=2.0), product of:
                0.15939656 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045518078 = queryNorm
                0.2708308 = fieldWeight in 4729, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4729)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    OCLC newsletter. 1998, no.234, Jul/Aug, S.22-24
  2. Hammwöhner, R.: TransRouter revisited : Decision support in the routing of translation projects (2000) 0.01
    0.010792375 = product of:
      0.02158475 = sum of:
        0.02158475 = product of:
          0.0431695 = sum of:
            0.0431695 = weight(_text_:22 in 5483) [ClassicSimilarity], result of:
              0.0431695 = score(doc=5483,freq=2.0), product of:
                0.15939656 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045518078 = queryNorm
                0.2708308 = fieldWeight in 5483, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5483)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    10.12.2000 18:22:35
  3. Melby, A.: Some notes on 'The proper place of men and machines in language translation' (1997) 0.01
    0.010792375 = product of:
      0.02158475 = sum of:
        0.02158475 = product of:
          0.0431695 = sum of:
            0.0431695 = weight(_text_:22 in 330) [ClassicSimilarity], result of:
              0.0431695 = score(doc=330,freq=2.0), product of:
                0.15939656 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045518078 = queryNorm
                0.2708308 = fieldWeight in 330, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=330)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    31. 7.1996 9:22:19
  4. Schneider, J.W.; Borlund, P.: ¬A bibliometric-based semiautomatic approach to identification of candidate thesaurus terms : parsing and filtering of noun phrases from citation contexts (2005) 0.01
    0.010792375 = product of:
      0.02158475 = sum of:
        0.02158475 = product of:
          0.0431695 = sum of:
            0.0431695 = weight(_text_:22 in 156) [ClassicSimilarity], result of:
              0.0431695 = score(doc=156,freq=2.0), product of:
                0.15939656 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045518078 = queryNorm
                0.2708308 = fieldWeight in 156, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=156)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    8. 3.2007 19:55:22
  5. Paolillo, J.C.: Linguistics and the information sciences (2009) 0.01
    0.010792375 = product of:
      0.02158475 = sum of:
        0.02158475 = product of:
          0.0431695 = sum of:
            0.0431695 = weight(_text_:22 in 3840) [ClassicSimilarity], result of:
              0.0431695 = score(doc=3840,freq=2.0), product of:
                0.15939656 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045518078 = queryNorm
                0.2708308 = fieldWeight in 3840, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3840)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    27. 8.2011 14:22:33
  6. Schneider, R.: Web 3.0 ante portas? : Integration von Social Web und Semantic Web (2008) 0.01
    0.010792375 = product of:
      0.02158475 = sum of:
        0.02158475 = product of:
          0.0431695 = sum of:
            0.0431695 = weight(_text_:22 in 4184) [ClassicSimilarity], result of:
              0.0431695 = score(doc=4184,freq=2.0), product of:
                0.15939656 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045518078 = queryNorm
                0.2708308 = fieldWeight in 4184, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4184)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 1.2011 10:38:28
  7. Betrand-Gastaldy, S.: ¬La modelisation de l'analyse documentaire : à la convergence de la semiotique, de la psychologie cognitive et de l'intelligence (1995) 0.01
    0.009469198 = product of:
      0.018938396 = sum of:
        0.018938396 = product of:
          0.037876792 = sum of:
            0.037876792 = weight(_text_:b in 5377) [ClassicSimilarity], result of:
              0.037876792 = score(doc=5377,freq=2.0), product of:
                0.16126883 = queryWeight, product of:
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23486741 = fieldWeight in 5377, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5377)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Connectedness: information, systems, people, organizations. Proceedings of CAIS/ACSI 95, the proceedings of the 23rd Annual Conference of the Canadian Association for Information Science. Ed. by Hope A. Olson and Denis B. Ward
  8. Nait-Baha, L.; Jackiewicz, A.; Djioua, B.; Laublet, P.: Query reformulation for information retrieval on the Web using the point of view methodology : preliminary results (2001) 0.01
    0.009469198 = product of:
      0.018938396 = sum of:
        0.018938396 = product of:
          0.037876792 = sum of:
            0.037876792 = weight(_text_:b in 249) [ClassicSimilarity], result of:
              0.037876792 = score(doc=249,freq=2.0), product of:
                0.16126883 = queryWeight, product of:
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23486741 = fieldWeight in 249, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.046875 = fieldNorm(doc=249)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  9. Ferret, O.; Grau, B.; Hurault-Plantet, M.; Illouz, G.; Jacquemin, C.; Monceaux, L.; Robba, I.; Vilnat, A.: How NLP can improve question answering (2002) 0.01
    0.009469198 = product of:
      0.018938396 = sum of:
        0.018938396 = product of:
          0.037876792 = sum of:
            0.037876792 = weight(_text_:b in 1850) [ClassicSimilarity], result of:
              0.037876792 = score(doc=1850,freq=2.0), product of:
                0.16126883 = queryWeight, product of:
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23486741 = fieldWeight in 1850, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1850)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  10. Al-Shawakfa, E.; Al-Badarneh, A.; Shatnawi, S.; Al-Rabab'ah, K.; Bani-Ismail, B.: ¬A comparison study of some Arabic root finding algorithms (2010) 0.01
    0.009469198 = product of:
      0.018938396 = sum of:
        0.018938396 = product of:
          0.037876792 = sum of:
            0.037876792 = weight(_text_:b in 3457) [ClassicSimilarity], result of:
              0.037876792 = score(doc=3457,freq=2.0), product of:
                0.16126883 = queryWeight, product of:
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23486741 = fieldWeight in 3457, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3457)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  11. Perovsek, M.; Kranjca, J.; Erjaveca, T.; Cestnika, B.; Lavraca, N.: TextFlows : a visual programming platform for text mining and natural language processing (2016) 0.01
    0.009469198 = product of:
      0.018938396 = sum of:
        0.018938396 = product of:
          0.037876792 = sum of:
            0.037876792 = weight(_text_:b in 2697) [ClassicSimilarity], result of:
              0.037876792 = score(doc=2697,freq=2.0), product of:
                0.16126883 = queryWeight, product of:
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23486741 = fieldWeight in 2697, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2697)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  12. Ghazzawi, N.; Robichaud, B.; Drouin, P.; Sadat, F.: Automatic extraction of specialized verbal units (2018) 0.01
    0.009469198 = product of:
      0.018938396 = sum of:
        0.018938396 = product of:
          0.037876792 = sum of:
            0.037876792 = weight(_text_:b in 4094) [ClassicSimilarity], result of:
              0.037876792 = score(doc=4094,freq=2.0), product of:
                0.16126883 = queryWeight, product of:
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23486741 = fieldWeight in 4094, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4094)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  13. Lu, C.; Bu, Y.; Wang, J.; Ding, Y.; Torvik, V.; Schnaars, M.; Zhang, C.: Examining scientific writing styles from the perspective of linguistic complexity : a cross-level moderation model (2019) 0.01
    0.009469198 = product of:
      0.018938396 = sum of:
        0.018938396 = product of:
          0.037876792 = sum of:
            0.037876792 = weight(_text_:b in 5219) [ClassicSimilarity], result of:
              0.037876792 = score(doc=5219,freq=2.0), product of:
                0.16126883 = queryWeight, product of:
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23486741 = fieldWeight in 5219, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5219)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Publishing articles in high-impact English journals is difficult for scholars around the world, especially for non-native English-speaking scholars (NNESs), most of whom struggle with proficiency in English. To uncover the differences in English scientific writing between native English-speaking scholars (NESs) and NNESs, we collected a large-scale data set containing more than 150,000 full-text articles published in PLoS between 2006 and 2015. We divided these articles into three groups according to the ethnic backgrounds of the first and corresponding authors, obtained by Ethnea, and examined the scientific writing styles in English from a two-fold perspective of linguistic complexity: (a) syntactic complexity, including measurements of sentence length and sentence complexity; and (b) lexical complexity, including measurements of lexical diversity, lexical density, and lexical sophistication. The observations suggest marginal differences between groups in syntactical and lexical complexity.
  14. Lund, B.D.; Wang, T.; Mannuru, N.R.; Nie, B.; Shimray, S.; Wang, Z.: ChatGPT and a new academic reality : artificial Intelligence-written research papers and the ethics of the large language models in scholarly publishing (2023) 0.01
    0.009469198 = product of:
      0.018938396 = sum of:
        0.018938396 = product of:
          0.037876792 = sum of:
            0.037876792 = weight(_text_:b in 943) [ClassicSimilarity], result of:
              0.037876792 = score(doc=943,freq=2.0), product of:
                0.16126883 = queryWeight, product of:
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23486741 = fieldWeight in 943, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.046875 = fieldNorm(doc=943)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  15. Dorr, B.J.: Large-scale dictionary construction for foreign language tutoring and interlingual machine translation (1997) 0.01
    0.009250606 = product of:
      0.018501213 = sum of:
        0.018501213 = product of:
          0.037002426 = sum of:
            0.037002426 = weight(_text_:22 in 3244) [ClassicSimilarity], result of:
              0.037002426 = score(doc=3244,freq=2.0), product of:
                0.15939656 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23214069 = fieldWeight in 3244, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3244)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    31. 7.1996 9:22:19
  16. Bian, G.-W.; Chen, H.-H.: Cross-language information access to multilingual collections on the Internet (2000) 0.01
    0.009250606 = product of:
      0.018501213 = sum of:
        0.018501213 = product of:
          0.037002426 = sum of:
            0.037002426 = weight(_text_:22 in 4436) [ClassicSimilarity], result of:
              0.037002426 = score(doc=4436,freq=2.0), product of:
                0.15939656 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23214069 = fieldWeight in 4436, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4436)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    16. 2.2000 14:22:39
  17. Rahmstorf, G.: Concept structures for large vocabularies (1998) 0.01
    0.009250606 = product of:
      0.018501213 = sum of:
        0.018501213 = product of:
          0.037002426 = sum of:
            0.037002426 = weight(_text_:22 in 75) [ClassicSimilarity], result of:
              0.037002426 = score(doc=75,freq=2.0), product of:
                0.15939656 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23214069 = fieldWeight in 75, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=75)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    30.12.2001 19:01:22
  18. Lawrie, D.; Mayfield, J.; McNamee, P.; Oard, P.W.: Cross-language person-entity linking from 20 languages (2015) 0.01
    0.009250606 = product of:
      0.018501213 = sum of:
        0.018501213 = product of:
          0.037002426 = sum of:
            0.037002426 = weight(_text_:22 in 1848) [ClassicSimilarity], result of:
              0.037002426 = score(doc=1848,freq=2.0), product of:
                0.15939656 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045518078 = queryNorm
                0.23214069 = fieldWeight in 1848, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1848)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The goal of entity linking is to associate references to an entity that is found in unstructured natural language content to an authoritative inventory of known entities. This article describes the construction of 6 test collections for cross-language person-entity linking that together span 22 languages. Fully automated components were used together with 2 crowdsourced validation stages to affordably generate ground-truth annotations with an accuracy comparable to that of a completely manual process. The resulting test collections each contain between 642 (Arabic) and 2,361 (Romanian) person references in non-English texts for which the correct resolution in English Wikipedia is known, plus a similar number of references for which no correct resolution into English Wikipedia is believed to exist. Fully automated cross-language person-name linking experiments with 20 non-English languages yielded a resolution accuracy of between 0.84 (Serbian) and 0.98 (Romanian), which compares favorably with previously reported cross-language entity linking results for Spanish.
  19. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; Agarwal, S.; Herbert-Voss, A.; Krueger, G.; Henighan, T.; Child, R.; Ramesh, A.; Ziegler, D.M.; Wu, J.; Winter, C.; Hesse, C.; Chen, M.; Sigler, E.; Litwin, M.; Gray, S.; Chess, B.; Clark, J.; Berner, C.; McCandlish, S.; Radford, A.; Sutskever, I.; Amodei, D.: Language models are few-shot learners (2020) 0.01
    0.008927646 = product of:
      0.017855292 = sum of:
        0.017855292 = product of:
          0.035710584 = sum of:
            0.035710584 = weight(_text_:b in 872) [ClassicSimilarity], result of:
              0.035710584 = score(doc=872,freq=4.0), product of:
                0.16126883 = queryWeight, product of:
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.045518078 = queryNorm
                0.22143513 = fieldWeight in 872, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.03125 = fieldNorm(doc=872)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  20. Shaalan, K.; Raza, H.: NERA: Named Entity Recognition for Arabic (2009) 0.01
    0.007890998 = product of:
      0.015781997 = sum of:
        0.015781997 = product of:
          0.031563994 = sum of:
            0.031563994 = weight(_text_:b in 2953) [ClassicSimilarity], result of:
              0.031563994 = score(doc=2953,freq=2.0), product of:
                0.16126883 = queryWeight, product of:
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.045518078 = queryNorm
                0.19572285 = fieldWeight in 2953, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.542962 = idf(docFreq=3476, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2953)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Name identification has been worked on quite intensively for the past few years, and has been incorporated into several products revolving around natural language processing tasks. Many researchers have attacked the name identification problem in a variety of languages, but only a few limited research efforts have focused on named entity recognition for Arabic script. This is due to the lack of resources for Arabic named entities and the limited amount of progress made in Arabic natural language processing in general. In this article, we present the results of our attempt at the recognition and extraction of the 10 most important categories of named entities in Arabic script: the person name, location, company, date, time, price, measurement, phone number, ISBN, and file name. We developed the system Named Entity Recognition for Arabic (NERA) using a rule-based approach. The resources created are: a Whitelist representing a dictionary of names, and a grammar, in the form of regular expressions, which are responsible for recognizing the named entities. A filtration mechanism is used that serves two different purposes: (a) revision of the results from a named entity extractor by using metadata, in terms of a Blacklist or rejecter, about ill-formed named entities and (b) disambiguation of identical or overlapping textual matches returned by different name entity extractors to get the correct choice. In NERA, we addressed major challenges posed by NER in the Arabic language arising due to the complexity of the language, peculiarities in the Arabic orthographic system, nonstandardization of the written text, ambiguity, and lack of resources. NERA has been effectively evaluated using our own tagged corpus; it achieved satisfactory results in terms of precision, recall, and F-measure.

Years

Languages

  • e 53
  • d 18
  • f 2
  • More… Less…

Types