Search (711 results, page 3 of 36)

  • × theme_ss:"Computerlinguistik"
  1. Sienel, J.; Weiss, M.; Laube, M.: Sprachtechnologien für die Informationsgesellschaft des 21. Jahrhunderts (2000) 0.01
    0.014156656 = product of:
      0.028313313 = sum of:
        0.028313313 = sum of:
          0.0027694327 = weight(_text_:a in 5557) [ClassicSimilarity], result of:
            0.0027694327 = score(doc=5557,freq=2.0), product of:
              0.043477926 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.037706986 = queryNorm
              0.06369744 = fieldWeight in 5557, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5557)
          0.02554388 = weight(_text_:22 in 5557) [ClassicSimilarity], result of:
            0.02554388 = score(doc=5557,freq=2.0), product of:
              0.13204344 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.037706986 = queryNorm
              0.19345059 = fieldWeight in 5557, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5557)
      0.5 = coord(1/2)
    
    Date
    26.12.2000 13:22:17
    Type
    a
  2. Pinker, S.: Wörter und Regeln : Die Natur der Sprache (2000) 0.01
    0.014156656 = product of:
      0.028313313 = sum of:
        0.028313313 = sum of:
          0.0027694327 = weight(_text_:a in 734) [ClassicSimilarity], result of:
            0.0027694327 = score(doc=734,freq=2.0), product of:
              0.043477926 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.037706986 = queryNorm
              0.06369744 = fieldWeight in 734, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=734)
          0.02554388 = weight(_text_:22 in 734) [ClassicSimilarity], result of:
            0.02554388 = score(doc=734,freq=2.0), product of:
              0.13204344 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.037706986 = queryNorm
              0.19345059 = fieldWeight in 734, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=734)
      0.5 = coord(1/2)
    
    Date
    19. 7.2002 14:22:31
    Footnote
    Rez. in: Franfurter Rundschau Nr.43 vom 20.2.2001, S.23 (A. Barthelmy)
  3. Yang, C.C.; Luk, J.: Automatic generation of English/Chinese thesaurus based on a parallel corpus in laws (2003) 0.01
    0.012298116 = product of:
      0.024596233 = sum of:
        0.024596233 = sum of:
          0.006715518 = weight(_text_:a in 1616) [ClassicSimilarity], result of:
            0.006715518 = score(doc=1616,freq=24.0), product of:
              0.043477926 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.037706986 = queryNorm
              0.1544581 = fieldWeight in 1616, product of:
                4.8989797 = tf(freq=24.0), with freq of:
                  24.0 = termFreq=24.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.02734375 = fieldNorm(doc=1616)
          0.017880715 = weight(_text_:22 in 1616) [ClassicSimilarity], result of:
            0.017880715 = score(doc=1616,freq=2.0), product of:
              0.13204344 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.037706986 = queryNorm
              0.1354154 = fieldWeight in 1616, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.02734375 = fieldNorm(doc=1616)
      0.5 = coord(1/2)
    
    Abstract
    The information available in languages other than English in the World Wide Web is increasing significantly. According to a report from Computer Economics in 1999, 54% of Internet users are English speakers ("English Will Dominate Web for Only Three More Years," Computer Economics, July 9, 1999, http://www.computereconomics. com/new4/pr/pr990610.html). However, it is predicted that there will be only 60% increase in Internet users among English speakers verses a 150% growth among nonEnglish speakers for the next five years. By 2005, 57% of Internet users will be non-English speakers. A report by CNN.com in 2000 showed that the number of Internet users in China had been increased from 8.9 million to 16.9 million from January to June in 2000 ("Report: China Internet users double to 17 million," CNN.com, July, 2000, http://cnn.org/2000/TECH/computing/07/27/ china.internet.reut/index.html). According to Nielsen/ NetRatings, there was a dramatic leap from 22.5 millions to 56.6 millions Internet users from 2001 to 2002. China had become the second largest global at-home Internet population in 2002 (US's Internet population was 166 millions) (Robyn Greenspan, "China Pulls Ahead of Japan," Internet.com, April 22, 2002, http://cyberatias.internet.com/big-picture/geographics/article/0,,5911_1013841,00. html). All of the evidences reveal the importance of crosslingual research to satisfy the needs in the near future. Digital library research has been focusing in structural and semantic interoperability in the past. Searching and retrieving objects across variations in protocols, formats and disciplines are widely explored (Schatz, B., & Chen, H. (1999). Digital libraries: technological advances and social impacts. IEEE Computer, Special Issue an Digital Libraries, February, 32(2), 45-50.; Chen, H., Yen, J., & Yang, C.C. (1999). International activities: development of Asian digital libraries. IEEE Computer, Special Issue an Digital Libraries, 32(2), 48-49.). However, research in crossing language boundaries, especially across European languages and Oriental languages, is still in the initial stage. In this proposal, we put our focus an cross-lingual semantic interoperability by developing automatic generation of a cross-lingual thesaurus based an English/Chinese parallel corpus. When the searchers encounter retrieval problems, Professional librarians usually consult the thesaurus to identify other relevant vocabularies. In the problem of searching across language boundaries, a cross-lingual thesaurus, which is generated by co-occurrence analysis and Hopfield network, can be used to generate additional semantically relevant terms that cannot be obtained from dictionary. In particular, the automatically generated cross-lingual thesaurus is able to capture the unknown words that do not exist in a dictionary, such as names of persons, organizations, and events. Due to Hong Kong's unique history background, both English and Chinese are used as official languages in all legal documents. Therefore, English/Chinese cross-lingual information retrieval is critical for applications in courts and the government. In this paper, we develop an automatic thesaurus by the Hopfield network based an a parallel corpus collected from the Web site of the Department of Justice of the Hong Kong Special Administrative Region (HKSAR) Government. Experiments are conducted to measure the precision and recall of the automatic generated English/Chinese thesaurus. The result Shows that such thesaurus is a promising tool to retrieve relevant terms, especially in the language that is not the same as the input term. The direct translation of the input term can also be retrieved in most of the cases.
    Footnote
    Teil eines Themenheftes: "Web retrieval and mining: A machine learning perspective"
    Type
    a
  4. Deventer, J.P. van; Kruger, C.J.; Johnson, R.D.: Delineating knowledge management through lexical analysis : a retrospective (2015) 0.01
    0.011504889 = product of:
      0.023009777 = sum of:
        0.023009777 = sum of:
          0.0051290616 = weight(_text_:a in 3807) [ClassicSimilarity], result of:
            0.0051290616 = score(doc=3807,freq=14.0), product of:
              0.043477926 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.037706986 = queryNorm
              0.11796933 = fieldWeight in 3807, product of:
                3.7416575 = tf(freq=14.0), with freq of:
                  14.0 = termFreq=14.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.02734375 = fieldNorm(doc=3807)
          0.017880715 = weight(_text_:22 in 3807) [ClassicSimilarity], result of:
            0.017880715 = score(doc=3807,freq=2.0), product of:
              0.13204344 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.037706986 = queryNorm
              0.1354154 = fieldWeight in 3807, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.02734375 = fieldNorm(doc=3807)
      0.5 = coord(1/2)
    
    Abstract
    Purpose Academic authors tend to define terms that meet their own needs. Knowledge Management (KM) is a term that comes to mind and is examined in this study. Lexicographical research identified KM terms used by authors from 1996 to 2006 in academic outlets to define KM. Data were collected based on strict criteria which included that definitions should be unique instances. From 2006 onwards, these authors could not identify new unique instances of definitions with repetitive usage of such definition instances. Analysis revealed that KM is directly defined by People (Person and Organisation), Processes (Codify, Share, Leverage, and Process) and Contextualised Content (Information). The paper aims to discuss these issues. Design/methodology/approach The aim of this paper is to add to the body of knowledge in the KM discipline and supply KM practitioners and scholars with insight into what is commonly regarded to be KM so as to reignite the debate on what one could consider as KM. The lexicon used by KM scholars was evaluated though the application of lexicographical research methods as extended though Knowledge Discovery and Text Analysis methods. Findings By simplifying term relationships through the application of lexicographical research methods, as extended though Knowledge Discovery and Text Analysis methods, it was found that KM is directly defined by People (Person and Organisation), Processes (Codify, Share, Leverage, Process) and Contextualised Content (Information). One would therefore be able to indicate that KM, from an academic point of view, refers to people processing contextualised content.
    Research limitations/implications In total, 42 definitions were identified spanning a period of 11 years. This represented the first use of KM through the estimated apex of terms used. From 2006 onwards definitions were used in repetition, and all definitions that were considered to repeat were therefore subsequently excluded as not being unique instances. All definitions listed are by no means complete and exhaustive. The definitions are viewed outside the scope and context in which they were originally formulated and then used to review the key concepts in the definitions themselves. Social implications When the authors refer to the aforementioned discussion of KM content as well as the presentation of the method followed in this paper, the authors may have a few implications for future research in KM. First the research validates ideas presented by the OECD in 2005 pertaining to KM. It also validates that through the evolution of KM, the authors ended with a description of KM that may be seen as a standardised description. If the authors as academics and practitioners, for example, refer to KM as the same construct and/or idea, it has the potential to speculatively, distinguish between what KM may or may not be. Originality/value By simplifying the term used to define KM, by focusing on the most common definitions, the paper assist in refocusing KM by reconsidering the dimensions that is the most common in how it has been defined over time. This would hopefully assist in reigniting discussions about KM and how it may be used to the benefit of an organisation.
    Date
    20. 1.2015 18:30:22
    Isbn
    a
  5. Rötzer, F.: KI-Programm besser als Menschen im Verständnis natürlicher Sprache (2018) 0.01
    0.011325325 = product of:
      0.02265065 = sum of:
        0.02265065 = sum of:
          0.0022155463 = weight(_text_:a in 4217) [ClassicSimilarity], result of:
            0.0022155463 = score(doc=4217,freq=2.0), product of:
              0.043477926 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.037706986 = queryNorm
              0.050957955 = fieldWeight in 4217, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.03125 = fieldNorm(doc=4217)
          0.020435104 = weight(_text_:22 in 4217) [ClassicSimilarity], result of:
            0.020435104 = score(doc=4217,freq=2.0), product of:
              0.13204344 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.037706986 = queryNorm
              0.15476047 = fieldWeight in 4217, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=4217)
      0.5 = coord(1/2)
    
    Date
    22. 1.2018 11:32:44
    Type
    a
  6. Way, E.C.: Knowledge representation and metaphor (oder: meaning) (1994) 0.01
    0.010217552 = product of:
      0.020435104 = sum of:
        0.020435104 = product of:
          0.04087021 = sum of:
            0.04087021 = weight(_text_:22 in 771) [ClassicSimilarity], result of:
              0.04087021 = score(doc=771,freq=2.0), product of:
                0.13204344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037706986 = queryNorm
                0.30952093 = fieldWeight in 771, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=771)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Footnote
    Bereits 1991 bei Kluwer publiziert // Rez. in: Knowledge organization 22(1995) no.1, S.48-49 (O. Sechser)
  7. Ruge, G.: Sprache und Computer : Wortbedeutung und Termassoziation. Methoden zur automatischen semantischen Klassifikation (1995) 0.01
    0.010217552 = product of:
      0.020435104 = sum of:
        0.020435104 = product of:
          0.04087021 = sum of:
            0.04087021 = weight(_text_:22 in 1534) [ClassicSimilarity], result of:
              0.04087021 = score(doc=1534,freq=2.0), product of:
                0.13204344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037706986 = queryNorm
                0.30952093 = fieldWeight in 1534, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1534)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Footnote
    Rez. in: Knowledge organization 22(1995) no.3/4, S.182-184 (M.T. Rolland)
  8. Lezius, W.: Morphy - Morphologie und Tagging für das Deutsche (2013) 0.01
    0.010217552 = product of:
      0.020435104 = sum of:
        0.020435104 = product of:
          0.04087021 = sum of:
            0.04087021 = weight(_text_:22 in 1490) [ClassicSimilarity], result of:
              0.04087021 = score(doc=1490,freq=2.0), product of:
                0.13204344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037706986 = queryNorm
                0.30952093 = fieldWeight in 1490, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1490)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 3.2015 9:30:24
  9. Schürmann, H.: Software scannt Radio- und Fernsehsendungen : Recherche in Nachrichtenarchiven erleichtert (2001) 0.01
    0.00990966 = product of:
      0.01981932 = sum of:
        0.01981932 = sum of:
          0.001938603 = weight(_text_:a in 5759) [ClassicSimilarity], result of:
            0.001938603 = score(doc=5759,freq=2.0), product of:
              0.043477926 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.037706986 = queryNorm
              0.044588212 = fieldWeight in 5759, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.02734375 = fieldNorm(doc=5759)
          0.017880715 = weight(_text_:22 in 5759) [ClassicSimilarity], result of:
            0.017880715 = score(doc=5759,freq=2.0), product of:
              0.13204344 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.037706986 = queryNorm
              0.1354154 = fieldWeight in 5759, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.02734375 = fieldNorm(doc=5759)
      0.5 = coord(1/2)
    
    Source
    Handelsblatt. Nr.79 vom 24.4.2001, S.22
    Type
    a
  10. Melzer, C.: ¬Der Maschine anpassen : PC-Spracherkennung - Programme sind mittlerweile alltagsreif (2005) 0.01
    0.00990966 = product of:
      0.01981932 = sum of:
        0.01981932 = sum of:
          0.001938603 = weight(_text_:a in 4044) [ClassicSimilarity], result of:
            0.001938603 = score(doc=4044,freq=2.0), product of:
              0.043477926 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.037706986 = queryNorm
              0.044588212 = fieldWeight in 4044, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.02734375 = fieldNorm(doc=4044)
          0.017880715 = weight(_text_:22 in 4044) [ClassicSimilarity], result of:
            0.017880715 = score(doc=4044,freq=2.0), product of:
              0.13204344 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.037706986 = queryNorm
              0.1354154 = fieldWeight in 4044, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.02734375 = fieldNorm(doc=4044)
      0.5 = coord(1/2)
    
    Date
    3. 5.1997 8:44:22
    Type
    a
  11. Lorenz, S.: Konzeption und prototypische Realisierung einer begriffsbasierten Texterschließung (2006) 0.01
    0.007663164 = product of:
      0.015326328 = sum of:
        0.015326328 = product of:
          0.030652655 = sum of:
            0.030652655 = weight(_text_:22 in 1746) [ClassicSimilarity], result of:
              0.030652655 = score(doc=1746,freq=2.0), product of:
                0.13204344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037706986 = queryNorm
                0.23214069 = fieldWeight in 1746, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1746)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 3.2015 9:17:30
  12. Semantik, Lexikographie und Computeranwendungen : Workshop ... (Bonn) : 1995.01.27-28 (1996) 0.01
    0.00638597 = product of:
      0.01277194 = sum of:
        0.01277194 = product of:
          0.02554388 = sum of:
            0.02554388 = weight(_text_:22 in 190) [ClassicSimilarity], result of:
              0.02554388 = score(doc=190,freq=2.0), product of:
                0.13204344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037706986 = queryNorm
                0.19345059 = fieldWeight in 190, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=190)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    14. 4.2007 10:04:22
  13. MacLeod, C.; Grisham, R.; Meyer, A.: COMLERX syntax : a large syntactic dictionary for natural language processing (1998) 0.00
    0.003357759 = product of:
      0.006715518 = sum of:
        0.006715518 = product of:
          0.013431036 = sum of:
            0.013431036 = weight(_text_:a in 3167) [ClassicSimilarity], result of:
              0.013431036 = score(doc=3167,freq=6.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.3089162 = fieldWeight in 3167, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3167)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  14. Sembok, T.M.T.; Rijsbergen, C.J. van: SILOL: a simple logical-linguistic document retrieval system (1990) 0.00
    0.0033233196 = product of:
      0.006646639 = sum of:
        0.006646639 = product of:
          0.013293278 = sum of:
            0.013293278 = weight(_text_:a in 6684) [ClassicSimilarity], result of:
              0.013293278 = score(doc=6684,freq=18.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.30574775 = fieldWeight in 6684, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6684)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Describes a system called SILOL which is based on a logical-linguistic model of document retrieval systems. SILOL uses a shallow semantic translation of natural language texts into a first order predicate representation in performing a document indexing and retrieval process. Some preliminary experiments have been carried out to test the retrieval effectiveness of this system. The results obtained show improvements in the level of retrieval effectiveness, which demonstrate that the approach of using a semantic theory of natural language and logic in document retrieval systems is a valid one
    Type
    a
  15. Campe, P.: Case, semantic roles, and grammatical relations : a comprehensive bibliography (1994) 0.00
    0.0033233196 = product of:
      0.006646639 = sum of:
        0.006646639 = product of:
          0.013293278 = sum of:
            0.013293278 = weight(_text_:a in 8663) [ClassicSimilarity], result of:
              0.013293278 = score(doc=8663,freq=8.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.30574775 = fieldWeight in 8663, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=8663)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Contains references to more than 6000 publications with a subject and a language index as well as a guide to the relevant languages and language families
  16. Solvberg, I.; Nordbo, I.; Aamodt, A.: Knowledge-based information retrieval (1991/92) 0.00
    0.0031332558 = product of:
      0.0062665115 = sum of:
        0.0062665115 = product of:
          0.012533023 = sum of:
            0.012533023 = weight(_text_:a in 546) [ClassicSimilarity], result of:
              0.012533023 = score(doc=546,freq=4.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.28826174 = fieldWeight in 546, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=546)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  17. Drouin, P.: Term extraction using non-technical corpora as a point of leverage (2003) 0.00
    0.0031332558 = product of:
      0.0062665115 = sum of:
        0.0062665115 = product of:
          0.012533023 = sum of:
            0.012533023 = weight(_text_:a in 8797) [ClassicSimilarity], result of:
              0.012533023 = score(doc=8797,freq=4.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.28826174 = fieldWeight in 8797, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=8797)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  18. Stede, M.: Lexicalization in natural language generation : a survey (1994/95) 0.00
    0.0031332558 = product of:
      0.0062665115 = sum of:
        0.0062665115 = product of:
          0.012533023 = sum of:
            0.012533023 = weight(_text_:a in 1913) [ClassicSimilarity], result of:
              0.012533023 = score(doc=1913,freq=16.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.28826174 = fieldWeight in 1913, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1913)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In natural language generation, a meaning representation of some kind is successively transformed into a sentence or a text. Naturally, a central subtask of this problem is the choice of words, or lexicalization. Proposes 4 major issues that determine how a generator tackles lexicalization, and surveys the contributions that research have made to them. Identifies open problems, and sketches a possible direction for research
    Type
    a
  19. Barriere, C.: Building a concept hierarchy from corpus analysis (2004) 0.00
    0.0031332558 = product of:
      0.0062665115 = sum of:
        0.0062665115 = product of:
          0.012533023 = sum of:
            0.012533023 = weight(_text_:a in 6787) [ClassicSimilarity], result of:
              0.012533023 = score(doc=6787,freq=4.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.28826174 = fieldWeight in 6787, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=6787)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  20. Chung, T.M.: ¬A corpus comparison approach for terminology extraction (2003) 0.00
    0.0031332558 = product of:
      0.0062665115 = sum of:
        0.0062665115 = product of:
          0.012533023 = sum of:
            0.012533023 = weight(_text_:a in 4072) [ClassicSimilarity], result of:
              0.012533023 = score(doc=4072,freq=4.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.28826174 = fieldWeight in 4072, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=4072)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a

Languages

Types

  • a 629
  • el 75
  • m 42
  • s 22
  • x 9
  • p 7
  • b 1
  • d 1
  • pat 1
  • r 1
  • More… Less…

Subjects

Classifications