Search (255 results, page 1 of 13)

  • × theme_ss:"Computerlinguistik"
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.08
    0.084174104 = sum of:
      0.062759444 = product of:
        0.25103778 = sum of:
          0.25103778 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
            0.25103778 = score(doc=562,freq=2.0), product of:
              0.4466721 = queryWeight, product of:
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.052685954 = queryNorm
              0.56201804 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.25 = coord(1/4)
      0.021414658 = product of:
        0.042829316 = sum of:
          0.042829316 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
            0.042829316 = score(doc=562,freq=2.0), product of:
              0.18449724 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052685954 = queryNorm
              0.23214069 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.5 = coord(1/2)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  2. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.07
    0.07357325 = sum of:
      0.062759444 = product of:
        0.25103778 = sum of:
          0.25103778 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
            0.25103778 = score(doc=862,freq=2.0), product of:
              0.4466721 = queryWeight, product of:
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.052685954 = queryNorm
              0.56201804 = fieldWeight in 862, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.046875 = fieldNorm(doc=862)
        0.25 = coord(1/4)
      0.010813803 = product of:
        0.021627607 = sum of:
          0.021627607 = weight(_text_:m in 862) [ClassicSimilarity], result of:
            0.021627607 = score(doc=862,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.1649624 = fieldWeight in 862, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.046875 = fieldNorm(doc=862)
        0.5 = coord(1/2)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  3. Lezius, W.; Rapp, R.; Wettler, M.: ¬A morphology-system and part-of-speech tagger for German (1996) 0.05
    0.053714104 = product of:
      0.10742821 = sum of:
        0.10742821 = sum of:
          0.036046013 = weight(_text_:m in 1693) [ClassicSimilarity], result of:
            0.036046013 = score(doc=1693,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.27493733 = fieldWeight in 1693, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.078125 = fieldNorm(doc=1693)
          0.071382195 = weight(_text_:22 in 1693) [ClassicSimilarity], result of:
            0.071382195 = score(doc=1693,freq=2.0), product of:
              0.18449724 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052685954 = queryNorm
              0.38690117 = fieldWeight in 1693, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.078125 = fieldNorm(doc=1693)
      0.5 = coord(1/2)
    
    Date
    22. 3.2015 9:37:18
  4. Way, E.C.: Knowledge representation and metaphor (oder: meaning) (1994) 0.04
    0.042971283 = product of:
      0.08594257 = sum of:
        0.08594257 = sum of:
          0.02883681 = weight(_text_:m in 771) [ClassicSimilarity], result of:
            0.02883681 = score(doc=771,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.21994986 = fieldWeight in 771, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.0625 = fieldNorm(doc=771)
          0.057105757 = weight(_text_:22 in 771) [ClassicSimilarity], result of:
            0.057105757 = score(doc=771,freq=2.0), product of:
              0.18449724 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052685954 = queryNorm
              0.30952093 = fieldWeight in 771, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=771)
      0.5 = coord(1/2)
    
    Footnote
    Bereits 1991 bei Kluwer publiziert // Rez. in: Knowledge organization 22(1995) no.1, S.48-49 (O. Sechser)
    Type
    m
  5. Ruge, G.: Sprache und Computer : Wortbedeutung und Termassoziation. Methoden zur automatischen semantischen Klassifikation (1995) 0.04
    0.042971283 = product of:
      0.08594257 = sum of:
        0.08594257 = sum of:
          0.02883681 = weight(_text_:m in 1534) [ClassicSimilarity], result of:
            0.02883681 = score(doc=1534,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.21994986 = fieldWeight in 1534, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.0625 = fieldNorm(doc=1534)
          0.057105757 = weight(_text_:22 in 1534) [ClassicSimilarity], result of:
            0.057105757 = score(doc=1534,freq=2.0), product of:
              0.18449724 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052685954 = queryNorm
              0.30952093 = fieldWeight in 1534, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=1534)
      0.5 = coord(1/2)
    
    Footnote
    Rez. in: Knowledge organization 22(1995) no.3/4, S.182-184 (M.T. Rolland)
    Type
    m
  6. Deventer, J.P. van; Kruger, C.J.; Johnson, R.D.: Delineating knowledge management through lexical analysis : a retrospective (2015) 0.04
    0.040498503 = sum of:
      0.02800662 = product of:
        0.11202648 = sum of:
          0.11202648 = weight(_text_:authors in 3807) [ClassicSimilarity], result of:
            0.11202648 = score(doc=3807,freq=14.0), product of:
              0.24018547 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.052685954 = queryNorm
              0.46641657 = fieldWeight in 3807, product of:
                3.7416575 = tf(freq=14.0), with freq of:
                  14.0 = termFreq=14.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.02734375 = fieldNorm(doc=3807)
        0.25 = coord(1/4)
      0.012491884 = product of:
        0.024983767 = sum of:
          0.024983767 = weight(_text_:22 in 3807) [ClassicSimilarity], result of:
            0.024983767 = score(doc=3807,freq=2.0), product of:
              0.18449724 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052685954 = queryNorm
              0.1354154 = fieldWeight in 3807, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.02734375 = fieldNorm(doc=3807)
        0.5 = coord(1/2)
    
    Abstract
    Purpose Academic authors tend to define terms that meet their own needs. Knowledge Management (KM) is a term that comes to mind and is examined in this study. Lexicographical research identified KM terms used by authors from 1996 to 2006 in academic outlets to define KM. Data were collected based on strict criteria which included that definitions should be unique instances. From 2006 onwards, these authors could not identify new unique instances of definitions with repetitive usage of such definition instances. Analysis revealed that KM is directly defined by People (Person and Organisation), Processes (Codify, Share, Leverage, and Process) and Contextualised Content (Information). The paper aims to discuss these issues. Design/methodology/approach The aim of this paper is to add to the body of knowledge in the KM discipline and supply KM practitioners and scholars with insight into what is commonly regarded to be KM so as to reignite the debate on what one could consider as KM. The lexicon used by KM scholars was evaluated though the application of lexicographical research methods as extended though Knowledge Discovery and Text Analysis methods. Findings By simplifying term relationships through the application of lexicographical research methods, as extended though Knowledge Discovery and Text Analysis methods, it was found that KM is directly defined by People (Person and Organisation), Processes (Codify, Share, Leverage, Process) and Contextualised Content (Information). One would therefore be able to indicate that KM, from an academic point of view, refers to people processing contextualised content.
    Research limitations/implications In total, 42 definitions were identified spanning a period of 11 years. This represented the first use of KM through the estimated apex of terms used. From 2006 onwards definitions were used in repetition, and all definitions that were considered to repeat were therefore subsequently excluded as not being unique instances. All definitions listed are by no means complete and exhaustive. The definitions are viewed outside the scope and context in which they were originally formulated and then used to review the key concepts in the definitions themselves. Social implications When the authors refer to the aforementioned discussion of KM content as well as the presentation of the method followed in this paper, the authors may have a few implications for future research in KM. First the research validates ideas presented by the OECD in 2005 pertaining to KM. It also validates that through the evolution of KM, the authors ended with a description of KM that may be seen as a standardised description. If the authors as academics and practitioners, for example, refer to KM as the same construct and/or idea, it has the potential to speculatively, distinguish between what KM may or may not be. Originality/value By simplifying the term used to define KM, by focusing on the most common definitions, the paper assist in refocusing KM by reconsidering the dimensions that is the most common in how it has been defined over time. This would hopefully assist in reigniting discussions about KM and how it may be used to the benefit of an organisation.
    Date
    20. 1.2015 18:30:22
  7. Kay, M.: ¬The proper place of men and machines in language translation (1997) 0.04
    0.037599873 = product of:
      0.075199746 = sum of:
        0.075199746 = sum of:
          0.025232209 = weight(_text_:m in 1178) [ClassicSimilarity], result of:
            0.025232209 = score(doc=1178,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.19245613 = fieldWeight in 1178, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.0546875 = fieldNorm(doc=1178)
          0.049967535 = weight(_text_:22 in 1178) [ClassicSimilarity], result of:
            0.049967535 = score(doc=1178,freq=2.0), product of:
              0.18449724 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052685954 = queryNorm
              0.2708308 = fieldWeight in 1178, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=1178)
      0.5 = coord(1/2)
    
    Date
    31. 7.1996 9:22:19
  8. Melby, A.: Some notes on 'The proper place of men and machines in language translation' (1997) 0.04
    0.037599873 = product of:
      0.075199746 = sum of:
        0.075199746 = sum of:
          0.025232209 = weight(_text_:m in 330) [ClassicSimilarity], result of:
            0.025232209 = score(doc=330,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.19245613 = fieldWeight in 330, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.0546875 = fieldNorm(doc=330)
          0.049967535 = weight(_text_:22 in 330) [ClassicSimilarity], result of:
            0.049967535 = score(doc=330,freq=2.0), product of:
              0.18449724 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052685954 = queryNorm
              0.2708308 = fieldWeight in 330, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=330)
      0.5 = coord(1/2)
    
    Abstract
    Responds to Kay, M.: The proper place of men and machines in language translation. Examines the appropriateness of machine translation (MT) under the following special circumstances: controlled domain-specific text and high-quality output; controlled domain-specific text and indicative output; dynamic general text and indicative output and dynamic general text and high-quality output. MT is appropriate in the 1st 3 cases but the 4th case requires human translation. Examines how MT research could be more useful for aiding human translation
    Date
    31. 7.1996 9:22:19
  9. Melucci, M.; Orio, N.: Design, implementation, and evaluation of a methodology for automatic stemmer generation (2007) 0.03
    0.03378712 = sum of:
      0.021171015 = product of:
        0.08468406 = sum of:
          0.08468406 = weight(_text_:authors in 268) [ClassicSimilarity], result of:
            0.08468406 = score(doc=268,freq=2.0), product of:
              0.24018547 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.052685954 = queryNorm
              0.35257778 = fieldWeight in 268, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0546875 = fieldNorm(doc=268)
        0.25 = coord(1/4)
      0.012616104 = product of:
        0.025232209 = sum of:
          0.025232209 = weight(_text_:m in 268) [ClassicSimilarity], result of:
            0.025232209 = score(doc=268,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.19245613 = fieldWeight in 268, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.0546875 = fieldNorm(doc=268)
        0.5 = coord(1/2)
    
    Abstract
    The authors describe a statistical approach based on hidden Markov models (HMMs), for generating stemmers automatically. The proposed approach requires little effort to insert new languages in the system even if minimal linguistic knowledge is available. This is a key advantage especially for digital libraries, which are often developed for a specific institution or government because the program can manage a great amount of documents written in local languages. The evaluation described in the article shows that the stemmers implemented by means of HMMs are as effective as those based on linguistic rules.
  10. Sienel, J.; Weiss, M.; Laube, M.: Sprachtechnologien für die Informationsgesellschaft des 21. Jahrhunderts (2000) 0.03
    0.030589739 = product of:
      0.061179478 = sum of:
        0.061179478 = sum of:
          0.02548838 = weight(_text_:m in 5557) [ClassicSimilarity], result of:
            0.02548838 = score(doc=5557,freq=4.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.19441006 = fieldWeight in 5557, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5557)
          0.035691097 = weight(_text_:22 in 5557) [ClassicSimilarity], result of:
            0.035691097 = score(doc=5557,freq=2.0), product of:
              0.18449724 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052685954 = queryNorm
              0.19345059 = fieldWeight in 5557, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5557)
      0.5 = coord(1/2)
    
    Date
    26.12.2000 13:22:17
  11. Computational linguistics for the new millennium : divergence or synergy? Proceedings of the International Symposium held at the Ruprecht-Karls Universität Heidelberg, 21-22 July 2000. Festschrift in honour of Peter Hellwig on the occasion of his 60th birthday (2002) 0.03
    0.030589739 = product of:
      0.061179478 = sum of:
        0.061179478 = sum of:
          0.02548838 = weight(_text_:m in 4900) [ClassicSimilarity], result of:
            0.02548838 = score(doc=4900,freq=4.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.19441006 = fieldWeight in 4900, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4900)
          0.035691097 = weight(_text_:22 in 4900) [ClassicSimilarity], result of:
            0.035691097 = score(doc=4900,freq=2.0), product of:
              0.18449724 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052685954 = queryNorm
              0.19345059 = fieldWeight in 4900, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4900)
      0.5 = coord(1/2)
    
    Editor
    Klenner, M. u. H. Visser
    Type
    m
  12. Levin, M.; Krawczyk, S.; Bethard, S.; Jurafsky, D.: Citation-based bootstrapping for large-scale author disambiguation (2012) 0.03
    0.03039746 = sum of:
      0.021385957 = product of:
        0.085543826 = sum of:
          0.085543826 = weight(_text_:authors in 246) [ClassicSimilarity], result of:
            0.085543826 = score(doc=246,freq=4.0), product of:
              0.24018547 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.052685954 = queryNorm
              0.35615736 = fieldWeight in 246, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0390625 = fieldNorm(doc=246)
        0.25 = coord(1/4)
      0.009011503 = product of:
        0.018023007 = sum of:
          0.018023007 = weight(_text_:m in 246) [ClassicSimilarity], result of:
            0.018023007 = score(doc=246,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.13746867 = fieldWeight in 246, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.0390625 = fieldNorm(doc=246)
        0.5 = coord(1/2)
    
    Abstract
    We present a new, two-stage, self-supervised algorithm for author disambiguation in large bibliographic databases. In the first "bootstrap" stage, a collection of high-precision features is used to bootstrap a training set with positive and negative examples of coreferring authors. A supervised feature-based classifier is then trained on the bootstrap clusters and used to cluster the authors in a larger unlabeled dataset. Our self-supervised approach shares the advantages of unsupervised approaches (no need for expensive hand labels) as well as supervised approaches (a rich set of features that can be discriminatively trained). The algorithm disambiguates 54,000,000 author instances in Thomson Reuters' Web of Knowledge with B3 F1 of.807. We analyze parameters and features, particularly those from citation networks, which have not been deeply investigated in author disambiguation. The most important citation feature is self-citation, which can be approximated without expensive extraction of the full network. For the supervised stage, the minor improvement due to other citation features (increasing F1 from.748 to.767) suggests they may not be worth the trouble of extracting from databases that don't already have them. A lean feature set without expensive abstract and title features performs 130 times faster with about equal F1.
  13. Sparck Jones, K.; Galliers, J.R.: Evaluating natural language processing systems : an analysis and review (1996) 0.03
    0.028960388 = sum of:
      0.018146584 = product of:
        0.072586335 = sum of:
          0.072586335 = weight(_text_:authors in 2934) [ClassicSimilarity], result of:
            0.072586335 = score(doc=2934,freq=2.0), product of:
              0.24018547 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.052685954 = queryNorm
              0.30220953 = fieldWeight in 2934, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.046875 = fieldNorm(doc=2934)
        0.25 = coord(1/4)
      0.010813803 = product of:
        0.021627607 = sum of:
          0.021627607 = weight(_text_:m in 2934) [ClassicSimilarity], result of:
            0.021627607 = score(doc=2934,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.1649624 = fieldWeight in 2934, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.046875 = fieldNorm(doc=2934)
        0.5 = coord(1/2)
    
    Abstract
    This comprehensive state-of-the-art book is the first devoted to the important and timely issue of evaluating NLP systems. It addresses the whole area of NLP system evaluation, including aims and scope, problems and methodology. The authors provide a wide-ranging and careful analysis of evaluation concepts, reinforced with extensive illustrations; they relate systems to their environments and develop a framework for proper evaluation. The discussion of principles is completed by a detailed review of practice and strategies in the field, covering both systems for specific tasks, like translation, and core language processors. The methodology lessons drawn from the analysis and review are applied in a series of example cases. A comprehensive bibliography, a subject index, and term glossary are included
    Type
    m
  14. Sidhom, S.; Hassoun, M.: Morpho-syntactic parsing for a text mining environment : An NP recognition model for knowledge visualization and information retrieval (2002) 0.03
    0.028960388 = sum of:
      0.018146584 = product of:
        0.072586335 = sum of:
          0.072586335 = weight(_text_:authors in 1852) [ClassicSimilarity], result of:
            0.072586335 = score(doc=1852,freq=2.0), product of:
              0.24018547 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.052685954 = queryNorm
              0.30220953 = fieldWeight in 1852, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.046875 = fieldNorm(doc=1852)
        0.25 = coord(1/4)
      0.010813803 = product of:
        0.021627607 = sum of:
          0.021627607 = weight(_text_:m in 1852) [ClassicSimilarity], result of:
            0.021627607 = score(doc=1852,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.1649624 = fieldWeight in 1852, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.046875 = fieldNorm(doc=1852)
        0.5 = coord(1/2)
    
    Abstract
    Sidhom and Hassoun discuss the crucial role of NLP tools in Knowledge Extraction and Management as well as in the design of Information Retrieval Systems. The authors focus more specifically an the morpho-syntactic issues by describing their morpho-syntactic analysis platform, which has been implemented to cover the automatic indexing and information retrieval topics. To this end they implemented the Cascaded "Augmented Transition Network (ATN)". They used this formalism in order to analyse French text descriptions of Multimedia documents. An implementation of an ATN parsing automaton is briefly described. The Platform in its logical operation is considered as an investigative tool towards the knowledge organization (based an an NP recognition model) and management of multiform e-documents (text, multimedia, audio, image) using their text descriptions.
  15. Lu, C.; Bu, Y.; Wang, J.; Ding, Y.; Torvik, V.; Schnaars, M.; Zhang, C.: Examining scientific writing styles from the perspective of linguistic complexity : a cross-level moderation model (2019) 0.03
    0.028960388 = sum of:
      0.018146584 = product of:
        0.072586335 = sum of:
          0.072586335 = weight(_text_:authors in 5219) [ClassicSimilarity], result of:
            0.072586335 = score(doc=5219,freq=2.0), product of:
              0.24018547 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.052685954 = queryNorm
              0.30220953 = fieldWeight in 5219, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.046875 = fieldNorm(doc=5219)
        0.25 = coord(1/4)
      0.010813803 = product of:
        0.021627607 = sum of:
          0.021627607 = weight(_text_:m in 5219) [ClassicSimilarity], result of:
            0.021627607 = score(doc=5219,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.1649624 = fieldWeight in 5219, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.046875 = fieldNorm(doc=5219)
        0.5 = coord(1/2)
    
    Abstract
    Publishing articles in high-impact English journals is difficult for scholars around the world, especially for non-native English-speaking scholars (NNESs), most of whom struggle with proficiency in English. To uncover the differences in English scientific writing between native English-speaking scholars (NESs) and NNESs, we collected a large-scale data set containing more than 150,000 full-text articles published in PLoS between 2006 and 2015. We divided these articles into three groups according to the ethnic backgrounds of the first and corresponding authors, obtained by Ethnea, and examined the scientific writing styles in English from a two-fold perspective of linguistic complexity: (a) syntactic complexity, including measurements of sentence length and sentence complexity; and (b) lexical complexity, including measurements of lexical diversity, lexical density, and lexical sophistication. The observations suggest marginal differences between groups in syntactical and lexical complexity.
  16. Warner, A.J.: Natural language processing (1987) 0.03
    0.028552879 = product of:
      0.057105757 = sum of:
        0.057105757 = product of:
          0.114211515 = sum of:
            0.114211515 = weight(_text_:22 in 337) [ClassicSimilarity], result of:
              0.114211515 = score(doc=337,freq=2.0), product of:
                0.18449724 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052685954 = queryNorm
                0.61904186 = fieldWeight in 337, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=337)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Annual review of information science and technology. 22(1987), S.79-108
  17. Pinker, S.: Wörter und Regeln : Die Natur der Sprache (2000) 0.03
    0.026857052 = product of:
      0.053714104 = sum of:
        0.053714104 = sum of:
          0.018023007 = weight(_text_:m in 734) [ClassicSimilarity], result of:
            0.018023007 = score(doc=734,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.13746867 = fieldWeight in 734, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.0390625 = fieldNorm(doc=734)
          0.035691097 = weight(_text_:22 in 734) [ClassicSimilarity], result of:
            0.035691097 = score(doc=734,freq=2.0), product of:
              0.18449724 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052685954 = queryNorm
              0.19345059 = fieldWeight in 734, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=734)
      0.5 = coord(1/2)
    
    Date
    19. 7.2002 14:22:31
    Type
    m
  18. Semantik, Lexikographie und Computeranwendungen : Workshop ... (Bonn) : 1995.01.27-28 (1996) 0.03
    0.026857052 = product of:
      0.053714104 = sum of:
        0.053714104 = sum of:
          0.018023007 = weight(_text_:m in 190) [ClassicSimilarity], result of:
            0.018023007 = score(doc=190,freq=2.0), product of:
              0.13110629 = queryWeight, product of:
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.052685954 = queryNorm
              0.13746867 = fieldWeight in 190, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.4884486 = idf(docFreq=9980, maxDocs=44218)
                0.0390625 = fieldNorm(doc=190)
          0.035691097 = weight(_text_:22 in 190) [ClassicSimilarity], result of:
            0.035691097 = score(doc=190,freq=2.0), product of:
              0.18449724 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052685954 = queryNorm
              0.19345059 = fieldWeight in 190, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=190)
      0.5 = coord(1/2)
    
    Date
    14. 4.2007 10:04:22
    Type
    m
  19. McMahon, J.G.; Smith, F.J.: Improved statistical language model performance with automatic generated word hierarchies (1996) 0.02
    0.024983767 = product of:
      0.049967535 = sum of:
        0.049967535 = product of:
          0.09993507 = sum of:
            0.09993507 = weight(_text_:22 in 3164) [ClassicSimilarity], result of:
              0.09993507 = score(doc=3164,freq=2.0), product of:
                0.18449724 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052685954 = queryNorm
                0.5416616 = fieldWeight in 3164, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3164)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Computational linguistics. 22(1996) no.2, S.217-248
  20. Ruge, G.: ¬A spreading activation network for automatic generation of thesaurus relationships (1991) 0.02
    0.024983767 = product of:
      0.049967535 = sum of:
        0.049967535 = product of:
          0.09993507 = sum of:
            0.09993507 = weight(_text_:22 in 4506) [ClassicSimilarity], result of:
              0.09993507 = score(doc=4506,freq=2.0), product of:
                0.18449724 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052685954 = queryNorm
                0.5416616 = fieldWeight in 4506, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4506)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    8.10.2000 11:52:22

Languages

  • e 179
  • d 70
  • m 6
  • ru 1
  • More… Less…

Types

  • a 144
  • m 92
  • s 25
  • el 22
  • x 4
  • b 2
  • d 2
  • p 2
  • r 1
  • More… Less…

Subjects

Classifications