Search (476 results, page 1 of 24)

  • × theme_ss:"Computerlinguistik"
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.29
    0.28663272 = product of:
      0.42994907 = sum of:
        0.05995991 = product of:
          0.17987972 = sum of:
            0.17987972 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.17987972 = score(doc=562,freq=2.0), product of:
                0.3200604 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.037751827 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.17987972 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.17987972 = score(doc=562,freq=2.0), product of:
            0.3200604 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037751827 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.17987972 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.17987972 = score(doc=562,freq=2.0), product of:
            0.3200604 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037751827 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.010229703 = product of:
          0.030689107 = sum of:
            0.030689107 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.030689107 = score(doc=562,freq=2.0), product of:
                0.13220046 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037751827 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
      0.6666667 = coord(4/6)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  2. Huo, W.: Automatic multi-word term extraction and its application to Web-page summarization (2012) 0.25
    0.25180098 = product of:
      0.37770146 = sum of:
        0.17987972 = weight(_text_:2f in 563) [ClassicSimilarity], result of:
          0.17987972 = score(doc=563,freq=2.0), product of:
            0.3200604 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037751827 = queryNorm
            0.56201804 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.007712319 = weight(_text_:information in 563) [ClassicSimilarity], result of:
          0.007712319 = score(doc=563,freq=2.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.116372846 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.17987972 = weight(_text_:2f in 563) [ClassicSimilarity], result of:
          0.17987972 = score(doc=563,freq=2.0), product of:
            0.3200604 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037751827 = queryNorm
            0.56201804 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.010229703 = product of:
          0.030689107 = sum of:
            0.030689107 = weight(_text_:22 in 563) [ClassicSimilarity], result of:
              0.030689107 = score(doc=563,freq=2.0), product of:
                0.13220046 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037751827 = queryNorm
                0.23214069 = fieldWeight in 563, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=563)
          0.33333334 = coord(1/3)
      0.6666667 = coord(4/6)
    
    Abstract
    In this thesis we propose three new word association measures for multi-word term extraction. We combine these association measures with LocalMaxs algorithm in our extraction model and compare the results of different multi-word term extraction methods. Our approach is language and domain independent and requires no training data. It can be applied to such tasks as text summarization, information retrieval, and document classification. We further explore the potential of using multi-word terms as an effective representation for general web-page summarization. We extract multi-word terms from human written summaries in a large collection of web-pages, and generate the summaries by aligning document words with these multi-word terms. Our system applies machine translation technology to learn the aligning process from a training set and focuses on selecting high quality multi-word terms from human written summaries to generate suitable results for web-page summarization.
    Content
    A Thesis presented to The University of Guelph In partial fulfilment of requirements for the degree of Master of Science in Computer Science. Vgl. Unter: http://www.inf.ufrgs.br%2F~ceramisch%2Fdownload_files%2Fpublications%2F2009%2Fp01.pdf.
    Date
    10. 1.2013 19:22:47
  3. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.21
    0.20985968 = product of:
      0.41971937 = sum of:
        0.05995991 = product of:
          0.17987972 = sum of:
            0.17987972 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.17987972 = score(doc=862,freq=2.0), product of:
                0.3200604 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.037751827 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
        0.17987972 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.17987972 = score(doc=862,freq=2.0), product of:
            0.3200604 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037751827 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.17987972 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.17987972 = score(doc=862,freq=2.0), product of:
            0.3200604 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037751827 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
      0.5 = coord(3/6)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  4. Doszkocs, T.E.; Zamora, A.: Dictionary services and spelling aids for Web searching (2004) 0.02
    0.016801655 = product of:
      0.050404966 = sum of:
        0.0090890555 = weight(_text_:information in 2541) [ClassicSimilarity], result of:
          0.0090890555 = score(doc=2541,freq=4.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.13714671 = fieldWeight in 2541, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2541)
        0.04131591 = product of:
          0.061973862 = sum of:
            0.025806403 = weight(_text_:29 in 2541) [ClassicSimilarity], result of:
              0.025806403 = score(doc=2541,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.19432661 = fieldWeight in 2541, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2541)
            0.03616746 = weight(_text_:22 in 2541) [ClassicSimilarity], result of:
              0.03616746 = score(doc=2541,freq=4.0), product of:
                0.13220046 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037751827 = queryNorm
                0.27358043 = fieldWeight in 2541, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2541)
          0.6666667 = coord(2/3)
      0.33333334 = coord(2/6)
    
    Abstract
    The Specialized Information Services Division (SIS) of the National Library of Medicine (NLM) provides Web access to more than a dozen scientific databases on toxicology and the environment on TOXNET . Search queries on TOXNET often include misspelled or variant English words, medical and scientific jargon and chemical names. Following the example of search engines like Google and ClinicalTrials.gov, we set out to develop a spelling "suggestion" system for increased recall and precision in TOXNET searching. This paper describes development of dictionary technology that can be used in a variety of applications such as orthographic verification, writing aid, natural language processing, and information storage and retrieval. The design of the technology allows building complex applications using the components developed in the earlier phases of the work in a modular fashion without extensive rewriting of computer code. Since many of the potential applications envisioned for this work have on-line or web-based interfaces, the dictionaries and other computer components must have fast response, and must be adaptable to open-ended database vocabularies, including chemical nomenclature. The dictionary vocabulary for this work was derived from SIS and other databases and specialized resources, such as NLM's Unified Medical Language Systems (UMLS) . The resulting technology, A-Z Dictionary (AZdict), has three major constituents: 1) the vocabulary list, 2) the word attributes that define part of speech and morphological relationships between words in the list, and 3) a set of programs that implements the retrieval of words and their attributes, and determines similarity between words (ChemSpell). These three components can be used in various applications such as spelling verification, spelling aid, part-of-speech tagging, paraphrasing, and many other natural language processing functions.
    Date
    14. 8.2004 17:22:56
    Source
    Online. 28(2004) no.3, S.22-29
  5. Warner, A.J.: Natural language processing (1987) 0.02
    0.015948465 = product of:
      0.047845393 = sum of:
        0.020566184 = weight(_text_:information in 337) [ClassicSimilarity], result of:
          0.020566184 = score(doc=337,freq=2.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.3103276 = fieldWeight in 337, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=337)
        0.02727921 = product of:
          0.081837624 = sum of:
            0.081837624 = weight(_text_:22 in 337) [ClassicSimilarity], result of:
              0.081837624 = score(doc=337,freq=2.0), product of:
                0.13220046 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037751827 = queryNorm
                0.61904186 = fieldWeight in 337, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=337)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Source
    Annual review of information science and technology. 22(1987), S.79-108
  6. Byrne, C.C.; McCracken, S.A.: ¬An adaptive thesaurus employing semantic distance, relational inheritance and nominal compound interpretation for linguistic support of information retrieval (1999) 0.01
    0.0140910465 = product of:
      0.042273138 = sum of:
        0.021813732 = weight(_text_:information in 4483) [ClassicSimilarity], result of:
          0.021813732 = score(doc=4483,freq=4.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.3291521 = fieldWeight in 4483, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=4483)
        0.020459406 = product of:
          0.061378215 = sum of:
            0.061378215 = weight(_text_:22 in 4483) [ClassicSimilarity], result of:
              0.061378215 = score(doc=4483,freq=2.0), product of:
                0.13220046 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037751827 = queryNorm
                0.46428138 = fieldWeight in 4483, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4483)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Date
    15. 3.2000 10:22:37
    Source
    Journal of information science. 25(1999) no.2, S.113-131
  7. Proszeky, G.: Language technology tools in the translator's practice (1999) 0.01
    0.014027129 = product of:
      0.042081386 = sum of:
        0.017995412 = weight(_text_:information in 6873) [ClassicSimilarity], result of:
          0.017995412 = score(doc=6873,freq=2.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.27153665 = fieldWeight in 6873, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=6873)
        0.024085974 = product of:
          0.07225792 = sum of:
            0.07225792 = weight(_text_:29 in 6873) [ClassicSimilarity], result of:
              0.07225792 = score(doc=6873,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.5441145 = fieldWeight in 6873, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6873)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Date
    30. 3.2002 18:29:40
    Source
    Journal of computing and information technoloyg. 7(1999) no.3, S.221-227
  8. Ciganik, M.: Pred koordinaciou a kooperaciou informacnych systemov (1997) 0.01
    0.01225237 = product of:
      0.036757108 = sum of:
        0.022993693 = weight(_text_:information in 950) [ClassicSimilarity], result of:
          0.022993693 = score(doc=950,freq=10.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.3469568 = fieldWeight in 950, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=950)
        0.013763415 = product of:
          0.041290242 = sum of:
            0.041290242 = weight(_text_:29 in 950) [ClassicSimilarity], result of:
              0.041290242 = score(doc=950,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.31092256 = fieldWeight in 950, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=950)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    The information requirements for library users can only be met if individual information systems are compatible, i.e. based on the use of a single information language. Points out that natural language is the best instrument for integration of information systems. Presents a model of the strucutre of natural language, extended by metaknowledge elements which makes it possible to analyse and represent text without the need for syntax analysis
    Footnote
    Übers. des Titels: Coordination of information systems
    Source
    Kniznice a informacie. 29(1997) no.10, S.389-396
  9. Mustafa El Hadi, W.: Evaluating human language technology : general applications to information access and management (2002) 0.01
    0.012023254 = product of:
      0.036069762 = sum of:
        0.015424638 = weight(_text_:information in 1840) [ClassicSimilarity], result of:
          0.015424638 = score(doc=1840,freq=2.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.23274569 = fieldWeight in 1840, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=1840)
        0.020645123 = product of:
          0.061935365 = sum of:
            0.061935365 = weight(_text_:29 in 1840) [ClassicSimilarity], result of:
              0.061935365 = score(doc=1840,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.46638384 = fieldWeight in 1840, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1840)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Source
    Knowledge organization. 29(2002) nos.3/4, S.124-134
  10. Paolillo, J.C.: Linguistics and the information sciences (2009) 0.01
    0.011913449 = product of:
      0.035740346 = sum of:
        0.02380569 = weight(_text_:information in 3840) [ClassicSimilarity], result of:
          0.02380569 = score(doc=3840,freq=14.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.3592092 = fieldWeight in 3840, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3840)
        0.011934653 = product of:
          0.03580396 = sum of:
            0.03580396 = weight(_text_:22 in 3840) [ClassicSimilarity], result of:
              0.03580396 = score(doc=3840,freq=2.0), product of:
                0.13220046 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037751827 = queryNorm
                0.2708308 = fieldWeight in 3840, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3840)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    Linguistics is the scientific study of language which emphasizes language spoken in everyday settings by human beings. It has a long history of interdisciplinarity, both internally and in contribution to other fields, including information science. A linguistic perspective is beneficial in many ways in information science, since it examines the relationship between the forms of meaningful expressions and their social, cognitive, institutional, and communicative context, these being two perspectives on information that are actively studied, to different degrees, in information science. Examples of issues relevant to information science are presented for which the approach taken under a linguistic perspective is illustrated.
    Date
    27. 8.2011 14:22:33
    Source
    Encyclopedia of library and information sciences. 3rd ed. Ed.: M.J. Bates
  11. Czejdo. B.D.; Tucci, R.P.: ¬A dataflow graphical language for database applications (1994) 0.01
    0.0117941275 = product of:
      0.035382383 = sum of:
        0.018178111 = weight(_text_:information in 559) [ClassicSimilarity], result of:
          0.018178111 = score(doc=559,freq=4.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.27429342 = fieldWeight in 559, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=559)
        0.01720427 = product of:
          0.051612806 = sum of:
            0.051612806 = weight(_text_:29 in 559) [ClassicSimilarity], result of:
              0.051612806 = score(doc=559,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.38865322 = fieldWeight in 559, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=559)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    Discusses a graphical language for information retrieval and processing. A lot of recent activity has occured in the area of improving access to database systems. However, current results are restricted to simple interfacing of database systems. Proposes a graphical language for specifying complex applications
    Date
    20.10.2000 13:29:46
    Source
    CIT - Journal of computing and information technology. 2(1994) no.1, S.39-50
  12. Rau, L.F.: Conceptual information extraction and retrieval from natural language input (198) 0.01
    0.0117941275 = product of:
      0.035382383 = sum of:
        0.018178111 = weight(_text_:information in 1955) [ClassicSimilarity], result of:
          0.018178111 = score(doc=1955,freq=4.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.27429342 = fieldWeight in 1955, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=1955)
        0.01720427 = product of:
          0.051612806 = sum of:
            0.051612806 = weight(_text_:29 in 1955) [ClassicSimilarity], result of:
              0.051612806 = score(doc=1955,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.38865322 = fieldWeight in 1955, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1955)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Date
    16. 8.1998 13:29:20
    Footnote
    Wiederabgedruckt in: Readings in information retrieval. Ed.: K. Sparck Jones u. P. Willett. San Francisco: Morgan Kaufmann 1997. S.527-533
  13. Liddy, E.D.: Natural language processing for information retrieval and knowledge discovery (1998) 0.01
    0.010684712 = product of:
      0.032054134 = sum of:
        0.020119483 = weight(_text_:information in 2345) [ClassicSimilarity], result of:
          0.020119483 = score(doc=2345,freq=10.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.3035872 = fieldWeight in 2345, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2345)
        0.011934653 = product of:
          0.03580396 = sum of:
            0.03580396 = weight(_text_:22 in 2345) [ClassicSimilarity], result of:
              0.03580396 = score(doc=2345,freq=2.0), product of:
                0.13220046 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037751827 = queryNorm
                0.2708308 = fieldWeight in 2345, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2345)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    Natural language processing (NLP) is a powerful technology for the vital tasks of information retrieval (IR) and knowledge discovery (KD) which, in turn, feed the visualization systems of the present and future and enable knowledge workers to focus more of their time on the vital tasks of analysis and prediction
    Date
    22. 9.1997 19:16:05
    Imprint
    Urbana-Champaign, IL : Illinois University at Urbana-Champaign, Graduate School of Library and Information Science
    Source
    Visualizing subject access for 21st century information resources: Papers presented at the 1997 Clinic on Library Applications of Data Processing, 2-4 Mar 1997, Graduate School of Library and Information Science, University of Illinois at Urbana-Champaign. Ed.: P.A. Cochrane et al
  14. Mustafa El Hadi, W.: Terminologies, ontologies and information access (2006) 0.01
    0.010524752 = product of:
      0.031574253 = sum of:
        0.017810838 = weight(_text_:information in 1488) [ClassicSimilarity], result of:
          0.017810838 = score(doc=1488,freq=6.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.2687516 = fieldWeight in 1488, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=1488)
        0.013763415 = product of:
          0.041290242 = sum of:
            0.041290242 = weight(_text_:29 in 1488) [ClassicSimilarity], result of:
              0.041290242 = score(doc=1488,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.31092256 = fieldWeight in 1488, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1488)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    Ontologies have become an important issue in research communities across several disciplines. This paper discusses some of the innovative techniques involving automatic terminology resources acquisition are briefly discussed. Suggests that NLP-based ontologies are useful in reducing the cost of ontology engineering. Emphasizes that linguistic ontologies covering both ontological and lexical information can offer solutions since they can be more easily updated by the resources of NLP products.
    Date
    29. 2.2008 16:25:23
    Source
    Knowledge organization, information systems and other essays: Professor A. Neelameghan Festschrift. Ed. by K.S. Raghavan and K.N. Prasad
  15. Riloff, E.: ¬An empirical study of automated dictionary construction for information extraction in three domains (1996) 0.01
    0.010483481 = product of:
      0.031450443 = sum of:
        0.017810838 = weight(_text_:information in 6752) [ClassicSimilarity], result of:
          0.017810838 = score(doc=6752,freq=6.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.2687516 = fieldWeight in 6752, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
        0.013639605 = product of:
          0.040918812 = sum of:
            0.040918812 = weight(_text_:22 in 6752) [ClassicSimilarity], result of:
              0.040918812 = score(doc=6752,freq=2.0), product of:
                0.13220046 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037751827 = queryNorm
                0.30952093 = fieldWeight in 6752, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6752)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    AutoSlog is a system that addresses the knowledge engineering bottleneck for information extraction. AutoSlog automatically creates domain specific dictionaries for information extraction, given an appropriate training corpus. Describes experiments with AutoSlog in terrorism, joint ventures and microelectronics domains. Compares the performance of AutoSlog across the 3 domains, discusses the lessons learned and presents results from 2 experiments which demonstrate that novice users can generate effective dictionaries using AutoSlog
    Date
    6. 3.1997 16:22:15
  16. Wenzel, F.: Semantische Eingrenzung im Freitext-Retrieval auf der Basis morphologischer Segmentierungen (1980) 0.01
    0.010019379 = product of:
      0.030058134 = sum of:
        0.012853865 = weight(_text_:information in 2037) [ClassicSimilarity], result of:
          0.012853865 = score(doc=2037,freq=2.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.19395474 = fieldWeight in 2037, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=2037)
        0.01720427 = product of:
          0.051612806 = sum of:
            0.051612806 = weight(_text_:29 in 2037) [ClassicSimilarity], result of:
              0.051612806 = score(doc=2037,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.38865322 = fieldWeight in 2037, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2037)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    The basic problem in freetext retrieval is that the retrieval language is not properly adapted to that of the author. Morphological segmentation, where words with the same root are grouped together in the inverted file, is a good eliminator of noise and information loss, providing high recall but low precision
    Source
    Nachrichten für Dokumentation. 31(1980) H.1, S.29-35
  17. Sheremet'eva, S.O.: Teoreticheskie i metodologicheskie problemy inzhenernoi lingvistiki (1998) 0.01
    0.010019379 = product of:
      0.030058134 = sum of:
        0.012853865 = weight(_text_:information in 6316) [ClassicSimilarity], result of:
          0.012853865 = score(doc=6316,freq=2.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.19395474 = fieldWeight in 6316, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=6316)
        0.01720427 = product of:
          0.051612806 = sum of:
            0.051612806 = weight(_text_:29 in 6316) [ClassicSimilarity], result of:
              0.051612806 = score(doc=6316,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.38865322 = fieldWeight in 6316, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=6316)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    Examines the major topical issues in the area of linguistic engineering: machine translation, text synthesis and information retrieval
    Date
    6. 3.1999 13:56:29
  18. Liu, S.; Liu, F.; Yu, C.; Meng, W.: ¬An effective approach to document retrieval via utilizing WordNet and recognizing phrases (2004) 0.01
    0.010019379 = product of:
      0.030058134 = sum of:
        0.012853865 = weight(_text_:information in 4078) [ClassicSimilarity], result of:
          0.012853865 = score(doc=4078,freq=2.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.19395474 = fieldWeight in 4078, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=4078)
        0.01720427 = product of:
          0.051612806 = sum of:
            0.051612806 = weight(_text_:29 in 4078) [ClassicSimilarity], result of:
              0.051612806 = score(doc=4078,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.38865322 = fieldWeight in 4078, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4078)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Date
    10.10.2005 10:29:08
    Source
    SIGIR'04: Proceedings of the 27th Annual International ACM-SIGIR Conference an Research and Development in Information Retrieval. Ed.: K. Järvelin, u.a
  19. Bowker, L.: Information retrieval in translation memory systems : assessment of current limitations and possibilities for future development (2002) 0.01
    0.0100128 = product of:
      0.030038398 = sum of:
        0.017995412 = weight(_text_:information in 1854) [ClassicSimilarity], result of:
          0.017995412 = score(doc=1854,freq=8.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.27153665 = fieldWeight in 1854, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1854)
        0.012042987 = product of:
          0.03612896 = sum of:
            0.03612896 = weight(_text_:29 in 1854) [ClassicSimilarity], result of:
              0.03612896 = score(doc=1854,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.27205724 = fieldWeight in 1854, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1854)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    A translation memory system is a new type of human language technology (HLT) tool that is gaining popularity among translators. Such tools allow translators to store previously translated texts in a type of aligned bilingual database, and to recycle relevant parts of these texts when producing new translations. Currently, these tools retrieve information from the database using superficial character string matching, which often results in poor precision and recall. This paper explains how translation memory systems work, and it considers some possible ways for introducing more sophisticated information retrieval techniques into such systems by taking syntactic and semantic similarity into account. Some of the suggested techniques are inspired by these used in other areas of HLT, and some by techniques used in information science.
    Source
    Knowledge organization. 29(2002) nos.3/4, S.198-203
  20. Rindflesch, T.C.; Aronson, A.R.: Semantic processing in information retrieval (1993) 0.01
    0.0100128 = product of:
      0.030038398 = sum of:
        0.017995412 = weight(_text_:information in 4121) [ClassicSimilarity], result of:
          0.017995412 = score(doc=4121,freq=8.0), product of:
            0.0662725 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037751827 = queryNorm
            0.27153665 = fieldWeight in 4121, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4121)
        0.012042987 = product of:
          0.03612896 = sum of:
            0.03612896 = weight(_text_:29 in 4121) [ClassicSimilarity], result of:
              0.03612896 = score(doc=4121,freq=2.0), product of:
                0.13279912 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037751827 = queryNorm
                0.27205724 = fieldWeight in 4121, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4121)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    Intuition suggests that one way to enhance the information retrieval process would be the use of phrases to characterize the contents of text. A number of researchers, however, have noted that phrases alone do not improve retrieval effectiveness. In this paper we briefly review the use of phrases in information retrieval and then suggest extensions to this paradigm using semantic information. We claim that semantic processing, which can be viewed as expressing relations between the concepts represented by phrases, will in fact enhance retrieval effectiveness. The availability of the UMLS® domain model, which we exploit extensively, significantly contributes to the feasibility of this processing.
    Date
    29. 6.2015 14:51:28

Languages

Types

  • a 407
  • m 36
  • el 32
  • s 19
  • x 10
  • p 3
  • d 2
  • b 1
  • More… Less…

Subjects

Classifications