Search (489 results, page 1 of 25)

  • × theme_ss:"Computerlinguistik"
  1. Haas, S.W.: Natural language processing : toward large-scale, robust systems (1996) 0.09
    0.09179538 = product of:
      0.13769306 = sum of:
        0.02001815 = weight(_text_:information in 7415) [ClassicSimilarity], result of:
          0.02001815 = score(doc=7415,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.21943474 = fieldWeight in 7415, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=7415)
        0.11767492 = sum of:
          0.061349012 = weight(_text_:systems in 7415) [ClassicSimilarity], result of:
            0.061349012 = score(doc=7415,freq=4.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.38414678 = fieldWeight in 7415, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0625 = fieldNorm(doc=7415)
          0.05632591 = weight(_text_:22 in 7415) [ClassicSimilarity], result of:
            0.05632591 = score(doc=7415,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.30952093 = fieldWeight in 7415, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=7415)
      0.6666667 = coord(2/3)
    
    Abstract
    State of the art review of natural language processing updating an earlier review published in ARIST 22(1987). Discusses important developments that have allowed for significant advances in the field of natural language processing: materials and resources; knowledge based systems and statistical approaches; and a strong emphasis on evaluation. Reviews some natural language processing applications and common problems still awaiting solution. Considers closely related applications such as language generation and th egeneration phase of machine translation which face the same problems as natural language processing. Covers natural language methodologies for information retrieval only briefly
    Source
    Annual review of information science and technology. 31(1996), S.83-119
  2. Liddy, E.D.: Natural language processing for information retrieval and knowledge discovery (1998) 0.08
    0.07662532 = product of:
      0.11493797 = sum of:
        0.02769504 = weight(_text_:information in 2345) [ClassicSimilarity], result of:
          0.02769504 = score(doc=2345,freq=10.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3035872 = fieldWeight in 2345, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2345)
        0.08724293 = sum of:
          0.03795776 = weight(_text_:systems in 2345) [ClassicSimilarity], result of:
            0.03795776 = score(doc=2345,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.23767869 = fieldWeight in 2345, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2345)
          0.04928517 = weight(_text_:22 in 2345) [ClassicSimilarity], result of:
            0.04928517 = score(doc=2345,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.2708308 = fieldWeight in 2345, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2345)
      0.6666667 = coord(2/3)
    
    Abstract
    Natural language processing (NLP) is a powerful technology for the vital tasks of information retrieval (IR) and knowledge discovery (KD) which, in turn, feed the visualization systems of the present and future and enable knowledge workers to focus more of their time on the vital tasks of analysis and prediction
    Date
    22. 9.1997 19:16:05
    Imprint
    Urbana-Champaign, IL : Illinois University at Urbana-Champaign, Graduate School of Library and Information Science
    Source
    Visualizing subject access for 21st century information resources: Papers presented at the 1997 Clinic on Library Applications of Data Processing, 2-4 Mar 1997, Graduate School of Library and Information Science, University of Illinois at Urbana-Champaign. Ed.: P.A. Cochrane et al
  3. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.07
    0.069105834 = product of:
      0.10365874 = sum of:
        0.082536526 = product of:
          0.24760957 = sum of:
            0.24760957 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.24760957 = score(doc=562,freq=2.0), product of:
                0.4405723 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.051966466 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.021122215 = product of:
          0.04224443 = sum of:
            0.04224443 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.04224443 = score(doc=562,freq=2.0), product of:
                0.1819777 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051966466 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  4. Doszkocs, T.E.; Zamora, A.: Dictionary services and spelling aids for Web searching (2004) 0.06
    0.059606384 = product of:
      0.089409575 = sum of:
        0.012511344 = weight(_text_:information in 2541) [ClassicSimilarity], result of:
          0.012511344 = score(doc=2541,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.13714671 = fieldWeight in 2541, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2541)
        0.07689823 = sum of:
          0.027112689 = weight(_text_:systems in 2541) [ClassicSimilarity], result of:
            0.027112689 = score(doc=2541,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.1697705 = fieldWeight in 2541, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2541)
          0.04978554 = weight(_text_:22 in 2541) [ClassicSimilarity], result of:
            0.04978554 = score(doc=2541,freq=4.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.27358043 = fieldWeight in 2541, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2541)
      0.6666667 = coord(2/3)
    
    Abstract
    The Specialized Information Services Division (SIS) of the National Library of Medicine (NLM) provides Web access to more than a dozen scientific databases on toxicology and the environment on TOXNET . Search queries on TOXNET often include misspelled or variant English words, medical and scientific jargon and chemical names. Following the example of search engines like Google and ClinicalTrials.gov, we set out to develop a spelling "suggestion" system for increased recall and precision in TOXNET searching. This paper describes development of dictionary technology that can be used in a variety of applications such as orthographic verification, writing aid, natural language processing, and information storage and retrieval. The design of the technology allows building complex applications using the components developed in the earlier phases of the work in a modular fashion without extensive rewriting of computer code. Since many of the potential applications envisioned for this work have on-line or web-based interfaces, the dictionaries and other computer components must have fast response, and must be adaptable to open-ended database vocabularies, including chemical nomenclature. The dictionary vocabulary for this work was derived from SIS and other databases and specialized resources, such as NLM's Unified Medical Language Systems (UMLS) . The resulting technology, A-Z Dictionary (AZdict), has three major constituents: 1) the vocabulary list, 2) the word attributes that define part of speech and morphological relationships between words in the list, and 3) a set of programs that implements the retrieval of words and their attributes, and determines similarity between words (ChemSpell). These three components can be used in various applications such as spelling verification, spelling aid, part-of-speech tagging, paraphrasing, and many other natural language processing functions.
    Date
    14. 8.2004 17:22:56
    Source
    Online. 28(2004) no.3, S.22-29
  5. Dorr, B.J.: Large-scale dictionary construction for foreign language tutoring and interlingual machine translation (1997) 0.06
    0.056930594 = product of:
      0.08539589 = sum of:
        0.010616227 = weight(_text_:information in 3244) [ClassicSimilarity], result of:
          0.010616227 = score(doc=3244,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.116372846 = fieldWeight in 3244, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3244)
        0.07477966 = sum of:
          0.032535225 = weight(_text_:systems in 3244) [ClassicSimilarity], result of:
            0.032535225 = score(doc=3244,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.2037246 = fieldWeight in 3244, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.046875 = fieldNorm(doc=3244)
          0.04224443 = weight(_text_:22 in 3244) [ClassicSimilarity], result of:
            0.04224443 = score(doc=3244,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.23214069 = fieldWeight in 3244, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=3244)
      0.6666667 = coord(2/3)
    
    Abstract
    Describes techniques for automatic construction of dictionaries for use in large-scale foreign language tutoring (FLT) and interlingual machine translation (MT) systems. The dictionaries are based on a language independent representation called lexical conceptual structure (LCS). Demonstrates that synonymous verb senses share distribution patterns. Shows how the syntax-semantics relation can be used to develop a lexical acquisition approach that contributes both toward the enrichment of existing online resources and toward the development of lexicons containing more complete information than is provided in any of these resources alone. Describes the structure of the LCS and shows how this representation is used in FLT and MT. Focuses on the problem of building LCS dictionaries for large-scale FLT and MT. Describes authoring tools for manual and semi-automatic construction of LCS dictionaries. Presents an approach that uses linguistic techniques for building word definitions automatically. The techniques have been implemented as part of a set of lixicon-development tools used in the MILT FLT project
    Date
    31. 7.1996 9:22:19
  6. Warner, A.J.: Natural language processing (1987) 0.06
    0.0564239 = product of:
      0.08463585 = sum of:
        0.02830994 = weight(_text_:information in 337) [ClassicSimilarity], result of:
          0.02830994 = score(doc=337,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3103276 = fieldWeight in 337, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=337)
        0.05632591 = product of:
          0.11265182 = sum of:
            0.11265182 = weight(_text_:22 in 337) [ClassicSimilarity], result of:
              0.11265182 = score(doc=337,freq=2.0), product of:
                0.1819777 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051966466 = queryNorm
                0.61904186 = fieldWeight in 337, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=337)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Annual review of information science and technology. 22(1987), S.79-108
  7. Kreymer, O.: ¬An evaluation of help mechanisms in natural language information retrieval systems (2002) 0.05
    0.05302041 = product of:
      0.07953061 = sum of:
        0.028087895 = weight(_text_:information in 2557) [ClassicSimilarity], result of:
          0.028087895 = score(doc=2557,freq=14.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3078936 = fieldWeight in 2557, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2557)
        0.051442713 = product of:
          0.102885425 = sum of:
            0.102885425 = weight(_text_:systems in 2557) [ClassicSimilarity], result of:
              0.102885425 = score(doc=2557,freq=20.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.64423376 = fieldWeight in 2557, product of:
                  4.472136 = tf(freq=20.0), with freq of:
                    20.0 = termFreq=20.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2557)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The field of natural language processing (NLP) demonstrates rapid changes in the design of information retrieval systems and human-computer interaction. While natural language is being looked on as the most effective tool for information retrieval in a contemporary information environment, the systems using it are only beginning to emerge. This study attempts to evaluate the current state of NLP information retrieval systems from the user's point of view: what techniques are used by these systems to guide their users through the search process? The analysis focused on the structure and components of the systems' help mechanisms. Results of the study demonstrated that systems which claimed to be using natural language searching in fact used a wide range of information retrieval techniques from real natural language processing to Boolean searching. As a result, the user assistance mechanisms of these systems also varied. While pseudo-NLP systems would suit a more traditional method of instruction, real NLP systems primarily utilised the methods of explanation and user-system dialogue.
    Source
    Online information review. 26(2002) no.1, S.30-39
  8. Byrne, C.C.; McCracken, S.A.: ¬An adaptive thesaurus employing semantic distance, relational inheritance and nominal compound interpretation for linguistic support of information retrieval (1999) 0.05
    0.0481811 = product of:
      0.07227165 = sum of:
        0.030027224 = weight(_text_:information in 4483) [ClassicSimilarity], result of:
          0.030027224 = score(doc=4483,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3291521 = fieldWeight in 4483, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=4483)
        0.04224443 = product of:
          0.08448886 = sum of:
            0.08448886 = weight(_text_:22 in 4483) [ClassicSimilarity], result of:
              0.08448886 = score(doc=4483,freq=2.0), product of:
                0.1819777 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051966466 = queryNorm
                0.46428138 = fieldWeight in 4483, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4483)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Date
    15. 3.2000 10:22:37
    Source
    Journal of information science. 25(1999) no.2, S.113-131
  9. Solvberg, I.; Nordbo, I.; Aamodt, A.: Knowledge-based information retrieval (1991/92) 0.05
    0.047793493 = product of:
      0.07169024 = sum of:
        0.02830994 = weight(_text_:information in 546) [ClassicSimilarity], result of:
          0.02830994 = score(doc=546,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3103276 = fieldWeight in 546, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=546)
        0.0433803 = product of:
          0.0867606 = sum of:
            0.0867606 = weight(_text_:systems in 546) [ClassicSimilarity], result of:
              0.0867606 = score(doc=546,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.5432656 = fieldWeight in 546, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.125 = fieldNorm(doc=546)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Future generations computer systems. 7(1991/92), S. -
  10. Ciganik, M.: Pred koordinaciou a kooperaciou informacnych systemov (1997) 0.05
    0.046146613 = product of:
      0.06921992 = sum of:
        0.031651475 = weight(_text_:information in 950) [ClassicSimilarity], result of:
          0.031651475 = score(doc=950,freq=10.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3469568 = fieldWeight in 950, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=950)
        0.037568443 = product of:
          0.075136885 = sum of:
            0.075136885 = weight(_text_:systems in 950) [ClassicSimilarity], result of:
              0.075136885 = score(doc=950,freq=6.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.4704818 = fieldWeight in 950, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0625 = fieldNorm(doc=950)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The information requirements for library users can only be met if individual information systems are compatible, i.e. based on the use of a single information language. Points out that natural language is the best instrument for integration of information systems. Presents a model of the strucutre of natural language, extended by metaknowledge elements which makes it possible to analyse and represent text without the need for syntax analysis
    Footnote
    Übers. des Titels: Coordination of information systems
  11. Czejdo. B.D.; Tucci, R.P.: ¬A dataflow graphical language for database applications (1994) 0.04
    0.04224388 = product of:
      0.06336582 = sum of:
        0.025022687 = weight(_text_:information in 559) [ClassicSimilarity], result of:
          0.025022687 = score(doc=559,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.27429342 = fieldWeight in 559, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=559)
        0.03834313 = product of:
          0.07668626 = sum of:
            0.07668626 = weight(_text_:systems in 559) [ClassicSimilarity], result of:
              0.07668626 = score(doc=559,freq=4.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.48018348 = fieldWeight in 559, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.078125 = fieldNorm(doc=559)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Discusses a graphical language for information retrieval and processing. A lot of recent activity has occured in the area of improving access to database systems. However, current results are restricted to simple interfacing of database systems. Proposes a graphical language for specifying complex applications
    Source
    CIT - Journal of computing and information technology. 2(1994) no.1, S.39-50
  12. Krause, J.: Principles of content analysis for information retrieval systems : an overview (1996) 0.04
    0.041819304 = product of:
      0.06272896 = sum of:
        0.024771197 = weight(_text_:information in 5270) [ClassicSimilarity], result of:
          0.024771197 = score(doc=5270,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.27153665 = fieldWeight in 5270, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=5270)
        0.03795776 = product of:
          0.07591552 = sum of:
            0.07591552 = weight(_text_:systems in 5270) [ClassicSimilarity], result of:
              0.07591552 = score(doc=5270,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.47535738 = fieldWeight in 5270, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.109375 = fieldNorm(doc=5270)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
  13. Metzler, D.P.; Haas, S.W.: ¬The constituent object parser : syntactic structure matching for information retrieval (1989) 0.04
    0.03850607 = product of:
      0.0577591 = sum of:
        0.03064641 = weight(_text_:information in 3607) [ClassicSimilarity], result of:
          0.03064641 = score(doc=3607,freq=6.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3359395 = fieldWeight in 3607, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=3607)
        0.027112689 = product of:
          0.054225378 = sum of:
            0.054225378 = weight(_text_:systems in 3607) [ClassicSimilarity], result of:
              0.054225378 = score(doc=3607,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.339541 = fieldWeight in 3607, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3607)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The constituent object parser is designed to improve the precision and recall performance of information retrieval by providing more powerful matching procedures. Describes the dependency tree representations and the relationship between the intended use of the parser and its design.
    Source
    ACM transactions on information systems. 7(1989) no.3, S.292-316
  14. Bowker, L.: Information retrieval in translation memory systems : assessment of current limitations and possibilities for future development (2002) 0.04
    0.03842906 = product of:
      0.057643585 = sum of:
        0.024771197 = weight(_text_:information in 1854) [ClassicSimilarity], result of:
          0.024771197 = score(doc=1854,freq=8.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.27153665 = fieldWeight in 1854, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1854)
        0.032872386 = product of:
          0.06574477 = sum of:
            0.06574477 = weight(_text_:systems in 1854) [ClassicSimilarity], result of:
              0.06574477 = score(doc=1854,freq=6.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.41167158 = fieldWeight in 1854, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1854)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    A translation memory system is a new type of human language technology (HLT) tool that is gaining popularity among translators. Such tools allow translators to store previously translated texts in a type of aligned bilingual database, and to recycle relevant parts of these texts when producing new translations. Currently, these tools retrieve information from the database using superficial character string matching, which often results in poor precision and recall. This paper explains how translation memory systems work, and it considers some possible ways for introducing more sophisticated information retrieval techniques into such systems by taking syntactic and semantic similarity into account. Some of the suggested techniques are inspired by these used in other areas of HLT, and some by techniques used in information science.
  15. Paolillo, J.C.: Linguistics and the information sciences (2009) 0.04
    0.038274534 = product of:
      0.057411797 = sum of:
        0.032769214 = weight(_text_:information in 3840) [ClassicSimilarity], result of:
          0.032769214 = score(doc=3840,freq=14.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3592092 = fieldWeight in 3840, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3840)
        0.024642585 = product of:
          0.04928517 = sum of:
            0.04928517 = weight(_text_:22 in 3840) [ClassicSimilarity], result of:
              0.04928517 = score(doc=3840,freq=2.0), product of:
                0.1819777 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051966466 = queryNorm
                0.2708308 = fieldWeight in 3840, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3840)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Linguistics is the scientific study of language which emphasizes language spoken in everyday settings by human beings. It has a long history of interdisciplinarity, both internally and in contribution to other fields, including information science. A linguistic perspective is beneficial in many ways in information science, since it examines the relationship between the forms of meaningful expressions and their social, cognitive, institutional, and communicative context, these being two perspectives on information that are actively studied, to different degrees, in information science. Examples of issues relevant to information science are presented for which the approach taken under a linguistic perspective is illustrated.
    Date
    27. 8.2011 14:22:33
    Source
    Encyclopedia of library and information sciences. 3rd ed. Ed.: M.J. Bates
  16. Sparck Jones, K.: Synonymy and semantic classification (1986) 0.04
    0.037357897 = product of:
      0.056036845 = sum of:
        0.017693711 = weight(_text_:information in 1304) [ClassicSimilarity], result of:
          0.017693711 = score(doc=1304,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.19395474 = fieldWeight in 1304, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=1304)
        0.03834313 = product of:
          0.07668626 = sum of:
            0.07668626 = weight(_text_:systems in 1304) [ClassicSimilarity], result of:
              0.07668626 = score(doc=1304,freq=4.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.48018348 = fieldWeight in 1304, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1304)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    PRECIS
    Computer systems / Programming languages / Grammar
    Series
    Edinburgh information technology series ; 1
    Subject
    Computer systems / Programming languages / Grammar
  17. Zaitseva, E.M.: Developing linguistic tools of thematic search in library information systems (2023) 0.04
    0.0368904 = product of:
      0.055335596 = sum of:
        0.025022687 = weight(_text_:information in 1187) [ClassicSimilarity], result of:
          0.025022687 = score(doc=1187,freq=16.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.27429342 = fieldWeight in 1187, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1187)
        0.030312907 = product of:
          0.060625814 = sum of:
            0.060625814 = weight(_text_:systems in 1187) [ClassicSimilarity], result of:
              0.060625814 = score(doc=1187,freq=10.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.37961838 = fieldWeight in 1187, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1187)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Within the R&D program "Information support of research by scientists and specialists on the basis of RNPLS&T Open Archive - the system of scientific knowledge aggregation", the RNPLS&T analyzes the use of linguistic tools of thematic search in the modern library information systems and the prospects for their development. The author defines the key common characteristics of e-catalogs of the largest Russian libraries revealed at the first stage of the analysis. Based on the specified common characteristics and detailed comparison analysis, the author outlines and substantiates the vectors for enhancing search inter faces of e-catalogs. The focus is made on linguistic tools of thematic search in library information systems; the key vectors are suggested: use of thematic search at different search levels with the clear-cut level differentiation; use of combined functionality within thematic search system; implementation of classification search in all e-catalogs; hierarchical representation of classifications; use of the matching systems for classification information retrieval languages, and in the long term classification and verbal information retrieval languages, and various verbal information retrieval languages. The author formulates practical recommendations to improve thematic search in library information systems.
  18. Liddy, E.D.: Natural language processing for information retrieval (2009) 0.04
    0.035561085 = product of:
      0.053341627 = sum of:
        0.031651475 = weight(_text_:information in 3854) [ClassicSimilarity], result of:
          0.031651475 = score(doc=3854,freq=10.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3469568 = fieldWeight in 3854, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=3854)
        0.02169015 = product of:
          0.0433803 = sum of:
            0.0433803 = weight(_text_:systems in 3854) [ClassicSimilarity], result of:
              0.0433803 = score(doc=3854,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.2716328 = fieldWeight in 3854, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3854)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Natural language processing (NLP) is the computerized approach to analyzing text that is based on both a set of theories and a set of technologies. Although NLP is a relatively recent area of research and application, compared with other information technology approaches, there have been sufficient successes to date that suggest that NLP-based information access technologies will continue to be a major area of research and development in information systems now and into the future.
    Source
    Encyclopedia of library and information sciences. 3rd ed. Ed.: M.J. Bates
  19. Riloff, E.: ¬An empirical study of automated dictionary construction for information extraction in three domains (1996) 0.04
    0.035120055 = product of:
      0.052680083 = sum of:
        0.024517128 = weight(_text_:information in 6752) [ClassicSimilarity], result of:
          0.024517128 = score(doc=6752,freq=6.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.2687516 = fieldWeight in 6752, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
        0.028162954 = product of:
          0.05632591 = sum of:
            0.05632591 = weight(_text_:22 in 6752) [ClassicSimilarity], result of:
              0.05632591 = score(doc=6752,freq=2.0), product of:
                0.1819777 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051966466 = queryNorm
                0.30952093 = fieldWeight in 6752, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6752)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    AutoSlog is a system that addresses the knowledge engineering bottleneck for information extraction. AutoSlog automatically creates domain specific dictionaries for information extraction, given an appropriate training corpus. Describes experiments with AutoSlog in terrorism, joint ventures and microelectronics domains. Compares the performance of AutoSlog across the 3 domains, discusses the lessons learned and presents results from 2 experiments which demonstrate that novice users can generate effective dictionaries using AutoSlog
    Date
    6. 3.1997 16:22:15
  20. Frappaolo, C.: Artificial intelligence and text retrieval : a current perspective on the state of the art (1992) 0.03
    0.03475692 = product of:
      0.052135378 = sum of:
        0.025022687 = weight(_text_:information in 7097) [ClassicSimilarity], result of:
          0.025022687 = score(doc=7097,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.27429342 = fieldWeight in 7097, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=7097)
        0.027112689 = product of:
          0.054225378 = sum of:
            0.054225378 = weight(_text_:systems in 7097) [ClassicSimilarity], result of:
              0.054225378 = score(doc=7097,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.339541 = fieldWeight in 7097, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.078125 = fieldNorm(doc=7097)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Brief discussion of the ways in which computerized information retrieval and database searching can be enhanced by integrating artificial intelligence with such search systems. Explores the possibility of integrating the powers and capabilities of artificial intelligence (specifically natural language processing) with text retrieval
    Imprint
    Medford, NJ : Learned Information Inc.

Languages

Types

  • a 412
  • m 44
  • el 31
  • s 21
  • x 10
  • p 3
  • d 2
  • b 1
  • r 1
  • More… Less…

Subjects

Classifications