Search (34 results, page 1 of 2)

  • × year_i:[1980 TO 1990}
  • × theme_ss:"Computerlinguistik"
  1. McCune, B.P.; Tong, R.M.; Dean, J.S.: Rubric: a system for rule-based information retrieval (1985) 0.01
    0.009709007 = product of:
      0.06796305 = sum of:
        0.01712272 = weight(_text_:information in 1945) [ClassicSimilarity], result of:
          0.01712272 = score(doc=1945,freq=4.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.3291521 = fieldWeight in 1945, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=1945)
        0.050840326 = weight(_text_:retrieval in 1945) [ClassicSimilarity], result of:
          0.050840326 = score(doc=1945,freq=4.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.5671716 = fieldWeight in 1945, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.09375 = fieldNorm(doc=1945)
      0.14285715 = coord(2/14)
    
    Footnote
    Wiederabgedruckt in: Readings in information retrieval. Ed.: K. Sparck Jones u. P. Willett. San Francisco: Morgan Kaufmann 1997. S.440-445.
  2. Schwarz, C.: Linguistische Hilfsmittel beim Information Retrieval (1984) 0.01
    0.009153739 = product of:
      0.06407617 = sum of:
        0.016143454 = weight(_text_:information in 545) [ClassicSimilarity], result of:
          0.016143454 = score(doc=545,freq=2.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.3103276 = fieldWeight in 545, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=545)
        0.047932718 = weight(_text_:retrieval in 545) [ClassicSimilarity], result of:
          0.047932718 = score(doc=545,freq=2.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.5347345 = fieldWeight in 545, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.125 = fieldNorm(doc=545)
      0.14285715 = coord(2/14)
    
  3. Wenzel, F.: Semantische Eingrenzung im Freitext-Retrieval auf der Basis morphologischer Segmentierungen (1980) 0.01
    0.00885405 = product of:
      0.061978348 = sum of:
        0.010089659 = weight(_text_:information in 2037) [ClassicSimilarity], result of:
          0.010089659 = score(doc=2037,freq=2.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.19395474 = fieldWeight in 2037, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=2037)
        0.05188869 = weight(_text_:retrieval in 2037) [ClassicSimilarity], result of:
          0.05188869 = score(doc=2037,freq=6.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.5788671 = fieldWeight in 2037, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.078125 = fieldNorm(doc=2037)
      0.14285715 = coord(2/14)
    
    Abstract
    The basic problem in freetext retrieval is that the retrieval language is not properly adapted to that of the author. Morphological segmentation, where words with the same root are grouped together in the inverted file, is a good eliminator of noise and information loss, providing high recall but low precision
  4. Rau, L.F.; Jacobs, P.S.; Zernik, U.: Information extraction and text summarization using linguistic knowledge acquisition (1989) 0.01
    0.008754653 = product of:
      0.061282568 = sum of:
        0.019771613 = weight(_text_:information in 6683) [ClassicSimilarity], result of:
          0.019771613 = score(doc=6683,freq=12.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.38007212 = fieldWeight in 6683, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=6683)
        0.041510954 = weight(_text_:retrieval in 6683) [ClassicSimilarity], result of:
          0.041510954 = score(doc=6683,freq=6.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.46309367 = fieldWeight in 6683, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=6683)
      0.14285715 = coord(2/14)
    
    Abstract
    Storing and accessing texts in a conceptual format has a number of advantages over traditional document retrieval methods. A conceptual format facilitates natural language access to text information. It can support imprecise and inexact queries, conceptual information summarisation, and, ultimately, document translation. Describes 2 methods which have been implemented in a prototype intelligent information retrieval system calles SCISOR (System for Conceptual Information Summarisation, Organization and Retrieval). Describes the text processing, language acquisition, and summarisation components of SCISOR
    Source
    Information processing and management. 25(1989) no.4, S.419-428
  5. Metzler, D.P.; Haas, S.W.: ¬The constituent object parser : syntactic structure matching for information retrieval (1989) 0.01
    0.008548964 = product of:
      0.059842743 = sum of:
        0.017475804 = weight(_text_:information in 3607) [ClassicSimilarity], result of:
          0.017475804 = score(doc=3607,freq=6.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.3359395 = fieldWeight in 3607, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=3607)
        0.042366937 = weight(_text_:retrieval in 3607) [ClassicSimilarity], result of:
          0.042366937 = score(doc=3607,freq=4.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.47264296 = fieldWeight in 3607, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.078125 = fieldNorm(doc=3607)
      0.14285715 = coord(2/14)
    
    Abstract
    The constituent object parser is designed to improve the precision and recall performance of information retrieval by providing more powerful matching procedures. Describes the dependency tree representations and the relationship between the intended use of the parser and its design.
    Source
    ACM transactions on information systems. 7(1989) no.3, S.292-316
  6. Rau, L.F.: Conceptual information extraction and retrieval from natural language input (198) 0.01
    0.008090839 = product of:
      0.05663587 = sum of:
        0.014268933 = weight(_text_:information in 1955) [ClassicSimilarity], result of:
          0.014268933 = score(doc=1955,freq=4.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.27429342 = fieldWeight in 1955, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=1955)
        0.042366937 = weight(_text_:retrieval in 1955) [ClassicSimilarity], result of:
          0.042366937 = score(doc=1955,freq=4.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.47264296 = fieldWeight in 1955, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.078125 = fieldNorm(doc=1955)
      0.14285715 = coord(2/14)
    
    Footnote
    Wiederabgedruckt in: Readings in information retrieval. Ed.: K. Sparck Jones u. P. Willett. San Francisco: Morgan Kaufmann 1997. S.527-533
  7. Schwarz, C.: Natural language and information retrieval : Kommentierte Literaturliste zu Systemen, Verfahren und Tools (1986) 0.01
    0.008009522 = product of:
      0.05606665 = sum of:
        0.014125523 = weight(_text_:information in 408) [ClassicSimilarity], result of:
          0.014125523 = score(doc=408,freq=2.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.27153665 = fieldWeight in 408, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=408)
        0.04194113 = weight(_text_:retrieval in 408) [ClassicSimilarity], result of:
          0.04194113 = score(doc=408,freq=2.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.46789268 = fieldWeight in 408, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.109375 = fieldNorm(doc=408)
      0.14285715 = coord(2/14)
    
  8. Informationslinguistische Texterschließung (1986) 0.01
    0.0070068724 = product of:
      0.049048103 = sum of:
        0.012357258 = weight(_text_:information in 186) [ClassicSimilarity], result of:
          0.012357258 = score(doc=186,freq=12.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.23754507 = fieldWeight in 186, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=186)
        0.036690846 = weight(_text_:retrieval in 186) [ClassicSimilarity], result of:
          0.036690846 = score(doc=186,freq=12.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.40932083 = fieldWeight in 186, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=186)
      0.14285715 = coord(2/14)
    
    LCSH
    Information storage and retrieval systems / Linguistics
    RSWK
    Information Retrieval / Aufsatzsammlung (DNB)
    Linguistik / Information Retrieval / Aufsatzsammlung (SWB / BVB)
    Subject
    Information Retrieval / Aufsatzsammlung (DNB)
    Linguistik / Information Retrieval / Aufsatzsammlung (SWB / BVB)
    Information storage and retrieval systems / Linguistics
  9. Porter, M.F.: ¬An algorithm for suffix stripping (1980) 0.01
    0.006865305 = product of:
      0.04805713 = sum of:
        0.012107591 = weight(_text_:information in 3122) [ClassicSimilarity], result of:
          0.012107591 = score(doc=3122,freq=2.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.23274569 = fieldWeight in 3122, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=3122)
        0.03594954 = weight(_text_:retrieval in 3122) [ClassicSimilarity], result of:
          0.03594954 = score(doc=3122,freq=2.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.40105087 = fieldWeight in 3122, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.09375 = fieldNorm(doc=3122)
      0.14285715 = coord(2/14)
    
    Footnote
    Wiederabgedruckt in: Readings in information retrieval. Ed.: K. Sparck Jones u. P. Willett. San Francisco: Morgan Kaufmann 1997. S.313-316.
  10. Metzler, D.P.; Haas, S.W.; Cosic, C.L.; Wheeler, L.H.: Constituent object parsing for information retrieval and similar text processing problems (1989) 0.01
    0.00683917 = product of:
      0.04787419 = sum of:
        0.013980643 = weight(_text_:information in 2858) [ClassicSimilarity], result of:
          0.013980643 = score(doc=2858,freq=6.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.2687516 = fieldWeight in 2858, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=2858)
        0.033893548 = weight(_text_:retrieval in 2858) [ClassicSimilarity], result of:
          0.033893548 = score(doc=2858,freq=4.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.37811437 = fieldWeight in 2858, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=2858)
      0.14285715 = coord(2/14)
    
    Abstract
    Describes the architecture and functioning of the Constituent Object Parser. This system has been developed specially for text processing applications such as information retrieval, which can benefit from structural comparisons between elements of text such as a query and a potentially relevant abstract. Describes the general way in which this objective influenced the design of the system.
    Source
    Journal of the American Society for Information Science. 40(1989) no.6, S.398-423
  11. Hayes, P.J.; Knecht, L.E.; Cellio, M.J.: ¬A news story categorization system (1988) 0.01
    0.005721087 = product of:
      0.04004761 = sum of:
        0.010089659 = weight(_text_:information in 1954) [ClassicSimilarity], result of:
          0.010089659 = score(doc=1954,freq=2.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.19395474 = fieldWeight in 1954, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=1954)
        0.029957948 = weight(_text_:retrieval in 1954) [ClassicSimilarity], result of:
          0.029957948 = score(doc=1954,freq=2.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.33420905 = fieldWeight in 1954, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.078125 = fieldNorm(doc=1954)
      0.14285715 = coord(2/14)
    
    Footnote
    Wiederabgedruckt in: Readings in information retrieval. Ed.: K. Sparck Jones u. P. Willett. San Francisco: Morgan Kaufmann 1997. S.518-526
  12. Salton, G.: Automatic processing of foreign language documents (1985) 0.01
    0.0056823203 = product of:
      0.03977624 = sum of:
        0.008071727 = weight(_text_:information in 3650) [ClassicSimilarity], result of:
          0.008071727 = score(doc=3650,freq=8.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.1551638 = fieldWeight in 3650, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=3650)
        0.03170451 = weight(_text_:retrieval in 3650) [ClassicSimilarity], result of:
          0.03170451 = score(doc=3650,freq=14.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.3536936 = fieldWeight in 3650, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=3650)
      0.14285715 = coord(2/14)
    
    Abstract
    The attempt to computerize a process, such as indexing, abstracting, classifying, or retrieving information, begins with an analysis of the process into its intellectual and nonintellectual components. That part of the process which is amenable to computerization is mechanical or algorithmic. What is not is intellectual or creative and requires human intervention. Gerard Salton has been an innovator, experimenter, and promoter in the area of mechanized information systems since the early 1960s. He has been particularly ingenious at analyzing the process of information retrieval into its algorithmic components. He received a doctorate in applied mathematics from Harvard University before moving to the computer science department at Cornell, where he developed a prototype automatic retrieval system called SMART. Working with this system he and his students contributed for over a decade to our theoretical understanding of the retrieval process. On a more practical level, they have contributed design criteria for operating retrieval systems. The following selection presents one of the early descriptions of the SMART system; it is valuable as it shows the direction automatic retrieval methods were to take beyond simple word-matching techniques. These include various word normalization techniques to improve recall, for instance, the separation of words into stems and affixes; the correlation and clustering, using statistical association measures, of related terms; and the identification, using a concept thesaurus, of synonymous, broader, narrower, and sibling terms. They include, as weIl, techniques, both linguistic and statistical, to deal with the thorny problem of how to automatically extract from texts index terms that consist of more than one word. They include weighting techniques and various documentrequest matching algorithms. Significant among the latter are those which produce a retrieval output of citations ranked in relevante order. During the 1970s, Salton and his students went an to further refine these various techniques, particularly the weighting and statistical association measures. Many of their early innovations seem commonplace today. Some of their later techniques are still ahead of their time and await technological developments for implementation. The particular focus of the selection that follows is an the evaluation of a particular component of the SMART system, a multilingual thesaurus. By mapping English language expressions and their German equivalents to a common concept number, the thesaurus permitted the automatic processing of German language documents against English language queries and vice versa. The results of the evaluation, as it turned out, were somewhat inconclusive. However, this SMART experiment suggested in a bold and optimistic way how one might proceed to answer such complex questions as What is meant by retrieval language compatability? How it is to be achieved, and how evaluated?
    Footnote
    Original in: Journal of the American Society for Information Science 21(1970) no.3, S.187-194.
  13. Fagan, J.L.: ¬The effectiveness of a nonsyntactic approach to automatic phrase indexing for document retrieval (1989) 0.01
    0.0055279788 = product of:
      0.03869585 = sum of:
        0.008737902 = weight(_text_:information in 1845) [ClassicSimilarity], result of:
          0.008737902 = score(doc=1845,freq=6.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.16796975 = fieldWeight in 1845, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1845)
        0.029957948 = weight(_text_:retrieval in 1845) [ClassicSimilarity], result of:
          0.029957948 = score(doc=1845,freq=8.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.33420905 = fieldWeight in 1845, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1845)
      0.14285715 = coord(2/14)
    
    Abstract
    It may be possible to improve the quality of automatic indexing systems by using complex descriptors, for example, phrases, in addition to the simple descriptors (words or word stems) that are normally used in automatically constructed representations of document content. This study is directed toward the goal of developing effective methods of identifying phrases in natural language text from which good quality phrase descriptors can be constructed. The effectiveness of one method, a simple nonsyntactic phrase indexing procedure, has been tested on five experimental document collections. The results have been analyzed in order to identify the inadequacies of the procedure, and to determine what kinds of information about text structure are needed in order to construct phrase descriptors that are good indicators of document content. Two primary conclusions have been reached: (1) In the retrieval experiments, the nonsyntactic phrase construction procedure did not consistently yield substantial improvements in effectiveness. It is therefore not likely that phrase indexing of this kind will prove to be an important method of enhancing the performance of automatic document indexing and retrieval systems in operational environments. (2) Many of the shortcomings of the nonsyntactic approach can be overcome by incorporating syntactic information into the phrase construction process. However, a general syntactic analysis facility may be required, since many useful sources of phrases cannot be exploited if only a limited inventory of syntactic patterns can be recognized. Further research should be conducted into methods of incorporating automatic syntactic analysis into content analysis for document retrieval.
    Source
    Journal of the American Society for Information Science. 40(1989) no.2, S.115-132
  14. Hahn, U.: Informationslinguistik : I: Einführung in das linguistische Information Retrieval (1985) 0.01
    0.0054824487 = product of:
      0.03837714 = sum of:
        0.009024465 = weight(_text_:information in 3115) [ClassicSimilarity], result of:
          0.009024465 = score(doc=3115,freq=10.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.1734784 = fieldWeight in 3115, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=3115)
        0.029352674 = weight(_text_:retrieval in 3115) [ClassicSimilarity], result of:
          0.029352674 = score(doc=3115,freq=12.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.32745665 = fieldWeight in 3115, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=3115)
      0.14285715 = coord(2/14)
    
    Abstract
    (1) "Informationslinguistik I: Einfuehrung in das linguistische Information Retrieval" (2) "Informationslinguistik II: linguistische und statistische Verfahren im experimentellen Information Retrieval" (3) "Intelligente Informationssysteme: Verfahren der Kuenstlichen Intelligenz im experimentellen Information Retrieval" Kursabschnitt zu natuerlichsprachlichen Systemen (4) Spezialkurse zum automatischen Uebersetzen, Indexing und Retrieval, Abstracting usf. dienen zur Vertiefung informationslinguistischer Spezialthemen Die Kurse (1) und (3) gehoeren zu dem Pool der Pflichtveranstaltungen aller Studenten des Diplom-Studiengangs Informationswissenschaft, waehrend (2) und (4) lediglich zu den Wahlpflichtveranstaltungen zaehlen, die aber obligatorisch fuer die Studenten des Diplomstudiengangs sind, die ihren Schwerpunkt (z.B. in Form der Diplomarbeit) im Bereich Informationslinguistik suchen - fuer alle anderen Studenten zaehlen diese Kurse zum Zusatz angebot an Lehrveranstaltungen.
    Content
    2. Teil u.d.T.: Linguistische und statistische Verfahren im experimentellen Information Retrieval
  15. Warner, A.J.: Natural language processing (1987) 0.01
    0.005365187 = product of:
      0.037556306 = sum of:
        0.016143454 = weight(_text_:information in 337) [ClassicSimilarity], result of:
          0.016143454 = score(doc=337,freq=2.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.3103276 = fieldWeight in 337, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=337)
        0.021412853 = product of:
          0.064238556 = sum of:
            0.064238556 = weight(_text_:22 in 337) [ClassicSimilarity], result of:
              0.064238556 = score(doc=337,freq=2.0), product of:
                0.103770934 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029633347 = queryNorm
                0.61904186 = fieldWeight in 337, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=337)
          0.33333334 = coord(1/3)
      0.14285715 = coord(2/14)
    
    Source
    Annual review of information science and technology. 22(1987), S.79-108
  16. Hahn, U.: Informationslinguistik : II: Einführung in das linguistische Information Retrieval (1985) 0.01
    0.00519879 = product of:
      0.03639153 = sum of:
        0.00865008 = weight(_text_:information in 3116) [ClassicSimilarity], result of:
          0.00865008 = score(doc=3116,freq=12.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.16628155 = fieldWeight in 3116, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=3116)
        0.027741449 = weight(_text_:retrieval in 3116) [ClassicSimilarity], result of:
          0.027741449 = score(doc=3116,freq=14.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.30948192 = fieldWeight in 3116, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.02734375 = fieldNorm(doc=3116)
      0.14285715 = coord(2/14)
    
    Abstract
    (1) "Informationslinguistik I: Einfuehrung in das linguistische Information Retrieval" (2) "Informationslinguistik II: linguistische und statistische Verfahren im experimentellen Information Retrieval" (3) "Intelligente Informationssysteme: Verfahren der Kuenstlichen Intelligenz im experimentellen Information Retrieval" Kursabschnitt zu natuerlichsprachlichen Systemen (4) Spezialkurse zum automatischen Uebersetzen, Indexing und Retrieval, Abstracting usf. dienen zur Vertiefung informationslinguistischer Spezialthemen Die Kurse (1) und (3) gehoeren zu dem Pool der Pflichtveranstaltungen aller Studenten des Diplom-Studiengangs Informationswissenschaft, waehrend (2) und (4) lediglich zu den Wahlpflichtveranstaltungen zaehlen, die aber obligatorisch fuer die Studenten des Diplomstudiengangs sind, die ihren Schwerpunkt (z.B. in Form der Diplomarbeit) im Bereich Informationslinguistik suchen - fuer alle anderen Studenten zaehlen diese Kurse zum Zusatz angebot an Lehrveranstaltungen.
    Das vorliegende Skript entspricht dem Inhalt des Kurses "Informationslinguistik II" im SS 1983 bzw. SS 1984. Es ist im Juli 1983 inhaltlich abgeschlossen und im Januar 1985 lediglich redaktionell ueberarbeitet worden. Die Erstellung des Skripts entspricht einem dezidierten Auftrag des Projekts "Informationsvermittlung", der die Entwicklung geeigneter Lehrmaterialien zum informationswissenschaftlichen Aufbaustudium vorsah. Aufgrund des engen Projektzeitrahmens (1982-84) kann das Skript nicht in dem Masse voll ausgereift und ausformuliert sein, wie es gaengigen Standards entspraeche. Im Unterschied zum Skript "Informationslinguistik I" (HAHN 1985) laesst das vorliegende Skript wahlweise eine eher methoden- oder mehr systembezogene Darstellung informationslinguistischer Konzepte des experimentellen Information Retrieval zu (beides zusammen schliesst der enge Zeitrahmen eines Sommersemesters ausl). Die Entscheidung darueber sollte wenn moeglich in Abhaengigkeit zur personellen Zusammensetzung des Kurses getroffen werden, wobei - sofern die bislang genachten Erfahrungen verallgemeinerbar sind - sich bei einem nicht ausschliesslich an einer informationslinguistischen Spezialisierung interessierten und damit heterogenen Publikum die mehr systembezogene Praesentation als fuer das Verstaendnis informationslinguistischer Fragestellungen und entsprechender Verfahrensloesungen guenstiger gezeigt hat. Innerhalb dieser Nuancierung besitzt aber auch dieses Skript schon eine akzeptable inhaltliche Stabilitaet. Nichtsdestotrotz sollte gerade die Veroeffentlichung des Skripts als Anregung dienen, kritische Kommentare, Anmerkungen und Ergaenzungen zu diesem curricularen Entwurf herauszufordern, um damit die weitere disziplinaere Klaerung der Informationslinguistik zu foerdern.
    Content
    1. Teil u.d.T.: Einführung in das linguistische Information Retrieval
  17. Asija, S.P.: Natural language interface without artifical intelligence (1989) 0.00
    0.0045768693 = product of:
      0.032038085 = sum of:
        0.008071727 = weight(_text_:information in 5711) [ClassicSimilarity], result of:
          0.008071727 = score(doc=5711,freq=2.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.1551638 = fieldWeight in 5711, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=5711)
        0.023966359 = weight(_text_:retrieval in 5711) [ClassicSimilarity], result of:
          0.023966359 = score(doc=5711,freq=2.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.26736724 = fieldWeight in 5711, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=5711)
      0.14285715 = coord(2/14)
    
    Abstract
    SWIFT-ANSWER (Special Word Indexed Full Text Alpha Numeric Storage With Easy Retrieval) is a natural language interface that allows searchers to communicate with the computer in their own languages. The system operates without the need for artificial intelligence.
    Imprint
    Medford, NJ : Learned Information
  18. Ruge, G.; Schwarz, C.: ¬Die Leistungsfähigkeit von linguistischen Verfahren in der Massentextverarbeitung (1989) 0.00
    0.0045768693 = product of:
      0.032038085 = sum of:
        0.008071727 = weight(_text_:information in 6969) [ClassicSimilarity], result of:
          0.008071727 = score(doc=6969,freq=2.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.1551638 = fieldWeight in 6969, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=6969)
        0.023966359 = weight(_text_:retrieval in 6969) [ClassicSimilarity], result of:
          0.023966359 = score(doc=6969,freq=2.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.26736724 = fieldWeight in 6969, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=6969)
      0.14285715 = coord(2/14)
    
    Abstract
    Dependenzstrukturen stellen syntagmatische Relationen von Worten in Texten dar. Ihre Anwendungsmöglichkeiten im Information Retrieval werden erläutert. Bei Siemens wurde ein System zur Transformation von Texten in Dependenzstrukturen entwickelt, wobei besonders darauf geachtet wurde, die Wirkung gegen den Aufwand abzuwiegen. Die letzte Version verarbeitet 20 MB Freitext in einer Stunde Realzeit auf einem Siemens BS2000 Großrechner. Analyse-Recall and Analyse-Precision liegen jeweils bei 0,85
  19. Schwarz, C.: THESYS: Thesaurus Syntax System : a fully automatic thesaurus building aid (1988) 0.00
    0.004334098 = product of:
      0.030338686 = sum of:
        0.020970564 = weight(_text_:retrieval in 1361) [ClassicSimilarity], result of:
          0.020970564 = score(doc=1361,freq=2.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.23394634 = fieldWeight in 1361, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1361)
        0.009368123 = product of:
          0.028104367 = sum of:
            0.028104367 = weight(_text_:22 in 1361) [ClassicSimilarity], result of:
              0.028104367 = score(doc=1361,freq=2.0), product of:
                0.103770934 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029633347 = queryNorm
                0.2708308 = fieldWeight in 1361, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1361)
          0.33333334 = coord(1/3)
      0.14285715 = coord(2/14)
    
    Abstract
    THESYS is based on the natural language processing of free-text databases. It yields statistically evaluated correlations between words of the database. These correlations correspond to traditional thesaurus relations. The person who has to build a thesaurus is thus assisted by the proposals made by THESYS. THESYS is being tested on commercial databases under real world conditions. It is part of a text processing project at Siemens, called TINA (Text-Inhalts-Analyse). Software from TINA is actually being applied and evaluated by the US Department of Commerce for patent search and indexing (REALIST: REtrieval Aids by Linguistics and STatistics)
    Date
    6. 1.1999 10:22:07
  20. Needham, R.M.; Sparck Jones, K.: Keywords and clumps (1985) 0.00
    0.0040628472 = product of:
      0.02843993 = sum of:
        0.0049941265 = weight(_text_:information in 3645) [ClassicSimilarity], result of:
          0.0049941265 = score(doc=3645,freq=4.0), product of:
            0.052020688 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.029633347 = queryNorm
            0.0960027 = fieldWeight in 3645, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=3645)
        0.023445804 = weight(_text_:retrieval in 3645) [ClassicSimilarity], result of:
          0.023445804 = score(doc=3645,freq=10.0), product of:
            0.08963835 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.029633347 = queryNorm
            0.26155996 = fieldWeight in 3645, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.02734375 = fieldNorm(doc=3645)
      0.14285715 = coord(2/14)
    
    Abstract
    The selection that follows was chosen as it represents "a very early paper an the possibilities allowed by computers an documentation." In the early 1960s computers were being used to provide simple automatic indexing systems wherein keywords were extracted from documents. The problem with such systems was that they lacked vocabulary control, thus documents related in subject matter were not always collocated in retrieval. To improve retrieval by improving recall is the raison d'être of vocabulary control tools such as classifications and thesauri. The question arose whether it was possible by automatic means to construct classes of terms, which when substituted, one for another, could be used to improve retrieval performance? One of the first theoretical approaches to this question was initiated by R. M. Needham and Karen Sparck Jones at the Cambridge Language Research Institute in England.t The question was later pursued using experimental methodologies by Sparck Jones, who, as a Senior Research Associate in the Computer Laboratory at the University of Cambridge, has devoted her life's work to research in information retrieval and automatic naturai language processing. Based an the principles of numerical taxonomy, automatic classification techniques start from the premise that two objects are similar to the degree that they share attributes in common. When these two objects are keywords, their similarity is measured in terms of the number of documents they index in common. Step 1 in automatic classification is to compute mathematically the degree to which two terms are similar. Step 2 is to group together those terms that are "most similar" to each other, forming equivalence classes of intersubstitutable terms. The technique for forming such classes varies and is the factor that characteristically distinguishes different approaches to automatic classification. The technique used by Needham and Sparck Jones, that of clumping, is described in the selection that follows. Questions that must be asked are whether the use of automatically generated classes really does improve retrieval performance and whether there is a true eco nomic advantage in substituting mechanical for manual labor. Several years after her work with clumping, Sparck Jones was to observe that while it was not wholly satisfactory in itself, it was valuable in that it stimulated research into automatic classification. To this it might be added that it was valuable in that it introduced to libraryl information science the methods of numerical taxonomy, thus stimulating us to think again about the fundamental nature and purpose of classification. In this connection it might be useful to review how automatically derived classes differ from those of manually constructed classifications: 1) the manner of their derivation is purely a posteriori, the ultimate operationalization of the principle of literary warrant; 2) the relationship between members forming such classes is essentially statistical; the members of a given class are similar to each other not because they possess the class-defining characteristic but by virtue of sharing a family resemblance; and finally, 3) automatically derived classes are not related meaningfully one to another, that is, they are not ordered in traditional hierarchical and precedence relationships.

Languages

  • e 18
  • d 16

Types

  • a 25
  • m 7
  • s 2
  • d 1
  • More… Less…