Search (37 results, page 1 of 2)

  • × theme_ss:"Computerlinguistik"
  • × type_ss:"a"
  • × year_i:[1980 TO 1990}
  1. Gräbnitz, V.: PASSAT: Programm zur automatischen Selektion von Stichwörtern aus Texten (1987) 0.03
    0.033442385 = product of:
      0.06688477 = sum of:
        0.06406432 = weight(_text_:von in 932) [ClassicSimilarity], result of:
          0.06406432 = score(doc=932,freq=4.0), product of:
            0.12806706 = queryWeight, product of:
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.04800207 = queryNorm
            0.5002404 = fieldWeight in 932, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.09375 = fieldNorm(doc=932)
        0.002820454 = product of:
          0.008461362 = sum of:
            0.008461362 = weight(_text_:a in 932) [ClassicSimilarity], result of:
              0.008461362 = score(doc=932,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.15287387 = fieldWeight in 932, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=932)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Source
    Inhaltserschließung von Massendaten zur Wirksamkeit informationslinguistischer Verfahren am Beispiel des Deutschen Patentinformationssystems. Hrsg. J. Krause
    Type
    a
  2. Krause, J.: Was leisten informationslinguistische Komponenten von Referenz-Retrievalsystemen für Massendaten? : Von der 'Pragmatik im Computer' zur Pragmatikanalyse als Designgrundlage (1986) 0.03
    0.033442385 = product of:
      0.06688477 = sum of:
        0.06406432 = weight(_text_:von in 7395) [ClassicSimilarity], result of:
          0.06406432 = score(doc=7395,freq=4.0), product of:
            0.12806706 = queryWeight, product of:
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.04800207 = queryNorm
            0.5002404 = fieldWeight in 7395, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.09375 = fieldNorm(doc=7395)
        0.002820454 = product of:
          0.008461362 = sum of:
            0.008461362 = weight(_text_:a in 7395) [ClassicSimilarity], result of:
              0.008461362 = score(doc=7395,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.15287387 = fieldWeight in 7395, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=7395)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Type
    a
  3. Ruge, G.; Schwarz, C.: ¬Die Leistungsfähigkeit von linguistischen Verfahren in der Massentextverarbeitung (1989) 0.03
    0.0270943 = product of:
      0.0541886 = sum of:
        0.0523083 = weight(_text_:von in 6969) [ClassicSimilarity], result of:
          0.0523083 = score(doc=6969,freq=6.0), product of:
            0.12806706 = queryWeight, product of:
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.04800207 = queryNorm
            0.40844458 = fieldWeight in 6969, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.0625 = fieldNorm(doc=6969)
        0.0018803024 = product of:
          0.005640907 = sum of:
            0.005640907 = weight(_text_:a in 6969) [ClassicSimilarity], result of:
              0.005640907 = score(doc=6969,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.10191591 = fieldWeight in 6969, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6969)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    Dependenzstrukturen stellen syntagmatische Relationen von Worten in Texten dar. Ihre Anwendungsmöglichkeiten im Information Retrieval werden erläutert. Bei Siemens wurde ein System zur Transformation von Texten in Dependenzstrukturen entwickelt, wobei besonders darauf geachtet wurde, die Wirkung gegen den Aufwand abzuwiegen. Die letzte Version verarbeitet 20 MB Freitext in einer Stunde Realzeit auf einem Siemens BS2000 Großrechner. Analyse-Recall and Analyse-Precision liegen jeweils bei 0,85
    Type
    a
  4. Werner, H.: Indexierung auf linguistischer Grundlage am Beispiel von JUDO-DS(1) (1982) 0.02
    0.024060382 = product of:
      0.048120763 = sum of:
        0.04530031 = weight(_text_:von in 3017) [ClassicSimilarity], result of:
          0.04530031 = score(doc=3017,freq=2.0), product of:
            0.12806706 = queryWeight, product of:
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.04800207 = queryNorm
            0.35372335 = fieldWeight in 3017, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.09375 = fieldNorm(doc=3017)
        0.002820454 = product of:
          0.008461362 = sum of:
            0.008461362 = weight(_text_:a in 3017) [ClassicSimilarity], result of:
              0.008461362 = score(doc=3017,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.15287387 = fieldWeight in 3017, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3017)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Type
    a
  5. Bátori, I.: ¬Der sprachliche Verarbeitungsprozeß als paradigmatischer Kern der linguistischen Datenverarbeitung (1982) 0.02
    0.024060382 = product of:
      0.048120763 = sum of:
        0.04530031 = weight(_text_:von in 8422) [ClassicSimilarity], result of:
          0.04530031 = score(doc=8422,freq=2.0), product of:
            0.12806706 = queryWeight, product of:
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.04800207 = queryNorm
            0.35372335 = fieldWeight in 8422, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.09375 = fieldNorm(doc=8422)
        0.002820454 = product of:
          0.008461362 = sum of:
            0.008461362 = weight(_text_:a in 8422) [ClassicSimilarity], result of:
              0.008461362 = score(doc=8422,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.15287387 = fieldWeight in 8422, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=8422)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Source
    Linguistische Datenverarbeitung: Versuch einer Standortbestimmung im Umfeld von Informationslinguistik und Künstlicher Intelligenz. Hrsg.: I. Bátori u.a
    Type
    a
  6. Luckhardt, H.-D.; Schmitt, D.: Saarbrücker Traslationsservice STS : ein Konzept für die computergestützte Übersetzung von Datenbanken in der Fachinformation (1987) 0.02
    0.024060382 = product of:
      0.048120763 = sum of:
        0.04530031 = weight(_text_:von in 263) [ClassicSimilarity], result of:
          0.04530031 = score(doc=263,freq=2.0), product of:
            0.12806706 = queryWeight, product of:
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.04800207 = queryNorm
            0.35372335 = fieldWeight in 263, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.09375 = fieldNorm(doc=263)
        0.002820454 = product of:
          0.008461362 = sum of:
            0.008461362 = weight(_text_:a in 263) [ClassicSimilarity], result of:
              0.008461362 = score(doc=263,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.15287387 = fieldWeight in 263, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=263)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Type
    a
  7. Gerstenkorn, A.: Indexierung mit Nominalgruppen (1980) 0.02
    0.020537097 = product of:
      0.041074194 = sum of:
        0.03775026 = weight(_text_:von in 6685) [ClassicSimilarity], result of:
          0.03775026 = score(doc=6685,freq=2.0), product of:
            0.12806706 = queryWeight, product of:
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.04800207 = queryNorm
            0.29476947 = fieldWeight in 6685, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.078125 = fieldNorm(doc=6685)
        0.0033239368 = product of:
          0.0099718105 = sum of:
            0.0099718105 = weight(_text_:a in 6685) [ClassicSimilarity], result of:
              0.0099718105 = score(doc=6685,freq=4.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.18016359 = fieldWeight in 6685, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=6685)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    Die Indexierung mit Nominalgruppen ist eine konsequente Fortsetzung der Entwicklung von der gleichordnenden zur syntaktischen Indexierung. Nominalgruppen eignen sich besonders zur Bezeichnung komplexer Begriffe (Themen) und sind benutzerfreundlich. Bei einer automatischen Indexierung mit Nominalgruppen sind keine vollständigen Satzanalysen nötig, auch Systeme mit einem partiellen Parser liefern brauchbare Ergebnisse. Das Problem eines Retrieval mit Nominalgruppen ist noch zu lösen
    Type
    a
  8. Stock, M.: Textwortmethode und Übersetzungsrelation : Eine Methode zum Aufbau von kombinierten Literaturnachweis- und Terminologiedatenbanken (1989) 0.02
    0.020050319 = product of:
      0.040100638 = sum of:
        0.03775026 = weight(_text_:von in 3412) [ClassicSimilarity], result of:
          0.03775026 = score(doc=3412,freq=2.0), product of:
            0.12806706 = queryWeight, product of:
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.04800207 = queryNorm
            0.29476947 = fieldWeight in 3412, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.078125 = fieldNorm(doc=3412)
        0.002350378 = product of:
          0.007051134 = sum of:
            0.007051134 = weight(_text_:a in 3412) [ClassicSimilarity], result of:
              0.007051134 = score(doc=3412,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.12739488 = fieldWeight in 3412, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3412)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Type
    a
  9. Warner, A.J.: Natural language processing (1987) 0.02
    0.019223286 = product of:
      0.07689314 = sum of:
        0.07689314 = product of:
          0.11533971 = sum of:
            0.011281814 = weight(_text_:a in 337) [ClassicSimilarity], result of:
              0.011281814 = score(doc=337,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.20383182 = fieldWeight in 337, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=337)
            0.10405789 = weight(_text_:22 in 337) [ClassicSimilarity], result of:
              0.10405789 = score(doc=337,freq=2.0), product of:
                0.16809508 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04800207 = queryNorm
                0.61904186 = fieldWeight in 337, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=337)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Source
    Annual review of information science and technology. 22(1987), S.79-108
    Type
    a
  10. Zimmermann, H.H.: Linguistische Verfahren zur Archivierung und zum Wiederfinden unstrukturierter Texte (1983) 0.01
    0.014035223 = product of:
      0.028070446 = sum of:
        0.026425181 = weight(_text_:von in 557) [ClassicSimilarity], result of:
          0.026425181 = score(doc=557,freq=2.0), product of:
            0.12806706 = queryWeight, product of:
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.04800207 = queryNorm
            0.20633863 = fieldWeight in 557, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.6679487 = idf(docFreq=8340, maxDocs=44218)
              0.0546875 = fieldNorm(doc=557)
        0.0016452647 = product of:
          0.004935794 = sum of:
            0.004935794 = weight(_text_:a in 557) [ClassicSimilarity], result of:
              0.004935794 = score(doc=557,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.089176424 = fieldWeight in 557, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=557)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    Die technologischen Entwicklungen der 80er und 90er Jahre werden den Bedarf an intelligenten Verfahren in der automatischen Texterschließung und -archivierung sprunghaft steigen lassen. Inzwischen existieren eine Reihe linguistischer Verfahren, die auch große Datenmengen, wie sie in der Büro- und Fachkommunikation auftreten, zu bewältigen helfen. Dabei ist eine wesentliche Zielsetzung, den Anwender von 'technischen' Handhabungen, wie sie die herkömmlichen Informations-Retrieval-Systeme noch erfordern, schrittweise zu entlasten in Richtung auf einen ehe natürlichsprachigen Zugang. Während dabei in den nächsten Jahren 'verstehensorietierte' Ansätze nur in ausgewählten Bereichen zum Einsatz kommen können, werden Verfahren auf morphologisch-syntaktischer Basis die bisherigen oberflächenorientierten Systeme schrittweise ersetzen
    Type
    a
  11. Schwarz, C.: THESYS: Thesaurus Syntax System : a fully automatic thesaurus building aid (1988) 0.01
    0.009232819 = product of:
      0.036931276 = sum of:
        0.036931276 = product of:
          0.055396914 = sum of:
            0.009871588 = weight(_text_:a in 1361) [ClassicSimilarity], result of:
              0.009871588 = score(doc=1361,freq=8.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.17835285 = fieldWeight in 1361, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1361)
            0.045525327 = weight(_text_:22 in 1361) [ClassicSimilarity], result of:
              0.045525327 = score(doc=1361,freq=2.0), product of:
                0.16809508 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04800207 = queryNorm
                0.2708308 = fieldWeight in 1361, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1361)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Abstract
    THESYS is based on the natural language processing of free-text databases. It yields statistically evaluated correlations between words of the database. These correlations correspond to traditional thesaurus relations. The person who has to build a thesaurus is thus assisted by the proposals made by THESYS. THESYS is being tested on commercial databases under real world conditions. It is part of a text processing project at Siemens, called TINA (Text-Inhalts-Analyse). Software from TINA is actually being applied and evaluated by the US Department of Commerce for patent search and indexing (REALIST: REtrieval Aids by Linguistics and STatistics)
    Date
    6. 1.1999 10:22:07
    Type
    a
  12. Rau, L.F.; Jacobs, P.S.; Zernik, U.: Information extraction and text summarization using linguistic knowledge acquisition (1989) 0.00
    0.0010511212 = product of:
      0.0042044846 = sum of:
        0.0042044846 = product of:
          0.012613453 = sum of:
            0.012613453 = weight(_text_:a in 6683) [ClassicSimilarity], result of:
              0.012613453 = score(doc=6683,freq=10.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.22789092 = fieldWeight in 6683, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6683)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    Storing and accessing texts in a conceptual format has a number of advantages over traditional document retrieval methods. A conceptual format facilitates natural language access to text information. It can support imprecise and inexact queries, conceptual information summarisation, and, ultimately, document translation. Describes 2 methods which have been implemented in a prototype intelligent information retrieval system calles SCISOR (System for Conceptual Information Summarisation, Organization and Retrieval). Describes the text processing, language acquisition, and summarisation components of SCISOR
    Type
    a
  13. McCune, B.P.; Tong, R.M.; Dean, J.S.: Rubric: a system for rule-based information retrieval (1985) 0.00
    9.97181E-4 = product of:
      0.003988724 = sum of:
        0.003988724 = product of:
          0.011966172 = sum of:
            0.011966172 = weight(_text_:a in 1945) [ClassicSimilarity], result of:
              0.011966172 = score(doc=1945,freq=4.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.2161963 = fieldWeight in 1945, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1945)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Type
    a
  14. Schwarz, C.: Linguistische Hilfsmittel beim Information Retrieval (1984) 0.00
    9.401512E-4 = product of:
      0.003760605 = sum of:
        0.003760605 = product of:
          0.011281814 = sum of:
            0.011281814 = weight(_text_:a in 545) [ClassicSimilarity], result of:
              0.011281814 = score(doc=545,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.20383182 = fieldWeight in 545, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=545)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Type
    a
  15. Schwarz, C.: Freitextrecherche: Grenzen und Möglichkeiten (1982) 0.00
    9.401512E-4 = product of:
      0.003760605 = sum of:
        0.003760605 = product of:
          0.011281814 = sum of:
            0.011281814 = weight(_text_:a in 1349) [ClassicSimilarity], result of:
              0.011281814 = score(doc=1349,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.20383182 = fieldWeight in 1349, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=1349)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Type
    a
  16. Endres-Niggemeyer, B.: Sprachverarbeitung im Informationsbereich (1989) 0.00
    9.401512E-4 = product of:
      0.003760605 = sum of:
        0.003760605 = product of:
          0.011281814 = sum of:
            0.011281814 = weight(_text_:a in 4860) [ClassicSimilarity], result of:
              0.011281814 = score(doc=4860,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.20383182 = fieldWeight in 4860, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=4860)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Type
    a
  17. Noel, J.: Syntax, semantics and pragmatics in the automatic analysis of texts (1980) 0.00
    9.401512E-4 = product of:
      0.003760605 = sum of:
        0.003760605 = product of:
          0.011281814 = sum of:
            0.011281814 = weight(_text_:a in 7512) [ClassicSimilarity], result of:
              0.011281814 = score(doc=7512,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.20383182 = fieldWeight in 7512, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=7512)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Type
    a
  18. Schwarz, C.: Probleme der syntaktischen Indexierung (1986) 0.00
    9.401512E-4 = product of:
      0.003760605 = sum of:
        0.003760605 = product of:
          0.011281814 = sum of:
            0.011281814 = weight(_text_:a in 8180) [ClassicSimilarity], result of:
              0.011281814 = score(doc=8180,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.20383182 = fieldWeight in 8180, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=8180)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Type
    a
  19. Becker, D.: Automated language processing (1981) 0.00
    9.401512E-4 = product of:
      0.003760605 = sum of:
        0.003760605 = product of:
          0.011281814 = sum of:
            0.011281814 = weight(_text_:a in 287) [ClassicSimilarity], result of:
              0.011281814 = score(doc=287,freq=2.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.20383182 = fieldWeight in 287, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=287)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Type
    a
  20. Salton, G.: Automatic processing of foreign language documents (1985) 0.00
    8.474409E-4 = product of:
      0.0033897636 = sum of:
        0.0033897636 = product of:
          0.010169291 = sum of:
            0.010169291 = weight(_text_:a in 3650) [ClassicSimilarity], result of:
              0.010169291 = score(doc=3650,freq=26.0), product of:
                0.055348642 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04800207 = queryNorm
                0.18373153 = fieldWeight in 3650, product of:
                  5.0990195 = tf(freq=26.0), with freq of:
                    26.0 = termFreq=26.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3650)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    The attempt to computerize a process, such as indexing, abstracting, classifying, or retrieving information, begins with an analysis of the process into its intellectual and nonintellectual components. That part of the process which is amenable to computerization is mechanical or algorithmic. What is not is intellectual or creative and requires human intervention. Gerard Salton has been an innovator, experimenter, and promoter in the area of mechanized information systems since the early 1960s. He has been particularly ingenious at analyzing the process of information retrieval into its algorithmic components. He received a doctorate in applied mathematics from Harvard University before moving to the computer science department at Cornell, where he developed a prototype automatic retrieval system called SMART. Working with this system he and his students contributed for over a decade to our theoretical understanding of the retrieval process. On a more practical level, they have contributed design criteria for operating retrieval systems. The following selection presents one of the early descriptions of the SMART system; it is valuable as it shows the direction automatic retrieval methods were to take beyond simple word-matching techniques. These include various word normalization techniques to improve recall, for instance, the separation of words into stems and affixes; the correlation and clustering, using statistical association measures, of related terms; and the identification, using a concept thesaurus, of synonymous, broader, narrower, and sibling terms. They include, as weIl, techniques, both linguistic and statistical, to deal with the thorny problem of how to automatically extract from texts index terms that consist of more than one word. They include weighting techniques and various documentrequest matching algorithms. Significant among the latter are those which produce a retrieval output of citations ranked in relevante order. During the 1970s, Salton and his students went an to further refine these various techniques, particularly the weighting and statistical association measures. Many of their early innovations seem commonplace today. Some of their later techniques are still ahead of their time and await technological developments for implementation. The particular focus of the selection that follows is an the evaluation of a particular component of the SMART system, a multilingual thesaurus. By mapping English language expressions and their German equivalents to a common concept number, the thesaurus permitted the automatic processing of German language documents against English language queries and vice versa. The results of the evaluation, as it turned out, were somewhat inconclusive. However, this SMART experiment suggested in a bold and optimistic way how one might proceed to answer such complex questions as What is meant by retrieval language compatability? How it is to be achieved, and how evaluated?
    Source
    Theory of subject analysis: a sourcebook. Ed.: L.M. Chan, et al
    Type
    a