Search (14 results, page 1 of 1)

  • × theme_ss:"Automatisches Indexieren"
  • × theme_ss:"Computerlinguistik"
  • × type_ss:"a"
  1. Stock, M.; Stock, W.G.: Literaturnachweis- und Terminologiedatenbank : die Erfassung von Fachliteratur und Fachterminologie eines Fachgebiets in einer kombinierten Datenbank (1991) 0.06
    0.056307487 = product of:
      0.22522995 = sum of:
        0.1341764 = weight(_text_:bibliographien in 3411) [ClassicSimilarity], result of:
          0.1341764 = score(doc=3411,freq=2.0), product of:
            0.17148633 = queryWeight, product of:
              7.0817666 = idf(docFreq=100, maxDocs=44218)
              0.024215192 = queryNorm
            0.782432 = fieldWeight in 3411, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0817666 = idf(docFreq=100, maxDocs=44218)
              0.078125 = fieldNorm(doc=3411)
        0.02276339 = weight(_text_:und in 3411) [ClassicSimilarity], result of:
          0.02276339 = score(doc=3411,freq=6.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.42413816 = fieldWeight in 3411, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3411)
        0.02276339 = weight(_text_:und in 3411) [ClassicSimilarity], result of:
          0.02276339 = score(doc=3411,freq=6.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.42413816 = fieldWeight in 3411, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3411)
        0.02276339 = weight(_text_:und in 3411) [ClassicSimilarity], result of:
          0.02276339 = score(doc=3411,freq=6.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.42413816 = fieldWeight in 3411, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3411)
        0.02276339 = weight(_text_:und in 3411) [ClassicSimilarity], result of:
          0.02276339 = score(doc=3411,freq=6.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.42413816 = fieldWeight in 3411, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3411)
      0.25 = coord(5/20)
    
    Abstract
    In wissenschaftlichen Spezialgebieten kann über den Aufbau einer Literaturdatenbank gleichzeitig eine Terminologiedatenbank mit erstellt werden. Als Dokumentationsmethode eignet sich die Textwortmethode mit Übersetzungrelation. Mit dem Softwarepaket LBase aufgebaute Druckbildprogramme gestatten die Ausgabe von Bibliographien und Wörterbüchern
  2. Gombocz, W.L.: Stichwort oder Schlagwort versus Textwort : Grazer und Düsseldorfer Philosophie-Dokumentation und -Information nach bzw. gemäß Norbert Henrichs (2000) 0.05
    0.051478792 = product of:
      0.17159596 = sum of:
        0.026284898 = weight(_text_:und in 3413) [ClassicSimilarity], result of:
          0.026284898 = score(doc=3413,freq=8.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.48975256 = fieldWeight in 3413, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3413)
        0.026284898 = weight(_text_:und in 3413) [ClassicSimilarity], result of:
          0.026284898 = score(doc=3413,freq=8.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.48975256 = fieldWeight in 3413, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3413)
        0.013349611 = weight(_text_:der in 3413) [ClassicSimilarity], result of:
          0.013349611 = score(doc=3413,freq=2.0), product of:
            0.054091092 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.024215192 = queryNorm
            0.2467987 = fieldWeight in 3413, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.078125 = fieldNorm(doc=3413)
        0.026284898 = weight(_text_:und in 3413) [ClassicSimilarity], result of:
          0.026284898 = score(doc=3413,freq=8.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.48975256 = fieldWeight in 3413, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3413)
        0.053106755 = product of:
          0.10621351 = sum of:
            0.10621351 = weight(_text_:philosophie in 3413) [ClassicSimilarity], result of:
              0.10621351 = score(doc=3413,freq=4.0), product of:
                0.12829916 = queryWeight, product of:
                  5.298292 = idf(docFreq=600, maxDocs=44218)
                  0.024215192 = queryNorm
                0.82785815 = fieldWeight in 3413, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.298292 = idf(docFreq=600, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3413)
          0.5 = coord(1/2)
        0.026284898 = weight(_text_:und in 3413) [ClassicSimilarity], result of:
          0.026284898 = score(doc=3413,freq=8.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.48975256 = fieldWeight in 3413, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3413)
      0.3 = coord(6/20)
    
    Field
    Philosophie
    Imprint
    Düsseldorf : Universitäts- und Landesbibliothek
    Series
    Schriften der Universitäts- und Landesbibliothek Düsseldorf; 32
  3. Stock, W.G.: Textwortmethode : Norbert Henrichs zum 65. (3) (2000) 0.04
    0.043075215 = product of:
      0.14358404 = sum of:
        0.018210711 = weight(_text_:und in 4891) [ClassicSimilarity], result of:
          0.018210711 = score(doc=4891,freq=6.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.33931053 = fieldWeight in 4891, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=4891)
        0.018210711 = weight(_text_:und in 4891) [ClassicSimilarity], result of:
          0.018210711 = score(doc=4891,freq=6.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.33931053 = fieldWeight in 4891, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=4891)
        0.028255802 = weight(_text_:der in 4891) [ClassicSimilarity], result of:
          0.028255802 = score(doc=4891,freq=14.0), product of:
            0.054091092 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.024215192 = queryNorm
            0.5223744 = fieldWeight in 4891, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0625 = fieldNorm(doc=4891)
        0.018210711 = weight(_text_:und in 4891) [ClassicSimilarity], result of:
          0.018210711 = score(doc=4891,freq=6.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.33931053 = fieldWeight in 4891, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=4891)
        0.0424854 = product of:
          0.0849708 = sum of:
            0.0849708 = weight(_text_:philosophie in 4891) [ClassicSimilarity], result of:
              0.0849708 = score(doc=4891,freq=4.0), product of:
                0.12829916 = queryWeight, product of:
                  5.298292 = idf(docFreq=600, maxDocs=44218)
                  0.024215192 = queryNorm
                0.6622865 = fieldWeight in 4891, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.298292 = idf(docFreq=600, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4891)
          0.5 = coord(1/2)
        0.018210711 = weight(_text_:und in 4891) [ClassicSimilarity], result of:
          0.018210711 = score(doc=4891,freq=6.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.33931053 = fieldWeight in 4891, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=4891)
      0.3 = coord(6/20)
    
    Abstract
    Nur wenige Dokumentationsmethoden werden mit dem Namen ihrer Entwickler assoziiert. Ausnahmen sind Melvil Dewey (DDC), S.R. Ranganathan (Colon Classification) - und Norbert Henrichs. Seine Textwortmethode ermöglicht die Indexierung und das Retrieval von Literatur aus Fachgebieten, die keine allseits akzeptierte Fachterminologie vorweisen, also viele Sozial- und Geisteswissenschaften, vorneweg die Philosophie. Für den Einsatz in der elektronischen Philosophie-Dokumentation hat Henrichs in den späten sechziger Jahren die Textwortmethode entworfen. Er ist damit nicht nur einer der Pioniere der Anwendung der elektronischen Datenverarbeitung in der Informationspraxis, sondern auch der Pionier bei der Dokumentation terminologisch nicht starrer Fachsprachen
  4. Stock, W.G.: Textwortmethode (2000) 0.04
    0.037574004 = product of:
      0.12524667 = sum of:
        0.018586228 = weight(_text_:und in 3408) [ClassicSimilarity], result of:
          0.018586228 = score(doc=3408,freq=4.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34630734 = fieldWeight in 3408, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3408)
        0.018586228 = weight(_text_:und in 3408) [ClassicSimilarity], result of:
          0.018586228 = score(doc=3408,freq=4.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34630734 = fieldWeight in 3408, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3408)
        0.013349611 = weight(_text_:der in 3408) [ClassicSimilarity], result of:
          0.013349611 = score(doc=3408,freq=2.0), product of:
            0.054091092 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.024215192 = queryNorm
            0.2467987 = fieldWeight in 3408, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.078125 = fieldNorm(doc=3408)
        0.018586228 = weight(_text_:und in 3408) [ClassicSimilarity], result of:
          0.018586228 = score(doc=3408,freq=4.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34630734 = fieldWeight in 3408, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3408)
        0.037552148 = product of:
          0.075104296 = sum of:
            0.075104296 = weight(_text_:philosophie in 3408) [ClassicSimilarity], result of:
              0.075104296 = score(doc=3408,freq=2.0), product of:
                0.12829916 = queryWeight, product of:
                  5.298292 = idf(docFreq=600, maxDocs=44218)
                  0.024215192 = queryNorm
                0.58538413 = fieldWeight in 3408, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.298292 = idf(docFreq=600, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3408)
          0.5 = coord(1/2)
        0.018586228 = weight(_text_:und in 3408) [ClassicSimilarity], result of:
          0.018586228 = score(doc=3408,freq=4.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34630734 = fieldWeight in 3408, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3408)
      0.3 = coord(6/20)
    
    Field
    Philosophie
    Imprint
    Düsseldorf : Universitäts- und Landesbibliothek
    Series
    Schriften der Universitäts- und Landesbibliothek Düsseldorf; 32
  5. Stock, M.: Textwortmethode und Übersetzungsrelation : Eine Methode zum Aufbau von kombinierten Literaturnachweis- und Terminologiedatenbanken (1989) 0.03
    0.03272481 = product of:
      0.13089924 = sum of:
        0.029387407 = weight(_text_:und in 3412) [ClassicSimilarity], result of:
          0.029387407 = score(doc=3412,freq=10.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.54756 = fieldWeight in 3412, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3412)
        0.029387407 = weight(_text_:und in 3412) [ClassicSimilarity], result of:
          0.029387407 = score(doc=3412,freq=10.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.54756 = fieldWeight in 3412, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3412)
        0.013349611 = weight(_text_:der in 3412) [ClassicSimilarity], result of:
          0.013349611 = score(doc=3412,freq=2.0), product of:
            0.054091092 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.024215192 = queryNorm
            0.2467987 = fieldWeight in 3412, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.078125 = fieldNorm(doc=3412)
        0.029387407 = weight(_text_:und in 3412) [ClassicSimilarity], result of:
          0.029387407 = score(doc=3412,freq=10.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.54756 = fieldWeight in 3412, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3412)
        0.029387407 = weight(_text_:und in 3412) [ClassicSimilarity], result of:
          0.029387407 = score(doc=3412,freq=10.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.54756 = fieldWeight in 3412, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=3412)
      0.25 = coord(5/20)
    
    Abstract
    Geisteswissenschaftliche Fachinformation erfordert eine enge Kooperation zwischen Literaturnachweis- und Terminologieinformationssystemen. Eine geeignete Dokumentationsmethode für die Auswertung geisteswissen- schaftlicher Literatur ist die Textwortwethode. Dem originalsprachig aufgenommenen Begriffsrepertoire ist ein einheitssprachiger Zugriff beizuordnen, der einerseits ein vollständiges und genaues Retrieval garantiert und andererseits den Aufbau fachspezifischer Wörterbücher vorantreibt
  6. Volk, M.; Mittermaier, H.; Schurig, A.; Biedassek, T.: Halbautomatische Volltextanalyse, Datenbankaufbau und Document Retrieval (1992) 0.03
    0.027644014 = product of:
      0.110576056 = sum of:
        0.024340155 = weight(_text_:und in 2571) [ClassicSimilarity], result of:
          0.024340155 = score(doc=2571,freq=14.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.4535172 = fieldWeight in 2571, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2571)
        0.024340155 = weight(_text_:und in 2571) [ClassicSimilarity], result of:
          0.024340155 = score(doc=2571,freq=14.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.4535172 = fieldWeight in 2571, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2571)
        0.013215441 = weight(_text_:der in 2571) [ClassicSimilarity], result of:
          0.013215441 = score(doc=2571,freq=4.0), product of:
            0.054091092 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.024215192 = queryNorm
            0.24431825 = fieldWeight in 2571, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2571)
        0.024340155 = weight(_text_:und in 2571) [ClassicSimilarity], result of:
          0.024340155 = score(doc=2571,freq=14.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.4535172 = fieldWeight in 2571, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2571)
        0.024340155 = weight(_text_:und in 2571) [ClassicSimilarity], result of:
          0.024340155 = score(doc=2571,freq=14.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.4535172 = fieldWeight in 2571, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2571)
      0.25 = coord(5/20)
    
    Abstract
    In diesem Aufsatz beschreiben wir ein System zur Analyse von Kurzartikeln. Das System arbeitet halbautomatisch. Das heißt, zunächst wird der Artikel vom System analysiert und dann dem benutzer zur Nachberarbeitung vorgelegt. Die so gewonnene Information wird in einem Datenbankeintrag abgelegt. Über die Datenbank - in dBase IV implementiert - sind dann Abfragen und Zugriffe auf die Originaltexte effizient möglich. Der Kern dieses Aufsatzes betrifft die halbautomatische Analyse. Wir beschreiben unser Verfahren für parametrisiertes Pattern Matching sowie linguistische Heuristiken zur Ermittlung von Nominalphrasen und Präpositionalphrasen. Das System wurde für den praktischen Einsatz im Bonner Büro des 'Forums InformatikerInnen Für Frieden und gesellschaftliche Verantwortung e.V. (FIFF)' entwickelt
    Source
    Datenanalyse, Klassifikation und Informationsverarbeitung: Methoden und Anwendungen in verschiedenen Fachgebieten. Hrsg.: H. Goebl u. M. Schader
  7. Zimmermann, H.H.: Wortrelationierung in der Sprachtechnik : Stilhilfen, Retrievalhilfen, Übersetzungshilfen (1992) 0.02
    0.02472613 = product of:
      0.09890452 = sum of:
        0.015770938 = weight(_text_:und in 1372) [ClassicSimilarity], result of:
          0.015770938 = score(doc=1372,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29385152 = fieldWeight in 1372, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.09375 = fieldNorm(doc=1372)
        0.015770938 = weight(_text_:und in 1372) [ClassicSimilarity], result of:
          0.015770938 = score(doc=1372,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29385152 = fieldWeight in 1372, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.09375 = fieldNorm(doc=1372)
        0.035820764 = weight(_text_:der in 1372) [ClassicSimilarity], result of:
          0.035820764 = score(doc=1372,freq=10.0), product of:
            0.054091092 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.024215192 = queryNorm
            0.6622304 = fieldWeight in 1372, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.09375 = fieldNorm(doc=1372)
        0.015770938 = weight(_text_:und in 1372) [ClassicSimilarity], result of:
          0.015770938 = score(doc=1372,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29385152 = fieldWeight in 1372, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.09375 = fieldNorm(doc=1372)
        0.015770938 = weight(_text_:und in 1372) [ClassicSimilarity], result of:
          0.015770938 = score(doc=1372,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29385152 = fieldWeight in 1372, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.09375 = fieldNorm(doc=1372)
      0.25 = coord(5/20)
    
    Series
    Fortschritte in der Wissensorganisation; Bd.2
    Source
    Kognitive Ansätze zum Ordnen und Darstellen von Wissen. 2. Tagung der Deutschen ISKO Sektion einschl. der Vorträge des Workshops "Thesauri als Werkzeuge der Sprachtechnologie", Weilburg, 15.-18.10.1991
  8. Rapke, K.: Automatische Indexierung von Volltexten für die Gruner+Jahr Pressedatenbank (2001) 0.02
    0.024120683 = product of:
      0.09648273 = sum of:
        0.018586228 = weight(_text_:und in 5863) [ClassicSimilarity], result of:
          0.018586228 = score(doc=5863,freq=16.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34630734 = fieldWeight in 5863, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5863)
        0.018586228 = weight(_text_:und in 5863) [ClassicSimilarity], result of:
          0.018586228 = score(doc=5863,freq=16.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34630734 = fieldWeight in 5863, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5863)
        0.022137824 = weight(_text_:der in 5863) [ClassicSimilarity], result of:
          0.022137824 = score(doc=5863,freq=22.0), product of:
            0.054091092 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.024215192 = queryNorm
            0.40926933 = fieldWeight in 5863, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5863)
        0.018586228 = weight(_text_:und in 5863) [ClassicSimilarity], result of:
          0.018586228 = score(doc=5863,freq=16.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34630734 = fieldWeight in 5863, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5863)
        0.018586228 = weight(_text_:und in 5863) [ClassicSimilarity], result of:
          0.018586228 = score(doc=5863,freq=16.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34630734 = fieldWeight in 5863, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5863)
      0.25 = coord(5/20)
    
    Abstract
    Retrievaltests sind die anerkannteste Methode, um neue Verfahren der Inhaltserschließung gegenüber traditionellen Verfahren zu rechtfertigen. Im Rahmen einer Diplomarbeit wurden zwei grundsätzlich unterschiedliche Systeme der automatischen inhaltlichen Erschließung anhand der Pressedatenbank des Verlagshauses Gruner + Jahr (G+J) getestet und evaluiert. Untersucht wurde dabei natürlichsprachliches Retrieval im Vergleich zu Booleschem Retrieval. Bei den beiden Systemen handelt es sich zum einen um Autonomy von Autonomy Inc. und DocCat, das von IBM an die Datenbankstruktur der G+J Pressedatenbank angepasst wurde. Ersteres ist ein auf natürlichsprachlichem Retrieval basierendes, probabilistisches System. DocCat demgegenüber basiert auf Booleschem Retrieval und ist ein lernendes System, das aufgrund einer intellektuell erstellten Trainingsvorlage indexiert. Methodisch geht die Evaluation vom realen Anwendungskontext der Textdokumentation von G+J aus. Die Tests werden sowohl unter statistischen wie auch qualitativen Gesichtspunkten bewertet. Ein Ergebnis der Tests ist, dass DocCat einige Mängel gegenüber der intellektuellen Inhaltserschließung aufweist, die noch behoben werden müssen, während das natürlichsprachliche Retrieval von Autonomy in diesem Rahmen und für die speziellen Anforderungen der G+J Textdokumentation so nicht einsetzbar ist
    Series
    Tagungen der Deutschen Gesellschaft für Informationswissenschaft und Informationspraxis; 4
    Source
    Information Research & Content Management: Orientierung, Ordnung und Organisation im Wissensmarkt; 23. DGI-Online-Tagung der DGI und 53. Jahrestagung der Deutschen Gesellschaft für Informationswissenschaft und Informationspraxis e.V. DGI, Frankfurt am Main, 8.-10.5.2001. Proceedings. Hrsg.: R. Schmidt
  9. Rapke, K.: Automatische Indexierung von Volltexten für die Gruner+Jahr Pressedatenbank (2001) 0.02
    0.023296207 = product of:
      0.09318483 = sum of:
        0.017632445 = weight(_text_:und in 6386) [ClassicSimilarity], result of:
          0.017632445 = score(doc=6386,freq=10.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.328536 = fieldWeight in 6386, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=6386)
        0.017632445 = weight(_text_:und in 6386) [ClassicSimilarity], result of:
          0.017632445 = score(doc=6386,freq=10.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.328536 = fieldWeight in 6386, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=6386)
        0.022655042 = weight(_text_:der in 6386) [ClassicSimilarity], result of:
          0.022655042 = score(doc=6386,freq=16.0), product of:
            0.054091092 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.024215192 = queryNorm
            0.4188313 = fieldWeight in 6386, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.046875 = fieldNorm(doc=6386)
        0.017632445 = weight(_text_:und in 6386) [ClassicSimilarity], result of:
          0.017632445 = score(doc=6386,freq=10.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.328536 = fieldWeight in 6386, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=6386)
        0.017632445 = weight(_text_:und in 6386) [ClassicSimilarity], result of:
          0.017632445 = score(doc=6386,freq=10.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.328536 = fieldWeight in 6386, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=6386)
      0.25 = coord(5/20)
    
    Abstract
    Retrieval Tests sind die anerkannteste Methode, um neue Verfahren der Inhaltserschließung gegenüber traditionellen Verfahren zu rechtfertigen. Im Rahmen einer Diplomarbeit wurden zwei grundsätzlich unterschiedliche Systeme der automatischen inhaltlichen Erschließung anhand der Pressedatenbank des Verlagshauses Gruner + Jahr (G+J) getestet und evaluiert. Untersucht wurde dabei natürlichsprachliches Retrieval im Vergleich zu Booleschem Retrieval. Bei den beiden Systemen handelt es sich zum einen um Autonomy von Autonomy Inc. und DocCat, das von IBM an die Datenbankstruktur der G+J Pressedatenbank angepasst wurde. Ersteres ist ein auf natürlichsprachlichem Retrieval basierendes, probabilistisches System. DocCat demgegenüber basiert auf Booleschem Retrieval und ist ein lernendes System, das auf Grund einer intellektuell erstellten Trainingsvorlage indexiert. Methodisch geht die Evaluation vom realen Anwendungskontext der Textdokumentation von G+J aus. Die Tests werden sowohl unter statistischen wie auch qualitativen Gesichtspunkten bewertet. Ein Ergebnis der Tests ist, dass DocCat einige Mängel gegenüber der intellektuellen Inhaltserschließung aufweist, die noch behoben werden müssen, während das natürlichsprachliche Retrieval von Autonomy in diesem Rahmen und für die speziellen Anforderungen der G+J Textdokumentation so nicht einsetzbar ist
    Source
    nfd Information - Wissenschaft und Praxis. 52(2001) H.5, S.251-262
  10. Larroche-Boutet, V.; Pöhl, K.: ¬Das Nominalsyntagna : über die Nutzbarmachung eines logico-semantischen Konzeptes für dokumentarische Fragestellungen (1993) 0.02
    0.020511115 = product of:
      0.08204446 = sum of:
        0.016096147 = weight(_text_:und in 5282) [ClassicSimilarity], result of:
          0.016096147 = score(doc=5282,freq=12.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29991096 = fieldWeight in 5282, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5282)
        0.016096147 = weight(_text_:und in 5282) [ClassicSimilarity], result of:
          0.016096147 = score(doc=5282,freq=12.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29991096 = fieldWeight in 5282, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5282)
        0.017659876 = weight(_text_:der in 5282) [ClassicSimilarity], result of:
          0.017659876 = score(doc=5282,freq=14.0), product of:
            0.054091092 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.024215192 = queryNorm
            0.326484 = fieldWeight in 5282, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5282)
        0.016096147 = weight(_text_:und in 5282) [ClassicSimilarity], result of:
          0.016096147 = score(doc=5282,freq=12.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29991096 = fieldWeight in 5282, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5282)
        0.016096147 = weight(_text_:und in 5282) [ClassicSimilarity], result of:
          0.016096147 = score(doc=5282,freq=12.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29991096 = fieldWeight in 5282, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5282)
      0.25 = coord(5/20)
    
    Abstract
    Am Anfang nachfolgender Ausführungen werden die für die Indexierung großer textmengen notwendigen strategischen Entscheidungen aufgezeigt: es müssen sowohl das Indexierungsverfahren (menschliche oder automatische Indexierung) als auch die Indexierungssparche (freie, kontrollierte oder natürliche Sprache) ausgewählt werden. Hierbei hat sich die Forschungsgruppe SYDO-LYON für natürlichsprachige automatische Vollindexierung entschieden. Auf der Grundlage der Unterscheidung zwischen prädikativen und referentiellen Textteilen wird d as Nominalsyntagma als kleinste referentielle Texteinheit definiert, dann das für die Konstituierung eines Nominalsyntagmas entscheidende Phänomen der Aktualisierung erläutert und schließlich auf die morphologischen Mittel zur Erkennung des Nominalsyntagmas hingewiesen. Alle Nominalsyntagma eines Textes werden als dessen potentielle Deskriptoren extrahiert, und Hilfsmittel für die Benutzer einer mit diesem Indexierungsverfahren arbeitenden Datenbank werden vorgestellt. Außerdem wird der begriff der Anapher (d.h. die Wiederaufnahme von Nominalsyntagmen durch Pronomen) kurz definiert, ihre Anwendung als Mittel zur Gewichtung des Deskriptorterme (durch Zählung ihrer Häufigkeit im text) aufgezeigt und morphologische uns syntaktische Regeln zur automatischen Bestimmung des von einem anaphorischen Pronomen aufgenommenen Nominalsyntagmas aufgestellt. Bevor abschließend Ziele und Grenzen der Arbeit diskutiert werden, wird noch auf einen Unterschied zwischen Nominalsyntagma und Deskriptorterm hingewiesen: das Nonimalsyntagma verweist auf ein Objekt, das ein Einzelobjekt oder eine Klasse sein kann, der Deskriptorterm verweist immer auf eine Klasse
  11. Lustig, G.: ¬Das Projekt WAI : Wörterbuchentwicklung für automatisches Indexing (1982) 0.01
    0.014719543 = product of:
      0.073597714 = sum of:
        0.018399429 = weight(_text_:und in 33) [ClassicSimilarity], result of:
          0.018399429 = score(doc=33,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34282678 = fieldWeight in 33, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.109375 = fieldNorm(doc=33)
        0.018399429 = weight(_text_:und in 33) [ClassicSimilarity], result of:
          0.018399429 = score(doc=33,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34282678 = fieldWeight in 33, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.109375 = fieldNorm(doc=33)
        0.018399429 = weight(_text_:und in 33) [ClassicSimilarity], result of:
          0.018399429 = score(doc=33,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34282678 = fieldWeight in 33, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.109375 = fieldNorm(doc=33)
        0.018399429 = weight(_text_:und in 33) [ClassicSimilarity], result of:
          0.018399429 = score(doc=33,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.34282678 = fieldWeight in 33, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.109375 = fieldNorm(doc=33)
      0.2 = coord(4/20)
    
    Source
    Deutscher Dokumentartag 1981, Mainz, 5.-8.10.1981: Kleincomputer in Information und Dokumentation. Bearb.: H. Strohl-Goebel
  12. Werner, H.: Indexierung auf linguistischer Grundlage am Beispiel von JUDO-DS(1) (1982) 0.01
    0.012616751 = product of:
      0.06308375 = sum of:
        0.015770938 = weight(_text_:und in 3017) [ClassicSimilarity], result of:
          0.015770938 = score(doc=3017,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29385152 = fieldWeight in 3017, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.09375 = fieldNorm(doc=3017)
        0.015770938 = weight(_text_:und in 3017) [ClassicSimilarity], result of:
          0.015770938 = score(doc=3017,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29385152 = fieldWeight in 3017, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.09375 = fieldNorm(doc=3017)
        0.015770938 = weight(_text_:und in 3017) [ClassicSimilarity], result of:
          0.015770938 = score(doc=3017,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29385152 = fieldWeight in 3017, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.09375 = fieldNorm(doc=3017)
        0.015770938 = weight(_text_:und in 3017) [ClassicSimilarity], result of:
          0.015770938 = score(doc=3017,freq=2.0), product of:
            0.05366975 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.024215192 = queryNorm
            0.29385152 = fieldWeight in 3017, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.09375 = fieldNorm(doc=3017)
      0.2 = coord(4/20)
    
    Source
    Deutscher Dokumentartag 1981, Mainz, 5.-8.10.1981: Kleincomputer in Information und Dokumentation. Bearb.: H. Strohl-Goebel
  13. Riloff, E.: ¬An empirical study of automated dictionary construction for information extraction in three domains (1996) 0.00
    6.5616483E-4 = product of:
      0.013123296 = sum of:
        0.013123296 = product of:
          0.026246592 = sum of:
            0.026246592 = weight(_text_:22 in 6752) [ClassicSimilarity], result of:
              0.026246592 = score(doc=6752,freq=2.0), product of:
                0.08479747 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.024215192 = queryNorm
                0.30952093 = fieldWeight in 6752, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6752)
          0.5 = coord(1/2)
      0.05 = coord(1/20)
    
    Date
    6. 3.1997 16:22:15
  14. Goller, C.; Löning, J.; Will, T.; Wolff, W.: Automatic document classification : a thourough evaluation of various methods (2000) 0.00
    4.0048832E-4 = product of:
      0.008009766 = sum of:
        0.008009766 = weight(_text_:der in 5480) [ClassicSimilarity], result of:
          0.008009766 = score(doc=5480,freq=2.0), product of:
            0.054091092 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.024215192 = queryNorm
            0.14807922 = fieldWeight in 5480, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.046875 = fieldNorm(doc=5480)
      0.05 = coord(1/20)
    
    Source
    Informationskompetenz - Basiskompetenz in der Informationsgesellschaft: Proceedings des 7. Internationalen Symposiums für Informationswissenschaft (ISI 2000), Hrsg.: G. Knorz u. R. Kuhlen