Search (266 results, page 1 of 14)

  • × theme_ss:"Automatisches Indexieren"
  1. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.16
    0.1631313 = product of:
      0.24469694 = sum of:
        0.02823696 = weight(_text_:information in 402) [ClassicSimilarity], result of:
          0.02823696 = score(doc=402,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.3103276 = fieldWeight in 402, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=402)
        0.21645999 = sum of:
          0.10409857 = weight(_text_:management in 402) [ClassicSimilarity], result of:
            0.10409857 = score(doc=402,freq=2.0), product of:
              0.17470726 = queryWeight, product of:
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.0518325 = queryNorm
              0.5958457 = fieldWeight in 402, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.125 = fieldNorm(doc=402)
          0.112361416 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
            0.112361416 = score(doc=402,freq=2.0), product of:
              0.18150859 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0518325 = queryNorm
              0.61904186 = fieldWeight in 402, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.125 = fieldNorm(doc=402)
      0.6666667 = coord(2/3)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  2. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.05
    0.054114997 = product of:
      0.16234499 = sum of:
        0.16234499 = sum of:
          0.078073926 = weight(_text_:management in 58) [ClassicSimilarity], result of:
            0.078073926 = score(doc=58,freq=2.0), product of:
              0.17470726 = queryWeight, product of:
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.0518325 = queryNorm
              0.44688427 = fieldWeight in 58, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.09375 = fieldNorm(doc=58)
          0.08427106 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
            0.08427106 = score(doc=58,freq=2.0), product of:
              0.18150859 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0518325 = queryNorm
              0.46428138 = fieldWeight in 58, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.09375 = fieldNorm(doc=58)
      0.33333334 = coord(1/3)
    
    Date
    14. 6.2015 22:12:44
    Source
    Deutscher Dokumentartag 1985, Nürnberg, 1.-4.10.1985: Fachinformation: Methodik - Management - Markt; neue Entwicklungen, Berufe, Produkte. Bearb.: H. Strohl-Goebel
  3. Thiel, T.J.: Automated indexing of information stored on optical disk electronic document image management systems (1994) 0.05
    0.05365639 = product of:
      0.080484584 = sum of:
        0.034941453 = weight(_text_:information in 1260) [ClassicSimilarity], result of:
          0.034941453 = score(doc=1260,freq=4.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.3840108 = fieldWeight in 1260, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=1260)
        0.045543127 = product of:
          0.09108625 = sum of:
            0.09108625 = weight(_text_:management in 1260) [ClassicSimilarity], result of:
              0.09108625 = score(doc=1260,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.521365 = fieldWeight in 1260, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1260)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Encyclopedia of library and information science. Vol.54, [=Suppl.17]
  4. Willett, P.: Recent trends in hierarchic document clustering : a critical review (1988) 0.05
    0.053524166 = product of:
      0.08028625 = sum of:
        0.02823696 = weight(_text_:information in 2604) [ClassicSimilarity], result of:
          0.02823696 = score(doc=2604,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.3103276 = fieldWeight in 2604, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=2604)
        0.052049287 = product of:
          0.10409857 = sum of:
            0.10409857 = weight(_text_:management in 2604) [ClassicSimilarity], result of:
              0.10409857 = score(doc=2604,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.5958457 = fieldWeight in 2604, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.125 = fieldNorm(doc=2604)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Information processing and management. 24(1988) no.5, S.577-597
  5. Hlava, M.M.K.: Automatic indexing : comparing rule-based and statistics-based indexing systems (2005) 0.05
    0.04924364 = product of:
      0.07386546 = sum of:
        0.02470734 = weight(_text_:information in 6265) [ClassicSimilarity], result of:
          0.02470734 = score(doc=6265,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.27153665 = fieldWeight in 6265, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=6265)
        0.04915812 = product of:
          0.09831624 = sum of:
            0.09831624 = weight(_text_:22 in 6265) [ClassicSimilarity], result of:
              0.09831624 = score(doc=6265,freq=2.0), product of:
                0.18150859 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0518325 = queryNorm
                0.5416616 = fieldWeight in 6265, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6265)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Information outlook. 9(2005) no.8, S.22-23
  6. Nohr, H.: Grundlagen der automatischen Indexierung : ein Lehrbuch (2003) 0.04
    0.044227973 = product of:
      0.06634196 = sum of:
        0.0122269625 = weight(_text_:information in 1767) [ClassicSimilarity], result of:
          0.0122269625 = score(doc=1767,freq=6.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.1343758 = fieldWeight in 1767, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=1767)
        0.054114997 = sum of:
          0.026024643 = weight(_text_:management in 1767) [ClassicSimilarity], result of:
            0.026024643 = score(doc=1767,freq=2.0), product of:
              0.17470726 = queryWeight, product of:
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.0518325 = queryNorm
              0.14896142 = fieldWeight in 1767, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.03125 = fieldNorm(doc=1767)
          0.028090354 = weight(_text_:22 in 1767) [ClassicSimilarity], result of:
            0.028090354 = score(doc=1767,freq=2.0), product of:
              0.18150859 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0518325 = queryNorm
              0.15476047 = fieldWeight in 1767, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1767)
      0.6666667 = coord(2/3)
    
    Date
    22. 6.2009 12:46:51
    Footnote
    Rez. in: nfd 54(2003) H.5, S.314 (W. Ratzek): "Um entscheidungsrelevante Daten aus der ständig wachsenden Flut von mehr oder weniger relevanten Dokumenten zu extrahieren, müssen Unternehmen, öffentliche Verwaltung oder Einrichtungen der Fachinformation effektive und effiziente Filtersysteme entwickeln, einsetzen und pflegen. Das vorliegende Lehrbuch von Holger Nohr bietet erstmalig eine grundlegende Einführung in das Thema "automatische Indexierung". Denn: "Wie man Information sammelt, verwaltet und verwendet, wird darüber entscheiden, ob man zu den Gewinnern oder Verlierern gehört" (Bill Gates), heißt es einleitend. Im ersten Kapitel "Einleitung" stehen die Grundlagen im Mittelpunkt. Die Zusammenhänge zwischen Dokumenten-Management-Systeme, Information Retrieval und Indexierung für Planungs-, Entscheidungs- oder Innovationsprozesse, sowohl in Profit- als auch Non-Profit-Organisationen werden beschrieben. Am Ende des einleitenden Kapitels geht Nohr auf die Diskussion um die intellektuelle und automatische Indexierung ein und leitet damit über zum zweiten Kapitel "automatisches Indexieren. Hier geht der Autor überblickartig unter anderem ein auf - Probleme der automatischen Sprachverarbeitung und Indexierung - verschiedene Verfahren der automatischen Indexierung z.B. einfache Stichwortextraktion / Volltextinvertierung, - statistische Verfahren, Pattern-Matching-Verfahren. Die "Verfahren der automatischen Indexierung" behandelt Nohr dann vertiefend und mit vielen Beispielen versehen im umfangreichsten dritten Kapitel. Das vierte Kapitel "Keyphrase Extraction" nimmt eine Passpartout-Status ein: "Eine Zwischenstufe auf dem Weg von der automatischen Indexierung hin zur automatischen Generierung textueller Zusammenfassungen (Automatic Text Summarization) stellen Ansätze dar, die Schlüsselphrasen aus Dokumenten extrahieren (Keyphrase Extraction). Die Grenzen zwischen den automatischen Verfahren der Indexierung und denen des Text Summarization sind fließend." (S. 91). Am Beispiel NCR"s Extractor/Copernic Summarizer beschreibt Nohr die Funktionsweise.
    Im fünften Kapitel "Information Extraction" geht Nohr auf eine Problemstellung ein, die in der Fachwelt eine noch stärkere Betonung verdiente: "Die stetig ansteigende Zahl elektronischer Dokumente macht neben einer automatischen Erschließung auch eine automatische Gewinnung der relevanten Informationen aus diesen Dokumenten wünschenswert, um diese z.B. für weitere Bearbeitungen oder Auswertungen in betriebliche Informationssysteme übernehmen zu können." (S. 103) "Indexierung und Retrievalverfahren" als voneinander abhängige Verfahren werden im sechsten Kapitel behandelt. Hier stehen Relevance Ranking und Relevance Feedback sowie die Anwendung informationslinguistischer Verfahren in der Recherche im Mittelpunkt. Die "Evaluation automatischer Indexierung" setzt den thematischen Schlusspunkt. Hier geht es vor allem um die Oualität einer Indexierung, um gängige Retrievalmaße in Retrievaltest und deren Einssatz. Weiterhin ist hervorzuheben, dass jedes Kapitel durch die Vorgabe von Lernzielen eingeleitet wird und zu den jeweiligen Kapiteln (im hinteren Teil des Buches) einige Kontrollfragen gestellt werden. Die sehr zahlreichen Beispiele aus der Praxis, ein Abkürzungsverzeichnis und ein Sachregister erhöhen den Nutzwert des Buches. Die Lektüre förderte beim Rezensenten das Verständnis für die Zusammenhänge von BID-Handwerkzeug, Wirtschaftsinformatik (insbesondere Data Warehousing) und Künstlicher Intelligenz. Die "Grundlagen der automatischen Indexierung" sollte auch in den bibliothekarischen Studiengängen zur Pflichtlektüre gehören. Holger Nohrs Lehrbuch ist auch für den BID-Profi geeignet, um die mehr oder weniger fundierten Kenntnisse auf dem Gebiet "automatisches Indexieren" schnell, leicht verständlich und informativ aufzufrischen."
  7. Greiner-Petter, A.; Schubotz, M.; Cohl, H.S.; Gipp, B.: Semantic preserving bijective mappings for expressions involving special functions between computer algebra systems and document preparation systems (2019) 0.04
    0.04273218 = product of:
      0.06409827 = sum of:
        0.009983272 = weight(_text_:information in 5499) [ClassicSimilarity], result of:
          0.009983272 = score(doc=5499,freq=4.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.10971737 = fieldWeight in 5499, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=5499)
        0.054114997 = sum of:
          0.026024643 = weight(_text_:management in 5499) [ClassicSimilarity], result of:
            0.026024643 = score(doc=5499,freq=2.0), product of:
              0.17470726 = queryWeight, product of:
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.0518325 = queryNorm
              0.14896142 = fieldWeight in 5499, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.03125 = fieldNorm(doc=5499)
          0.028090354 = weight(_text_:22 in 5499) [ClassicSimilarity], result of:
            0.028090354 = score(doc=5499,freq=2.0), product of:
              0.18150859 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0518325 = queryNorm
              0.15476047 = fieldWeight in 5499, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=5499)
      0.6666667 = coord(2/3)
    
    Date
    20. 1.2015 18:30:22
    Footnote
    Beitrag in einem Special Issue: Information Science in the German-speaking Countries.
    Source
    Aslib journal of information management. 71(2019) no.3, S.415-439
  8. Anderson, J.D.; Pérez-Carballo, J.: ¬The nature of indexing: how humans and machines analyze messages and texts for retrieval : Part I: Research and the nature of human indexing (2001) 0.04
    0.040143125 = product of:
      0.060214683 = sum of:
        0.02117772 = weight(_text_:information in 3136) [ClassicSimilarity], result of:
          0.02117772 = score(doc=3136,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.23274569 = fieldWeight in 3136, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=3136)
        0.039036963 = product of:
          0.078073926 = sum of:
            0.078073926 = weight(_text_:management in 3136) [ClassicSimilarity], result of:
              0.078073926 = score(doc=3136,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.44688427 = fieldWeight in 3136, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3136)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Information processing and management. 37(2001) no.2, S.231-254
  9. Biebricher, N.; Fuhr, N.; Lustig, G.; Schwantner, M.; Knorz, G.: ¬The automatic indexing system AIR/PHYS : from research to application (1988) 0.04
    0.04004742 = product of:
      0.060071126 = sum of:
        0.02495818 = weight(_text_:information in 1952) [ClassicSimilarity], result of:
          0.02495818 = score(doc=1952,freq=4.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.27429342 = fieldWeight in 1952, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=1952)
        0.035112944 = product of:
          0.07022589 = sum of:
            0.07022589 = weight(_text_:22 in 1952) [ClassicSimilarity], result of:
              0.07022589 = score(doc=1952,freq=2.0), product of:
                0.18150859 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0518325 = queryNorm
                0.38690117 = fieldWeight in 1952, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1952)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Date
    16. 8.1998 12:51:22
    Footnote
    Wiederabgedruckt in: Readings in information retrieval. Ed.: K. Sparck Jones u. P. Willett. San Francisco: Morgan Kaufmann 1997. S.513-517.
    Source
    Proceedings of the 11th annual conference on research and development in information retrieval. Ed.: Y. Chiaramella
  10. Smart, G.: Using language analysis to manage information (1993) 0.04
    0.03839635 = product of:
      0.057594523 = sum of:
        0.03156988 = weight(_text_:information in 4423) [ClassicSimilarity], result of:
          0.03156988 = score(doc=4423,freq=10.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.3469568 = fieldWeight in 4423, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=4423)
        0.026024643 = product of:
          0.052049287 = sum of:
            0.052049287 = weight(_text_:management in 4423) [ClassicSimilarity], result of:
              0.052049287 = score(doc=4423,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.29792285 = fieldWeight in 4423, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4423)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The ESPRIT project SIMPR developed software to analyse documents and generate indexes for them. Of immediate application as a document indexing and classification system, this also offers a technology for information modelling that has broader implications, supporting many new uses for information management softeware. The project was based on the assumption that information can only be managed successfully by computer systems that can view the information contained in a document through the language in which the document is written, and that systems need to be sufficiently flexible to respond to the changing requirements of document use
  11. Kutschekmanesch, S.; Lutes, B.; Moelle, K.; Thiel, U.; Tzeras, K.: Automated multilingual indexing : a synthesis of rule-based and thesaurus-based methods (1998) 0.04
    0.035174027 = product of:
      0.05276104 = sum of:
        0.017648099 = weight(_text_:information in 4157) [ClassicSimilarity], result of:
          0.017648099 = score(doc=4157,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.19395474 = fieldWeight in 4157, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=4157)
        0.035112944 = product of:
          0.07022589 = sum of:
            0.07022589 = weight(_text_:22 in 4157) [ClassicSimilarity], result of:
              0.07022589 = score(doc=4157,freq=2.0), product of:
                0.18150859 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0518325 = queryNorm
                0.38690117 = fieldWeight in 4157, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4157)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Information und Märkte: 50. Deutscher Dokumentartag 1998, Kongreß der Deutschen Gesellschaft für Dokumentation e.V. (DGD), Rheinische Friedrich-Wilhelms-Universität Bonn, 22.-24. September 1998. Hrsg. von Marlies Ockenfeld u. Gerhard J. Mantwill
  12. Riloff, E.: ¬An empirical study of automated dictionary construction for information extraction in three domains (1996) 0.04
    0.035029523 = product of:
      0.05254428 = sum of:
        0.024453925 = weight(_text_:information in 6752) [ClassicSimilarity], result of:
          0.024453925 = score(doc=6752,freq=6.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.2687516 = fieldWeight in 6752, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
        0.028090354 = product of:
          0.056180708 = sum of:
            0.056180708 = weight(_text_:22 in 6752) [ClassicSimilarity], result of:
              0.056180708 = score(doc=6752,freq=2.0), product of:
                0.18150859 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0518325 = queryNorm
                0.30952093 = fieldWeight in 6752, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6752)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    AutoSlog is a system that addresses the knowledge engineering bottleneck for information extraction. AutoSlog automatically creates domain specific dictionaries for information extraction, given an appropriate training corpus. Describes experiments with AutoSlog in terrorism, joint ventures and microelectronics domains. Compares the performance of AutoSlog across the 3 domains, discusses the lessons learned and presents results from 2 experiments which demonstrate that novice users can generate effective dictionaries using AutoSlog
    Date
    6. 3.1997 16:22:15
  13. Mars, N.J.I.: ¬The management of scientific information, or, how to cope with the flood (1996) 0.03
    0.03365238 = product of:
      0.05047857 = sum of:
        0.024453925 = weight(_text_:information in 7414) [ClassicSimilarity], result of:
          0.024453925 = score(doc=7414,freq=6.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.2687516 = fieldWeight in 7414, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=7414)
        0.026024643 = product of:
          0.052049287 = sum of:
            0.052049287 = weight(_text_:management in 7414) [ClassicSimilarity], result of:
              0.052049287 = score(doc=7414,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.29792285 = fieldWeight in 7414, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0625 = fieldNorm(doc=7414)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Research in the Knowledge-Based Systems Group of the University of Twente in the Netherlands is aimed at reducing information overload. One approach is to support indexing by the traditional method of assigning content descriptions to find documents. A second way is to use a computer program to determine what the document says without descriptors. Discusses automated indexing and direct access to information
  14. Anderson, J.D.; Pérez-Carballo, J.: ¬The nature of indexing: how humans and machines analyze messages and texts for retrieval : Part II: Machine indexing, and the allocation of human versus machine effort (2001) 0.03
    0.0334526 = product of:
      0.0501789 = sum of:
        0.017648099 = weight(_text_:information in 368) [ClassicSimilarity], result of:
          0.017648099 = score(doc=368,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.19395474 = fieldWeight in 368, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=368)
        0.032530803 = product of:
          0.06506161 = sum of:
            0.06506161 = weight(_text_:management in 368) [ClassicSimilarity], result of:
              0.06506161 = score(doc=368,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.37240356 = fieldWeight in 368, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.078125 = fieldNorm(doc=368)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Information processing and management. 37(2001) no.2, S.255-277
  15. Milstead, J.L.: Methodologies for subject analysis in bibliographic databases (1992) 0.03
    0.033116385 = product of:
      0.049674578 = sum of:
        0.017470727 = weight(_text_:information in 2311) [ClassicSimilarity], result of:
          0.017470727 = score(doc=2311,freq=4.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.1920054 = fieldWeight in 2311, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2311)
        0.032203853 = product of:
          0.064407706 = sum of:
            0.064407706 = weight(_text_:management in 2311) [ClassicSimilarity], result of:
              0.064407706 = score(doc=2311,freq=4.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.36866072 = fieldWeight in 2311, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2311)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The goal of the study was to determine the state of the art of subject analysis as applied to large bibliographic data bases. The intent was to gather and evaluate information, casting it in a form that could be applied by management. There was no attempt to determine actual costs or trade-offs among costs and possible benefits. Commercial automatic indexing packages were also reviewed. The overall conclusion was that data base producers should begin working seriously on upgrading their thesauri and codifying their indexing policies as a means of moving toward development of machine aids to indexing, but that fully automatic indexing is not yet ready for wholesale implementation
    Source
    Information processing and management. 28(1992) no.3, S.407-431
  16. Frants, V.I.; Kamenoff, N.I.; Shapiro, J.: ¬One approach to classification of users and automatic clustering of documents (1993) 0.03
    0.030660793 = product of:
      0.04599119 = sum of:
        0.019966545 = weight(_text_:information in 4569) [ClassicSimilarity], result of:
          0.019966545 = score(doc=4569,freq=4.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.21943474 = fieldWeight in 4569, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=4569)
        0.026024643 = product of:
          0.052049287 = sum of:
            0.052049287 = weight(_text_:management in 4569) [ClassicSimilarity], result of:
              0.052049287 = score(doc=4569,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.29792285 = fieldWeight in 4569, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4569)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Shows how to automatically construct a classification of users and a clustering of documents on the basis of users' information needs by creating clusters of documents and cross-references among clusters using users' search requests. Examines feedback in the construction of this classification and clustering so that the classification can be changed over time to reflect the changing needs of the users
    Source
    Information processing and management. 29(1993) no.2, S.187-195
  17. Haas, S.; He, S.: Toward the automatic identification of sublanguage vocabulary (1993) 0.03
    0.030660793 = product of:
      0.04599119 = sum of:
        0.019966545 = weight(_text_:information in 4891) [ClassicSimilarity], result of:
          0.019966545 = score(doc=4891,freq=4.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.21943474 = fieldWeight in 4891, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=4891)
        0.026024643 = product of:
          0.052049287 = sum of:
            0.052049287 = weight(_text_:management in 4891) [ClassicSimilarity], result of:
              0.052049287 = score(doc=4891,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.29792285 = fieldWeight in 4891, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4891)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Describes a method developed for automatic identification of sublanguage vocabulary words as they occur in abstracts. Describes the sublanguage vocabulary identification procedures using abstracts from computer science and library and information science as sublanguage sources. Evaluates the results using three criteria. Discuss the practical and theoretical significance of this research and plans for further experiments
    Source
    Information processing and management. 29(1993) no.6, S.721-744
  18. Flores, F.N.; Moreira, V.P.: Assessing the impact of stemming accuracy on information retrieval : a multilingual perspective (2016) 0.03
    0.030303856 = product of:
      0.045455784 = sum of:
        0.025937302 = weight(_text_:information in 3187) [ClassicSimilarity], result of:
          0.025937302 = score(doc=3187,freq=12.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.2850541 = fieldWeight in 3187, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3187)
        0.019518482 = product of:
          0.039036963 = sum of:
            0.039036963 = weight(_text_:management in 3187) [ClassicSimilarity], result of:
              0.039036963 = score(doc=3187,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.22344214 = fieldWeight in 3187, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3187)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The quality of stemming algorithms is typically measured in two different ways: (i) how accurately they map the variant forms of a word to the same stem; or (ii) how much improvement they bring to Information Retrieval systems. In this article, we evaluate various stemming algorithms, in four languages, in terms of accuracy and in terms of their aid to Information Retrieval. The aim is to assess whether the most accurate stemmers are also the ones that bring the biggest gain in Information Retrieval. Experiments in English, French, Portuguese, and Spanish show that this is not always the case, as stemmers with higher error rates yield better retrieval quality. As a byproduct, we also identified the most accurate stemmers and the best for Information Retrieval purposes.
    Source
    Information processing and management. 52(2016) no.5, S.840-854
  19. Fauzi, F.; Belkhatir, M.: Multifaceted conceptual image indexing on the world wide web (2013) 0.03
    0.028797261 = product of:
      0.043195892 = sum of:
        0.02367741 = weight(_text_:information in 2721) [ClassicSimilarity], result of:
          0.02367741 = score(doc=2721,freq=10.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.2602176 = fieldWeight in 2721, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2721)
        0.019518482 = product of:
          0.039036963 = sum of:
            0.039036963 = weight(_text_:management in 2721) [ClassicSimilarity], result of:
              0.039036963 = score(doc=2721,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.22344214 = fieldWeight in 2721, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2721)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    In this paper, we describe a user-centered design of an automated multifaceted concept-based indexing framework which analyzes the semantics of the Web image contextual information and classifies it into five broad semantic concept facets: signal, object, abstract, scene, and relational; and identifies the semantic relationships between the concepts. An important aspect of our indexing model is that it relates to the users' levels of image descriptions. Also, a major contribution relies on the fact that the classification is performed automatically with the raw image contextual information extracted from any general webpage and is not solely based on image tags like state-of-the-art solutions. Human Language Technology techniques and an external knowledge base are used to analyze the information both syntactically and semantically. Experimental results on a human-annotated Web image collection and corresponding contextual information indicate that our method outperforms empirical frameworks employing tf-idf and location-based tf-idf weighting schemes as well as n-gram indexing in a recall/precision based evaluation framework.
    Source
    Information processing and management. 49(2013) no.2, S.420-440
  20. Lepsky, K.; Vorhauer, J.: Lingo - ein open source System für die Automatische Indexierung deutschsprachiger Dokumente (2006) 0.03
    0.028139224 = product of:
      0.042208835 = sum of:
        0.01411848 = weight(_text_:information in 3581) [ClassicSimilarity], result of:
          0.01411848 = score(doc=3581,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.1551638 = fieldWeight in 3581, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=3581)
        0.028090354 = product of:
          0.056180708 = sum of:
            0.056180708 = weight(_text_:22 in 3581) [ClassicSimilarity], result of:
              0.056180708 = score(doc=3581,freq=2.0), product of:
                0.18150859 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0518325 = queryNorm
                0.30952093 = fieldWeight in 3581, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3581)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Lingo ist ein frei verfügbares System (open source) zur automatischen Indexierung der deutschen Sprache. Bei der Entwicklung von lingo standen hohe Konfigurierbarkeit und Flexibilität des Systems für unterschiedliche Einsatzmöglichkeiten im Vordergrund. Der Beitrag zeigt den Nutzen einer linguistisch basierten automatischen Indexierung für das Information Retrieval auf. Die für eine Retrievalverbesserung zur Verfügung stehende linguistische Funktionalität von lingo wird vorgestellt und an Beispielen erläutert: Grundformerkennung, Kompositumerkennung bzw. Kompositumzerlegung, Wortrelationierung, lexikalische und algorithmische Mehrwortgruppenerkennung, OCR-Fehlerkorrektur. Der offene Systemaufbau von lingo wird beschrieben, mögliche Einsatzszenarien und Anwendungsgrenzen werden benannt.
    Date
    24. 3.2006 12:22:02

Languages

Types

  • a 231
  • x 14
  • m 13
  • el 9
  • s 8
  • d 1
  • More… Less…