Search (382 results, page 1 of 20)

  • × theme_ss:"Automatisches Indexieren"
  1. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.05
    0.05025269 = product of:
      0.10050538 = sum of:
        0.10050538 = sum of:
          0.009830814 = weight(_text_:a in 402) [ClassicSimilarity], result of:
            0.009830814 = score(doc=402,freq=2.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.20383182 = fieldWeight in 402, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.125 = fieldNorm(doc=402)
          0.09067457 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
            0.09067457 = score(doc=402,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.61904186 = fieldWeight in 402, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.125 = fieldNorm(doc=402)
      0.5 = coord(1/2)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
    Type
    a
  2. Plaunt, C.; Norgard, B.A.: ¬An association-based method for automatic indexing with a controlled vocabulary (1998) 0.05
    0.046528738 = sum of:
      0.008003834 = product of:
        0.048023 = sum of:
          0.048023 = weight(_text_:authors in 1794) [ClassicSimilarity], result of:
            0.048023 = score(doc=1794,freq=2.0), product of:
              0.19068757 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.041828327 = queryNorm
              0.25184128 = fieldWeight in 1794, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1794)
        0.16666667 = coord(1/6)
      0.038524903 = sum of:
        0.010189101 = weight(_text_:a in 1794) [ClassicSimilarity], result of:
          0.010189101 = score(doc=1794,freq=22.0), product of:
            0.048230026 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.041828327 = queryNorm
            0.21126054 = fieldWeight in 1794, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1794)
        0.028335804 = weight(_text_:22 in 1794) [ClassicSimilarity], result of:
          0.028335804 = score(doc=1794,freq=2.0), product of:
            0.14647567 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.041828327 = queryNorm
            0.19345059 = fieldWeight in 1794, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1794)
    
    Abstract
    In this article, we describe and test a two-stage algorithm based on a lexical collocation technique which maps from the lexical clues contained in a document representation into a controlled vocabulary list of subject headings. Using a collection of 4.626 INSPEC documents, we create a 'dictionary' of associations between the lexical items contained in the titles, authors, and abstracts, and controlled vocabulary subject headings assigned to those records by human indexers using a likelihood ratio statistic as the measure of association. In the deployment stage, we use the dictiony to predict which of the controlled vocabulary subject headings best describe new documents when they are presented to the system. Our evaluation of this algorithm, in which we compare the automatically assigned subject headings to the subject headings assigned to the test documents by human catalogers, shows that we can obtain results comparable to, and consistent with, human cataloging. In effect we have cast this as a classic partial match information retrieval problem. We consider the problem to be one of 'retrieving' (or assigning) the most probably 'relevant' (or correct) controlled vocabulary subject headings to a document based on the clues contained in that document
    Date
    11. 9.2000 19:53:22
    Type
    a
  3. Fuhr, N.; Niewelt, B.: ¬Ein Retrievaltest mit automatisch indexierten Dokumenten (1984) 0.04
    0.043971106 = product of:
      0.08794221 = sum of:
        0.08794221 = sum of:
          0.008601963 = weight(_text_:a in 262) [ClassicSimilarity], result of:
            0.008601963 = score(doc=262,freq=2.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.17835285 = fieldWeight in 262, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.109375 = fieldNorm(doc=262)
          0.07934025 = weight(_text_:22 in 262) [ClassicSimilarity], result of:
            0.07934025 = score(doc=262,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.5416616 = fieldWeight in 262, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.109375 = fieldNorm(doc=262)
      0.5 = coord(1/2)
    
    Date
    20.10.2000 12:22:23
    Type
    a
  4. Hlava, M.M.K.: Automatic indexing : comparing rule-based and statistics-based indexing systems (2005) 0.04
    0.043971106 = product of:
      0.08794221 = sum of:
        0.08794221 = sum of:
          0.008601963 = weight(_text_:a in 6265) [ClassicSimilarity], result of:
            0.008601963 = score(doc=6265,freq=2.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.17835285 = fieldWeight in 6265, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.109375 = fieldNorm(doc=6265)
          0.07934025 = weight(_text_:22 in 6265) [ClassicSimilarity], result of:
            0.07934025 = score(doc=6265,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.5416616 = fieldWeight in 6265, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.109375 = fieldNorm(doc=6265)
      0.5 = coord(1/2)
    
    Source
    Information outlook. 9(2005) no.8, S.22-23
    Type
    a
  5. Greiner-Petter, A.; Schubotz, M.; Cohl, H.S.; Gipp, B.: Semantic preserving bijective mappings for expressions involving special functions between computer algebra systems and document preparation systems (2019) 0.04
    0.039254673 = sum of:
      0.011090438 = product of:
        0.066542625 = sum of:
          0.066542625 = weight(_text_:authors in 5499) [ClassicSimilarity], result of:
            0.066542625 = score(doc=5499,freq=6.0), product of:
              0.19068757 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.041828327 = queryNorm
              0.34896153 = fieldWeight in 5499, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.03125 = fieldNorm(doc=5499)
        0.16666667 = coord(1/6)
      0.028164236 = sum of:
        0.0054955925 = weight(_text_:a in 5499) [ClassicSimilarity], result of:
          0.0054955925 = score(doc=5499,freq=10.0), product of:
            0.048230026 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.041828327 = queryNorm
            0.11394546 = fieldWeight in 5499, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=5499)
        0.022668643 = weight(_text_:22 in 5499) [ClassicSimilarity], result of:
          0.022668643 = score(doc=5499,freq=2.0), product of:
            0.14647567 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.041828327 = queryNorm
            0.15476047 = fieldWeight in 5499, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.03125 = fieldNorm(doc=5499)
    
    Abstract
    Purpose Modern mathematicians and scientists of math-related disciplines often use Document Preparation Systems (DPS) to write and Computer Algebra Systems (CAS) to calculate mathematical expressions. Usually, they translate the expressions manually between DPS and CAS. This process is time-consuming and error-prone. The purpose of this paper is to automate this translation. This paper uses Maple and Mathematica as the CAS, and LaTeX as the DPS. Design/methodology/approach Bruce Miller at the National Institute of Standards and Technology (NIST) developed a collection of special LaTeX macros that create links from mathematical symbols to their definitions in the NIST Digital Library of Mathematical Functions (DLMF). The authors are using these macros to perform rule-based translations between the formulae in the DLMF and CAS. Moreover, the authors develop software to ease the creation of new rules and to discover inconsistencies. Findings The authors created 396 mappings and translated 58.8 percent of DLMF formulae (2,405 expressions) successfully between Maple and DLMF. For a significant percentage, the special function definitions in Maple and the DLMF were different. An atomic symbol in one system maps to a composite expression in the other system. The translator was also successfully used for automatic verification of mathematical online compendia and CAS. The evaluation techniques discovered two errors in the DLMF and one defect in Maple. Originality/value This paper introduces the first translation tool for special functions between LaTeX and CAS. The approach improves error-prone manual translations and can be used to verify mathematical online compendia and CAS.
    Date
    20. 1.2015 18:30:22
    Type
    a
  6. Mesquita, L.A.P.; Souza, R.R.; Baracho Porto, R.M.A.: Noun phrases in automatic indexing: : a structural analysis of the distribution of relevant terms in doctoral theses (2014) 0.04
    0.038590357 = sum of:
      0.006403066 = product of:
        0.038418397 = sum of:
          0.038418397 = weight(_text_:authors in 1442) [ClassicSimilarity], result of:
            0.038418397 = score(doc=1442,freq=2.0), product of:
              0.19068757 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.041828327 = queryNorm
              0.20147301 = fieldWeight in 1442, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.03125 = fieldNorm(doc=1442)
        0.16666667 = coord(1/6)
      0.03218729 = sum of:
        0.009518646 = weight(_text_:a in 1442) [ClassicSimilarity], result of:
          0.009518646 = score(doc=1442,freq=30.0), product of:
            0.048230026 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.041828327 = queryNorm
            0.19735932 = fieldWeight in 1442, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=1442)
        0.022668643 = weight(_text_:22 in 1442) [ClassicSimilarity], result of:
          0.022668643 = score(doc=1442,freq=2.0), product of:
            0.14647567 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.041828327 = queryNorm
            0.15476047 = fieldWeight in 1442, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.03125 = fieldNorm(doc=1442)
    
    Abstract
    The main objective of this research was to analyze whether there was a characteristic distribution behavior of relevant terms over a scientific text that could contribute as a criterion for their process of automatic indexing. The terms considered in this study were only full noun phrases contained in the texts themselves. The texts were considered a total of 98 doctoral theses of the eight areas of knowledge in a same university. Initially, 20 full noun phrases were automatically extracted from each text as candidates to be the most relevant terms, and each author of each text assigned a relevance value 0-6 (not relevant and highly relevant, respectively) for each of the 20 noun phrases sent. Only, 22.1 % of noun phrases were considered not relevant. A relevance values of the terms assigned by the authors were associated with their positions in the text. Each full noun phrases found in the text was considered as a valid linear position. The results that were obtained showed values resulting from this distribution by considering two types of position: linear, with values consolidated into ten equal consecutive parts; and structural, considering parts of the text (such as introduction, development and conclusion). As a result of considerable importance, all areas of knowledge related to the Natural Sciences showed a characteristic behavior in the distribution of relevant terms, as well as all areas of knowledge related to Social Sciences showed the same characteristic behavior of distribution, but distinct from the Natural Sciences. The difference of the distribution behavior between the Natural and Social Sciences can be clearly visualized through graphs. All behaviors, including the general behavior of all areas of knowledge together, were characterized in polynomial equations and can be applied in future as criteria for automatic indexing. Until the present date this work has become inedited of for two reasons: to present a method for characterizing the distribution of relevant terms in a scientific text, and also, through this method, pointing out a quantitative trait difference between the Natural and Social Sciences.
    Source
    Knowledge organization in the 21st century: between historical patterns and future prospects. Proceedings of the Thirteenth International ISKO Conference 19-22 May 2014, Kraków, Poland. Ed.: Wieslaw Babik
    Type
    a
  7. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.04
    0.03768952 = product of:
      0.07537904 = sum of:
        0.07537904 = sum of:
          0.007373111 = weight(_text_:a in 58) [ClassicSimilarity], result of:
            0.007373111 = score(doc=58,freq=2.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.15287387 = fieldWeight in 58, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.09375 = fieldNorm(doc=58)
          0.06800593 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
            0.06800593 = score(doc=58,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.46428138 = fieldWeight in 58, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.09375 = fieldNorm(doc=58)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:44
    Type
    a
  8. Hauer, M.: Automatische Indexierung (2000) 0.04
    0.03768952 = product of:
      0.07537904 = sum of:
        0.07537904 = sum of:
          0.007373111 = weight(_text_:a in 5887) [ClassicSimilarity], result of:
            0.007373111 = score(doc=5887,freq=2.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.15287387 = fieldWeight in 5887, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.09375 = fieldNorm(doc=5887)
          0.06800593 = weight(_text_:22 in 5887) [ClassicSimilarity], result of:
            0.06800593 = score(doc=5887,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.46428138 = fieldWeight in 5887, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.09375 = fieldNorm(doc=5887)
      0.5 = coord(1/2)
    
    Source
    Wissen in Aktion: Wege des Knowledge Managements. 22. Online-Tagung der DGI, Frankfurt am Main, 2.-4.5.2000. Proceedings. Hrsg.: R. Schmidt
    Type
    a
  9. Fuhr, N.: Rankingexperimente mit gewichteter Indexierung (1986) 0.04
    0.03768952 = product of:
      0.07537904 = sum of:
        0.07537904 = sum of:
          0.007373111 = weight(_text_:a in 2051) [ClassicSimilarity], result of:
            0.007373111 = score(doc=2051,freq=2.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.15287387 = fieldWeight in 2051, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.09375 = fieldNorm(doc=2051)
          0.06800593 = weight(_text_:22 in 2051) [ClassicSimilarity], result of:
            0.06800593 = score(doc=2051,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.46428138 = fieldWeight in 2051, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.09375 = fieldNorm(doc=2051)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:56
    Type
    a
  10. Hauer, M.: Tiefenindexierung im Bibliothekskatalog : 17 Jahre intelligentCAPTURE (2019) 0.04
    0.03768952 = product of:
      0.07537904 = sum of:
        0.07537904 = sum of:
          0.007373111 = weight(_text_:a in 5629) [ClassicSimilarity], result of:
            0.007373111 = score(doc=5629,freq=2.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.15287387 = fieldWeight in 5629, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.09375 = fieldNorm(doc=5629)
          0.06800593 = weight(_text_:22 in 5629) [ClassicSimilarity], result of:
            0.06800593 = score(doc=5629,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.46428138 = fieldWeight in 5629, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.09375 = fieldNorm(doc=5629)
      0.5 = coord(1/2)
    
    Source
    B.I.T.online. 22(2019) H.2, S.163-166
    Type
    a
  11. Kutschekmanesch, S.; Lutes, B.; Moelle, K.; Thiel, U.; Tzeras, K.: Automated multilingual indexing : a synthesis of rule-based and thesaurus-based methods (1998) 0.03
    0.032680452 = product of:
      0.065360904 = sum of:
        0.065360904 = sum of:
          0.008689295 = weight(_text_:a in 4157) [ClassicSimilarity], result of:
            0.008689295 = score(doc=4157,freq=4.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.18016359 = fieldWeight in 4157, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.078125 = fieldNorm(doc=4157)
          0.05667161 = weight(_text_:22 in 4157) [ClassicSimilarity], result of:
            0.05667161 = score(doc=4157,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.38690117 = fieldWeight in 4157, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.078125 = fieldNorm(doc=4157)
      0.5 = coord(1/2)
    
    Source
    Information und Märkte: 50. Deutscher Dokumentartag 1998, Kongreß der Deutschen Gesellschaft für Dokumentation e.V. (DGD), Rheinische Friedrich-Wilhelms-Universität Bonn, 22.-24. September 1998. Hrsg. von Marlies Ockenfeld u. Gerhard J. Mantwill
    Type
    a
  12. Stankovic, R. et al.: Indexing of textual databases based on lexical resources : a case study for Serbian (2016) 0.03
    0.032680452 = product of:
      0.065360904 = sum of:
        0.065360904 = sum of:
          0.008689295 = weight(_text_:a in 2759) [ClassicSimilarity], result of:
            0.008689295 = score(doc=2759,freq=4.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.18016359 = fieldWeight in 2759, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.078125 = fieldNorm(doc=2759)
          0.05667161 = weight(_text_:22 in 2759) [ClassicSimilarity], result of:
            0.05667161 = score(doc=2759,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.38690117 = fieldWeight in 2759, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.078125 = fieldNorm(doc=2759)
      0.5 = coord(1/2)
    
    Date
    1. 2.2016 18:25:22
    Type
    a
  13. Biebricher, N.; Fuhr, N.; Lustig, G.; Schwantner, M.; Knorz, G.: ¬The automatic indexing system AIR/PHYS : from research to application (1988) 0.03
    0.031407934 = product of:
      0.06281587 = sum of:
        0.06281587 = sum of:
          0.0061442587 = weight(_text_:a in 1952) [ClassicSimilarity], result of:
            0.0061442587 = score(doc=1952,freq=2.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.12739488 = fieldWeight in 1952, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.078125 = fieldNorm(doc=1952)
          0.05667161 = weight(_text_:22 in 1952) [ClassicSimilarity], result of:
            0.05667161 = score(doc=1952,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.38690117 = fieldWeight in 1952, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.078125 = fieldNorm(doc=1952)
      0.5 = coord(1/2)
    
    Date
    16. 8.1998 12:51:22
    Type
    a
  14. Tsareva, P.V.: Algoritmy dlya raspoznavaniya pozitivnykh i negativnykh vkhozdenii deskriptorov v tekst i protsedura avtomaticheskoi klassifikatsii tekstov (1999) 0.03
    0.031407934 = product of:
      0.06281587 = sum of:
        0.06281587 = sum of:
          0.0061442587 = weight(_text_:a in 374) [ClassicSimilarity], result of:
            0.0061442587 = score(doc=374,freq=2.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.12739488 = fieldWeight in 374, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.078125 = fieldNorm(doc=374)
          0.05667161 = weight(_text_:22 in 374) [ClassicSimilarity], result of:
            0.05667161 = score(doc=374,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.38690117 = fieldWeight in 374, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.078125 = fieldNorm(doc=374)
      0.5 = coord(1/2)
    
    Date
    1. 4.2002 10:22:41
    Type
    a
  15. Tsujii, J.-I.: Automatic acquisition of semantic collocation from corpora (1995) 0.03
    0.026925512 = product of:
      0.053851023 = sum of:
        0.053851023 = sum of:
          0.008513736 = weight(_text_:a in 4709) [ClassicSimilarity], result of:
            0.008513736 = score(doc=4709,freq=6.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.17652355 = fieldWeight in 4709, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0625 = fieldNorm(doc=4709)
          0.045337286 = weight(_text_:22 in 4709) [ClassicSimilarity], result of:
            0.045337286 = score(doc=4709,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.30952093 = fieldWeight in 4709, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=4709)
      0.5 = coord(1/2)
    
    Abstract
    Proposes automatic linguistic knowledge acquisition from sublanguage corpora. The system combines existing linguistic knowledge and human intervention with corpus based techniques. The algorithm involves a gradual approximation which works to converge linguistic knowledge gradually towards desirable results. The 1st experiment revealed the characteristic of this algorithm and the others proved the effectiveness of this algorithm for a real corpus
    Date
    31. 7.1996 9:22:19
    Type
    a
  16. Riloff, E.: ¬An empirical study of automated dictionary construction for information extraction in three domains (1996) 0.03
    0.026144361 = product of:
      0.052288722 = sum of:
        0.052288722 = sum of:
          0.0069514355 = weight(_text_:a in 6752) [ClassicSimilarity], result of:
            0.0069514355 = score(doc=6752,freq=4.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.14413087 = fieldWeight in 6752, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0625 = fieldNorm(doc=6752)
          0.045337286 = weight(_text_:22 in 6752) [ClassicSimilarity], result of:
            0.045337286 = score(doc=6752,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.30952093 = fieldWeight in 6752, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=6752)
      0.5 = coord(1/2)
    
    Abstract
    AutoSlog is a system that addresses the knowledge engineering bottleneck for information extraction. AutoSlog automatically creates domain specific dictionaries for information extraction, given an appropriate training corpus. Describes experiments with AutoSlog in terrorism, joint ventures and microelectronics domains. Compares the performance of AutoSlog across the 3 domains, discusses the lessons learned and presents results from 2 experiments which demonstrate that novice users can generate effective dictionaries using AutoSlog
    Date
    6. 3.1997 16:22:15
    Type
    a
  17. Lepsky, K.; Vorhauer, J.: Lingo - ein open source System für die Automatische Indexierung deutschsprachiger Dokumente (2006) 0.03
    0.025126345 = product of:
      0.05025269 = sum of:
        0.05025269 = sum of:
          0.004915407 = weight(_text_:a in 3581) [ClassicSimilarity], result of:
            0.004915407 = score(doc=3581,freq=2.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.10191591 = fieldWeight in 3581, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0625 = fieldNorm(doc=3581)
          0.045337286 = weight(_text_:22 in 3581) [ClassicSimilarity], result of:
            0.045337286 = score(doc=3581,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.30952093 = fieldWeight in 3581, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=3581)
      0.5 = coord(1/2)
    
    Date
    24. 3.2006 12:22:02
    Type
    a
  18. Probst, M.; Mittelbach, J.: Maschinelle Indexierung in der Sacherschließung wissenschaftlicher Bibliotheken (2006) 0.03
    0.025126345 = product of:
      0.05025269 = sum of:
        0.05025269 = sum of:
          0.004915407 = weight(_text_:a in 1755) [ClassicSimilarity], result of:
            0.004915407 = score(doc=1755,freq=2.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.10191591 = fieldWeight in 1755, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0625 = fieldNorm(doc=1755)
          0.045337286 = weight(_text_:22 in 1755) [ClassicSimilarity], result of:
            0.045337286 = score(doc=1755,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.30952093 = fieldWeight in 1755, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=1755)
      0.5 = coord(1/2)
    
    Date
    22. 3.2008 12:35:19
    Type
    a
  19. Newman, D.J.; Block, S.: Probabilistic topic decomposition of an eighteenth-century American newspaper (2006) 0.02
    0.024136044 = product of:
      0.04827209 = sum of:
        0.04827209 = sum of:
          0.008601963 = weight(_text_:a in 5291) [ClassicSimilarity], result of:
            0.008601963 = score(doc=5291,freq=8.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.17835285 = fieldWeight in 5291, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0546875 = fieldNorm(doc=5291)
          0.039670125 = weight(_text_:22 in 5291) [ClassicSimilarity], result of:
            0.039670125 = score(doc=5291,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.2708308 = fieldWeight in 5291, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=5291)
      0.5 = coord(1/2)
    
    Abstract
    We use a probabilistic mixture decomposition method to determine topics in the Pennsylvania Gazette, a major colonial U.S. newspaper from 1728-1800. We assess the value of several topic decomposition techniques for historical research and compare the accuracy and efficacy of various methods. After determining the topics covered by the 80,000 articles and advertisements in the entire 18th century run of the Gazette, we calculate how the prevalence of those topics changed over time, and give historically relevant examples of our findings. This approach reveals important information about the content of this colonial newspaper, and suggests the value of such approaches to a more complete understanding of early American print culture and society.
    Date
    22. 7.2006 17:32:00
    Type
    a
  20. Bordoni, L.; Pazienza, M.T.: Documents automatic indexing in an environmental domain (1997) 0.02
    0.023559822 = product of:
      0.047119644 = sum of:
        0.047119644 = sum of:
          0.0074495184 = weight(_text_:a in 530) [ClassicSimilarity], result of:
            0.0074495184 = score(doc=530,freq=6.0), product of:
              0.048230026 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041828327 = queryNorm
              0.1544581 = fieldWeight in 530, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0546875 = fieldNorm(doc=530)
          0.039670125 = weight(_text_:22 in 530) [ClassicSimilarity], result of:
            0.039670125 = score(doc=530,freq=2.0), product of:
              0.14647567 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041828327 = queryNorm
              0.2708308 = fieldWeight in 530, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=530)
      0.5 = coord(1/2)
    
    Abstract
    Describes an application of Natural Language Processing (NLP) techniques, in HIRMA (Hypertextual Information Retrieval Managed by ARIOSTO), to the problem of document indexing by referring to a system which incorporates natural language processing techniques to determine the subject of the text of documents and to associate them with relevant semantic indexes. Describes briefly the overall system, details of its implementation on a corpus of scientific abstracts related to environmental topics and experimental evidence of the system's behaviour. Analyzes in detail an experiment designed to evaluate the system's retrieval ability in terms of recall and precision
    Source
    International forum on information and documentation. 22(1997) no.1, S.17-28
    Type
    a

Languages

Types

  • a 364
  • el 31
  • x 5
  • m 4
  • s 3
  • d 1
  • p 1
  • More… Less…

Classifications