Search (382 results, page 1 of 20)

  • × theme_ss:"Automatisches Indexieren"
  1. Gödert, W.; Liebig, M.: Maschinelle Indexierung auf dem Prüfstand : Ergebnisse eines Retrievaltests zum MILOS II Projekt (1997) 0.04
    0.040197317 = product of:
      0.10049329 = sum of:
        0.007388207 = weight(_text_:a in 1174) [ClassicSimilarity], result of:
          0.007388207 = score(doc=1174,freq=6.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.1544581 = fieldWeight in 1174, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1174)
        0.093105085 = weight(_text_:68 in 1174) [ClassicSimilarity], result of:
          0.093105085 = score(doc=1174,freq=2.0), product of:
            0.2234734 = queryWeight, product of:
              5.386969 = idf(docFreq=549, maxDocs=44218)
              0.04148407 = queryNorm
            0.41662714 = fieldWeight in 1174, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.386969 = idf(docFreq=549, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1174)
      0.4 = coord(2/5)
    
    Abstract
    The test ran between Nov 95-Aug 96 in Cologne Fachhochschule fur Bibliothekswesen (College of Librarianship).The test basis was a database of 190,000 book titles published between 1990-95. MILOS II mechanized indexing methods proved helpful in avoiding or reducing numbers of unsatisfied/no result retrieval searches. Retrieval from mechanised indexing is 3 times more successful than from title keyword data. MILOS II also used a standardized semantic vocabulary. Mechanised indexing demands high quality software and output data
    Source
    Bibliotheksdienst. 31(1997) H.1, S.59-68
    Type
    a
  2. Zhitomirsky-Geffet, M.; Prebor, G.; Bloch, O.: Improving proverb search and retrieval with a generic multidimensional ontology (2017) 0.04
    0.035791118 = product of:
      0.08947779 = sum of:
        0.009673434 = weight(_text_:a in 3320) [ClassicSimilarity], result of:
          0.009673434 = score(doc=3320,freq=14.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.20223314 = fieldWeight in 3320, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=3320)
        0.07980436 = weight(_text_:68 in 3320) [ClassicSimilarity], result of:
          0.07980436 = score(doc=3320,freq=2.0), product of:
            0.2234734 = queryWeight, product of:
              5.386969 = idf(docFreq=549, maxDocs=44218)
              0.04148407 = queryNorm
            0.35710898 = fieldWeight in 3320, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.386969 = idf(docFreq=549, maxDocs=44218)
              0.046875 = fieldNorm(doc=3320)
      0.4 = coord(2/5)
    
    Abstract
    The goal of this research is to develop a generic ontological model for proverbs that unifies potential classification criteria and various characteristics of proverbs to enable their effective retrieval and large-scale analysis. Because proverbs can be described and indexed by multiple characteristics and criteria, we built a multidimensional ontology suitable for proverb classification. To evaluate the effectiveness of the constructed ontology for improving search and retrieval of proverbs, a large-scale user experiment was arranged with 70 users who were asked to search a proverb repository using ontology-based and free-text search interfaces. The comparative analysis of the results shows that the use of this ontology helped to substantially improve the search recall, precision, user satisfaction, and efficiency and to minimize user effort during the search process. A practical contribution of this work is an automated web-based proverb search and retrieval system which incorporates the proposed ontological scheme and an initial corpus of ontology-based annotated proverbs.
    Source
    Journal of the Association for Information Science and Technology. 68(2017) no.1, S.141-153
    Type
    a
  3. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.02
    0.021885622 = product of:
      0.054714054 = sum of:
        0.009749904 = weight(_text_:a in 402) [ClassicSimilarity], result of:
          0.009749904 = score(doc=402,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.20383182 = fieldWeight in 402, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.125 = fieldNorm(doc=402)
        0.04496415 = product of:
          0.0899283 = sum of:
            0.0899283 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.0899283 = score(doc=402,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
    Type
    a
  4. Fuhr, N.; Niewelt, B.: ¬Ein Retrievaltest mit automatisch indexierten Dokumenten (1984) 0.02
    0.019149918 = product of:
      0.047874793 = sum of:
        0.008531166 = weight(_text_:a in 262) [ClassicSimilarity], result of:
          0.008531166 = score(doc=262,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.17835285 = fieldWeight in 262, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=262)
        0.03934363 = product of:
          0.07868726 = sum of:
            0.07868726 = weight(_text_:22 in 262) [ClassicSimilarity], result of:
              0.07868726 = score(doc=262,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.5416616 = fieldWeight in 262, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=262)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Date
    20.10.2000 12:22:23
    Type
    a
  5. Hlava, M.M.K.: Automatic indexing : comparing rule-based and statistics-based indexing systems (2005) 0.02
    0.019149918 = product of:
      0.047874793 = sum of:
        0.008531166 = weight(_text_:a in 6265) [ClassicSimilarity], result of:
          0.008531166 = score(doc=6265,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.17835285 = fieldWeight in 6265, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=6265)
        0.03934363 = product of:
          0.07868726 = sum of:
            0.07868726 = weight(_text_:22 in 6265) [ClassicSimilarity], result of:
              0.07868726 = score(doc=6265,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.5416616 = fieldWeight in 6265, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6265)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Source
    Information outlook. 9(2005) no.8, S.22-23
    Type
    a
  6. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.02
    0.016414216 = product of:
      0.041035537 = sum of:
        0.007312428 = weight(_text_:a in 58) [ClassicSimilarity], result of:
          0.007312428 = score(doc=58,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.15287387 = fieldWeight in 58, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=58)
        0.03372311 = product of:
          0.06744622 = sum of:
            0.06744622 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
              0.06744622 = score(doc=58,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.46428138 = fieldWeight in 58, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=58)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Date
    14. 6.2015 22:12:44
    Type
    a
  7. Hauer, M.: Automatische Indexierung (2000) 0.02
    0.016414216 = product of:
      0.041035537 = sum of:
        0.007312428 = weight(_text_:a in 5887) [ClassicSimilarity], result of:
          0.007312428 = score(doc=5887,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.15287387 = fieldWeight in 5887, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=5887)
        0.03372311 = product of:
          0.06744622 = sum of:
            0.06744622 = weight(_text_:22 in 5887) [ClassicSimilarity], result of:
              0.06744622 = score(doc=5887,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.46428138 = fieldWeight in 5887, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=5887)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Source
    Wissen in Aktion: Wege des Knowledge Managements. 22. Online-Tagung der DGI, Frankfurt am Main, 2.-4.5.2000. Proceedings. Hrsg.: R. Schmidt
    Type
    a
  8. Fuhr, N.: Rankingexperimente mit gewichteter Indexierung (1986) 0.02
    0.016414216 = product of:
      0.041035537 = sum of:
        0.007312428 = weight(_text_:a in 2051) [ClassicSimilarity], result of:
          0.007312428 = score(doc=2051,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.15287387 = fieldWeight in 2051, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=2051)
        0.03372311 = product of:
          0.06744622 = sum of:
            0.06744622 = weight(_text_:22 in 2051) [ClassicSimilarity], result of:
              0.06744622 = score(doc=2051,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.46428138 = fieldWeight in 2051, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2051)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Date
    14. 6.2015 22:12:56
    Type
    a
  9. Hauer, M.: Tiefenindexierung im Bibliothekskatalog : 17 Jahre intelligentCAPTURE (2019) 0.02
    0.016414216 = product of:
      0.041035537 = sum of:
        0.007312428 = weight(_text_:a in 5629) [ClassicSimilarity], result of:
          0.007312428 = score(doc=5629,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.15287387 = fieldWeight in 5629, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=5629)
        0.03372311 = product of:
          0.06744622 = sum of:
            0.06744622 = weight(_text_:22 in 5629) [ClassicSimilarity], result of:
              0.06744622 = score(doc=5629,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.46428138 = fieldWeight in 5629, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=5629)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Source
    B.I.T.online. 22(2019) H.2, S.163-166
    Type
    a
  10. Kutschekmanesch, S.; Lutes, B.; Moelle, K.; Thiel, U.; Tzeras, K.: Automated multilingual indexing : a synthesis of rule-based and thesaurus-based methods (1998) 0.01
    0.014688149 = product of:
      0.036720373 = sum of:
        0.00861778 = weight(_text_:a in 4157) [ClassicSimilarity], result of:
          0.00861778 = score(doc=4157,freq=4.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.18016359 = fieldWeight in 4157, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=4157)
        0.028102593 = product of:
          0.056205187 = sum of:
            0.056205187 = weight(_text_:22 in 4157) [ClassicSimilarity], result of:
              0.056205187 = score(doc=4157,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.38690117 = fieldWeight in 4157, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4157)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Source
    Information und Märkte: 50. Deutscher Dokumentartag 1998, Kongreß der Deutschen Gesellschaft für Dokumentation e.V. (DGD), Rheinische Friedrich-Wilhelms-Universität Bonn, 22.-24. September 1998. Hrsg. von Marlies Ockenfeld u. Gerhard J. Mantwill
    Type
    a
  11. Stankovic, R. et al.: Indexing of textual databases based on lexical resources : a case study for Serbian (2016) 0.01
    0.014688149 = product of:
      0.036720373 = sum of:
        0.00861778 = weight(_text_:a in 2759) [ClassicSimilarity], result of:
          0.00861778 = score(doc=2759,freq=4.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.18016359 = fieldWeight in 2759, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=2759)
        0.028102593 = product of:
          0.056205187 = sum of:
            0.056205187 = weight(_text_:22 in 2759) [ClassicSimilarity], result of:
              0.056205187 = score(doc=2759,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.38690117 = fieldWeight in 2759, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2759)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Date
    1. 2.2016 18:25:22
    Type
    a
  12. Biebricher, N.; Fuhr, N.; Lustig, G.; Schwantner, M.; Knorz, G.: ¬The automatic indexing system AIR/PHYS : from research to application (1988) 0.01
    0.0136785135 = product of:
      0.034196284 = sum of:
        0.0060936897 = weight(_text_:a in 1952) [ClassicSimilarity], result of:
          0.0060936897 = score(doc=1952,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.12739488 = fieldWeight in 1952, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=1952)
        0.028102593 = product of:
          0.056205187 = sum of:
            0.056205187 = weight(_text_:22 in 1952) [ClassicSimilarity], result of:
              0.056205187 = score(doc=1952,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.38690117 = fieldWeight in 1952, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1952)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Date
    16. 8.1998 12:51:22
    Type
    a
  13. Tsareva, P.V.: Algoritmy dlya raspoznavaniya pozitivnykh i negativnykh vkhozdenii deskriptorov v tekst i protsedura avtomaticheskoi klassifikatsii tekstov (1999) 0.01
    0.0136785135 = product of:
      0.034196284 = sum of:
        0.0060936897 = weight(_text_:a in 374) [ClassicSimilarity], result of:
          0.0060936897 = score(doc=374,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.12739488 = fieldWeight in 374, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=374)
        0.028102593 = product of:
          0.056205187 = sum of:
            0.056205187 = weight(_text_:22 in 374) [ClassicSimilarity], result of:
              0.056205187 = score(doc=374,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.38690117 = fieldWeight in 374, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=374)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Date
    1. 4.2002 10:22:41
    Type
    a
  14. Tsujii, J.-I.: Automatic acquisition of semantic collocation from corpora (1995) 0.01
    0.012370296 = product of:
      0.03092574 = sum of:
        0.008443666 = weight(_text_:a in 4709) [ClassicSimilarity], result of:
          0.008443666 = score(doc=4709,freq=6.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.17652355 = fieldWeight in 4709, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=4709)
        0.022482075 = product of:
          0.04496415 = sum of:
            0.04496415 = weight(_text_:22 in 4709) [ClassicSimilarity], result of:
              0.04496415 = score(doc=4709,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.30952093 = fieldWeight in 4709, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4709)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Proposes automatic linguistic knowledge acquisition from sublanguage corpora. The system combines existing linguistic knowledge and human intervention with corpus based techniques. The algorithm involves a gradual approximation which works to converge linguistic knowledge gradually towards desirable results. The 1st experiment revealed the characteristic of this algorithm and the others proved the effectiveness of this algorithm for a real corpus
    Date
    31. 7.1996 9:22:19
    Type
    a
  15. Riloff, E.: ¬An empirical study of automated dictionary construction for information extraction in three domains (1996) 0.01
    0.011750519 = product of:
      0.029376298 = sum of:
        0.006894224 = weight(_text_:a in 6752) [ClassicSimilarity], result of:
          0.006894224 = score(doc=6752,freq=4.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.14413087 = fieldWeight in 6752, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
        0.022482075 = product of:
          0.04496415 = sum of:
            0.04496415 = weight(_text_:22 in 6752) [ClassicSimilarity], result of:
              0.04496415 = score(doc=6752,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.30952093 = fieldWeight in 6752, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6752)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    AutoSlog is a system that addresses the knowledge engineering bottleneck for information extraction. AutoSlog automatically creates domain specific dictionaries for information extraction, given an appropriate training corpus. Describes experiments with AutoSlog in terrorism, joint ventures and microelectronics domains. Compares the performance of AutoSlog across the 3 domains, discusses the lessons learned and presents results from 2 experiments which demonstrate that novice users can generate effective dictionaries using AutoSlog
    Date
    6. 3.1997 16:22:15
    Type
    a
  16. Newman, D.J.; Block, S.: Probabilistic topic decomposition of an eighteenth-century American newspaper (2006) 0.01
    0.011281192 = product of:
      0.02820298 = sum of:
        0.008531166 = weight(_text_:a in 5291) [ClassicSimilarity], result of:
          0.008531166 = score(doc=5291,freq=8.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.17835285 = fieldWeight in 5291, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5291)
        0.019671815 = product of:
          0.03934363 = sum of:
            0.03934363 = weight(_text_:22 in 5291) [ClassicSimilarity], result of:
              0.03934363 = score(doc=5291,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.2708308 = fieldWeight in 5291, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5291)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    We use a probabilistic mixture decomposition method to determine topics in the Pennsylvania Gazette, a major colonial U.S. newspaper from 1728-1800. We assess the value of several topic decomposition techniques for historical research and compare the accuracy and efficacy of various methods. After determining the topics covered by the 80,000 articles and advertisements in the entire 18th century run of the Gazette, we calculate how the prevalence of those topics changed over time, and give historically relevant examples of our findings. This approach reveals important information about the content of this colonial newspaper, and suggests the value of such approaches to a more complete understanding of early American print culture and society.
    Date
    22. 7.2006 17:32:00
    Type
    a
  17. Lepsky, K.; Vorhauer, J.: Lingo - ein open source System für die Automatische Indexierung deutschsprachiger Dokumente (2006) 0.01
    0.010942811 = product of:
      0.027357027 = sum of:
        0.004874952 = weight(_text_:a in 3581) [ClassicSimilarity], result of:
          0.004874952 = score(doc=3581,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.10191591 = fieldWeight in 3581, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=3581)
        0.022482075 = product of:
          0.04496415 = sum of:
            0.04496415 = weight(_text_:22 in 3581) [ClassicSimilarity], result of:
              0.04496415 = score(doc=3581,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.30952093 = fieldWeight in 3581, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3581)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Date
    24. 3.2006 12:22:02
    Type
    a
  18. Probst, M.; Mittelbach, J.: Maschinelle Indexierung in der Sacherschließung wissenschaftlicher Bibliotheken (2006) 0.01
    0.010942811 = product of:
      0.027357027 = sum of:
        0.004874952 = weight(_text_:a in 1755) [ClassicSimilarity], result of:
          0.004874952 = score(doc=1755,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.10191591 = fieldWeight in 1755, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=1755)
        0.022482075 = product of:
          0.04496415 = sum of:
            0.04496415 = weight(_text_:22 in 1755) [ClassicSimilarity], result of:
              0.04496415 = score(doc=1755,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.30952093 = fieldWeight in 1755, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1755)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Date
    22. 3.2008 12:35:19
    Type
    a
  19. Bordoni, L.; Pazienza, M.T.: Documents automatic indexing in an environmental domain (1997) 0.01
    0.010824009 = product of:
      0.02706002 = sum of:
        0.007388207 = weight(_text_:a in 530) [ClassicSimilarity], result of:
          0.007388207 = score(doc=530,freq=6.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.1544581 = fieldWeight in 530, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=530)
        0.019671815 = product of:
          0.03934363 = sum of:
            0.03934363 = weight(_text_:22 in 530) [ClassicSimilarity], result of:
              0.03934363 = score(doc=530,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.2708308 = fieldWeight in 530, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=530)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Describes an application of Natural Language Processing (NLP) techniques, in HIRMA (Hypertextual Information Retrieval Managed by ARIOSTO), to the problem of document indexing by referring to a system which incorporates natural language processing techniques to determine the subject of the text of documents and to associate them with relevant semantic indexes. Describes briefly the overall system, details of its implementation on a corpus of scientific abstracts related to environmental topics and experimental evidence of the system's behaviour. Analyzes in detail an experiment designed to evaluate the system's retrieval ability in terms of recall and precision
    Source
    International forum on information and documentation. 22(1997) no.1, S.17-28
    Type
    a
  20. Wolfekuhler, M.R.; Punch, W.F.: Finding salient features for personal Web pages categories (1997) 0.01
    0.010824009 = product of:
      0.02706002 = sum of:
        0.007388207 = weight(_text_:a in 2673) [ClassicSimilarity], result of:
          0.007388207 = score(doc=2673,freq=6.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.1544581 = fieldWeight in 2673, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2673)
        0.019671815 = product of:
          0.03934363 = sum of:
            0.03934363 = weight(_text_:22 in 2673) [ClassicSimilarity], result of:
              0.03934363 = score(doc=2673,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.2708308 = fieldWeight in 2673, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2673)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Examines techniques that discover features in sets of pre-categorized documents, such that similar documents can be found on the WWW. Examines techniques which will classifiy training examples with high accuracy, then explains why this is not necessarily useful. Describes a method for extracting word clusters from the raw document features. Results show that the clustering technique is successful in discovering word groups in personal Web pages which can be used to find similar information on the WWW
    Date
    1. 8.1996 22:08:06
    Footnote
    Contribution to a special issue of papers from the 6th International World Wide Web conference, held 7-11 Apr 1997, Santa Clara, California
    Type
    a

Languages

Types

  • a 364
  • el 31
  • x 5
  • m 4
  • s 3
  • d 1
  • p 1
  • More… Less…

Classifications