Search (332 results, page 1 of 17)

  • × theme_ss:"Retrievalalgorithmen"
  1. Thompson, P.: Looking back: on relevance, probabilistic indexing and information retrieval (2008) 0.05
    0.048595686 = product of:
      0.11338993 = sum of:
        0.025110744 = weight(_text_:information in 2074) [ClassicSimilarity], result of:
          0.025110744 = score(doc=2074,freq=12.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.38007212 = fieldWeight in 2074, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=2074)
        0.07455817 = weight(_text_:retrieval in 2074) [ClassicSimilarity], result of:
          0.07455817 = score(doc=2074,freq=12.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.6549133 = fieldWeight in 2074, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=2074)
        0.013721022 = product of:
          0.041163065 = sum of:
            0.041163065 = weight(_text_:29 in 2074) [ClassicSimilarity], result of:
              0.041163065 = score(doc=2074,freq=2.0), product of:
                0.13239008 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037635546 = queryNorm
                0.31092256 = fieldWeight in 2074, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2074)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    Forty-eight years ago Maron and Kuhns published their paper, "On Relevance, Probabilistic Indexing and Information Retrieval" (1960). This was the first paper to present a probabilistic approach to information retrieval, and perhaps the first paper on ranked retrieval. Although it is one of the most widely cited papers in the field of information retrieval, many researchers today may not be familiar with its influence. This paper describes the Maron and Kuhns article and the influence that it has had on the field of information retrieval.
    Date
    31. 7.2008 19:58:29
    Source
    Information processing and management. 44(2008) no.2, S.963-970
  2. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.05
    0.046531934 = product of:
      0.10857451 = sum of:
        0.020502837 = weight(_text_:information in 402) [ClassicSimilarity], result of:
          0.020502837 = score(doc=402,freq=2.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.3103276 = fieldWeight in 402, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=402)
        0.06087649 = weight(_text_:retrieval in 402) [ClassicSimilarity], result of:
          0.06087649 = score(doc=402,freq=2.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.5347345 = fieldWeight in 402, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.125 = fieldNorm(doc=402)
        0.027195185 = product of:
          0.081585556 = sum of:
            0.081585556 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.081585556 = score(doc=402,freq=2.0), product of:
                0.13179328 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037635546 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  3. Crestani, F.: Combination of similarity measures for effective spoken document retrieval (2003) 0.04
    0.040808015 = product of:
      0.0952187 = sum of:
        0.017939983 = weight(_text_:information in 4690) [ClassicSimilarity], result of:
          0.017939983 = score(doc=4690,freq=2.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.27153665 = fieldWeight in 4690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=4690)
        0.053266928 = weight(_text_:retrieval in 4690) [ClassicSimilarity], result of:
          0.053266928 = score(doc=4690,freq=2.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.46789268 = fieldWeight in 4690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.109375 = fieldNorm(doc=4690)
        0.024011787 = product of:
          0.07203536 = sum of:
            0.07203536 = weight(_text_:29 in 4690) [ClassicSimilarity], result of:
              0.07203536 = score(doc=4690,freq=2.0), product of:
                0.13239008 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037635546 = queryNorm
                0.5441145 = fieldWeight in 4690, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4690)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Source
    Journal of information science. 29(2003) no.2, S.87-96
  4. Cole, C.: Intelligent information retrieval: diagnosing information need : Part II: uncertainty expansion in a prototype of a diagnostic IR tool (1998) 0.04
    0.039802656 = product of:
      0.09287286 = sum of:
        0.026633967 = weight(_text_:information in 6432) [ClassicSimilarity], result of:
          0.026633967 = score(doc=6432,freq=6.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.40312737 = fieldWeight in 6432, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=6432)
        0.045657367 = weight(_text_:retrieval in 6432) [ClassicSimilarity], result of:
          0.045657367 = score(doc=6432,freq=2.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.40105087 = fieldWeight in 6432, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.09375 = fieldNorm(doc=6432)
        0.020581532 = product of:
          0.061744597 = sum of:
            0.061744597 = weight(_text_:29 in 6432) [ClassicSimilarity], result of:
              0.061744597 = score(doc=6432,freq=2.0), product of:
                0.13239008 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037635546 = queryNorm
                0.46638384 = fieldWeight in 6432, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6432)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Date
    11. 8.2001 14:48:29
    Source
    Information processing and management. 34(1998) no.6, S.721-731
  5. Losada, D.E.; Barreiro, A.: Emebedding term similarity and inverse document frequency into a logical model of information retrieval (2003) 0.04
    0.037209008 = product of:
      0.08682102 = sum of:
        0.020502837 = weight(_text_:information in 1422) [ClassicSimilarity], result of:
          0.020502837 = score(doc=1422,freq=8.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.3103276 = fieldWeight in 1422, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=1422)
        0.052720588 = weight(_text_:retrieval in 1422) [ClassicSimilarity], result of:
          0.052720588 = score(doc=1422,freq=6.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.46309367 = fieldWeight in 1422, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=1422)
        0.013597593 = product of:
          0.040792778 = sum of:
            0.040792778 = weight(_text_:22 in 1422) [ClassicSimilarity], result of:
              0.040792778 = score(doc=1422,freq=2.0), product of:
                0.13179328 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037635546 = queryNorm
                0.30952093 = fieldWeight in 1422, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1422)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    We propose a novel approach to incorporate term similarity and inverse document frequency into a logical model of information retrieval. The ability of the logic to handle expressive representations along with the use of such classical notions are promising characteristics for IR systems. The approach proposed here has been efficiently implemented and experiments against test collections are presented.
    Date
    22. 3.2003 19:27:23
    Footnote
    Beitrag eines Themenheftes: Mathematical, logical, and formal methods in information retrieval
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.4, S.285-301
  6. Crestani, F.; Dominich, S.; Lalmas, M.; Rijsbergen, C.J.K. van: Mathematical, logical, and formal methods in information retrieval : an introduction to the special issue (2003) 0.04
    0.03640709 = product of:
      0.08494988 = sum of:
        0.018833058 = weight(_text_:information in 1451) [ClassicSimilarity], result of:
          0.018833058 = score(doc=1451,freq=12.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.2850541 = fieldWeight in 1451, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1451)
        0.055918623 = weight(_text_:retrieval in 1451) [ClassicSimilarity], result of:
          0.055918623 = score(doc=1451,freq=12.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.49118498 = fieldWeight in 1451, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=1451)
        0.0101981945 = product of:
          0.030594582 = sum of:
            0.030594582 = weight(_text_:22 in 1451) [ClassicSimilarity], result of:
              0.030594582 = score(doc=1451,freq=2.0), product of:
                0.13179328 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037635546 = queryNorm
                0.23214069 = fieldWeight in 1451, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1451)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    Research an the use of mathematical, logical, and formal methods, has been central to Information Retrieval research for a long time. Research in this area is important not only because it helps enhancing retrieval effectiveness, but also because it helps clarifying the underlying concepts of Information Retrieval. In this article we outline some of the major aspects of the subject, and summarize the papers of this special issue with respect to how they relate to these aspects. We conclude by highlighting some directions of future research, which are needed to better understand the formal characteristics of Information Retrieval.
    Date
    22. 3.2003 19:27:36
    Footnote
    Einführung zu den Beiträgen eines Themenheftes: Mathematical, logical, and formal methods in information retrieval
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.4, S.281-284
  7. Smeaton, A.F.; Rijsbergen, C.J. van: ¬The retrieval effects of query expansion on a feedback document retrieval system (1983) 0.03
    0.033159092 = product of:
      0.116056815 = sum of:
        0.092261024 = weight(_text_:retrieval in 2134) [ClassicSimilarity], result of:
          0.092261024 = score(doc=2134,freq=6.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.8104139 = fieldWeight in 2134, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.109375 = fieldNorm(doc=2134)
        0.023795787 = product of:
          0.07138736 = sum of:
            0.07138736 = weight(_text_:22 in 2134) [ClassicSimilarity], result of:
              0.07138736 = score(doc=2134,freq=2.0), product of:
                0.13179328 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037635546 = queryNorm
                0.5416616 = fieldWeight in 2134, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2134)
          0.33333334 = coord(1/3)
      0.2857143 = coord(2/7)
    
    Date
    30. 3.2001 13:32:22
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  8. Faloutsos, C.: Signature files (1992) 0.03
    0.032815546 = product of:
      0.0765696 = sum of:
        0.010251419 = weight(_text_:information in 3499) [ClassicSimilarity], result of:
          0.010251419 = score(doc=3499,freq=2.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.1551638 = fieldWeight in 3499, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=3499)
        0.052720588 = weight(_text_:retrieval in 3499) [ClassicSimilarity], result of:
          0.052720588 = score(doc=3499,freq=6.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.46309367 = fieldWeight in 3499, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=3499)
        0.013597593 = product of:
          0.040792778 = sum of:
            0.040792778 = weight(_text_:22 in 3499) [ClassicSimilarity], result of:
              0.040792778 = score(doc=3499,freq=2.0), product of:
                0.13179328 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037635546 = queryNorm
                0.30952093 = fieldWeight in 3499, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3499)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    Presents a survey and discussion on signature-based text retrieval methods. It describes the main idea behind the signature approach and its advantages over other text retrieval methods, it provides a classification of the signature methods that have appeared in the literature, it describes the main representatives of each class, together with the relative advantages and drawbacks, and it gives a list of applications as well as commercial or university prototypes that use the signature approach
    Date
    7. 5.1999 15:22:48
    Source
    Information retrieval: data structures and algorithms. Ed.: W.B. Frakes u. R. Baeza-Yates
  9. Qi, Q.; Hessen, D.J.; Heijden, P.G.M. van der: Improving information retrieval through correspondenceanalysis instead of latent semantic analysis (2023) 0.03
    0.031345837 = product of:
      0.073140286 = sum of:
        0.017192151 = weight(_text_:information in 1045) [ClassicSimilarity], result of:
          0.017192151 = score(doc=1045,freq=10.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.2602176 = fieldWeight in 1045, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1045)
        0.045657367 = weight(_text_:retrieval in 1045) [ClassicSimilarity], result of:
          0.045657367 = score(doc=1045,freq=8.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.40105087 = fieldWeight in 1045, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=1045)
        0.010290766 = product of:
          0.030872298 = sum of:
            0.030872298 = weight(_text_:29 in 1045) [ClassicSimilarity], result of:
              0.030872298 = score(doc=1045,freq=2.0), product of:
                0.13239008 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037635546 = queryNorm
                0.23319192 = fieldWeight in 1045, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1045)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    The initial dimensions extracted by latent semantic analysis (LSA) of a document-term matrixhave been shown to mainly display marginal effects, which are irrelevant for informationretrieval. To improve the performance of LSA, usually the elements of the raw document-term matrix are weighted and the weighting exponent of singular values can be adjusted.An alternative information retrieval technique that ignores the marginal effects is correspon-dence analysis (CA). In this paper, the information retrieval performance of LSA and CA isempirically compared. Moreover, it is explored whether the two weightings also improve theperformance of CA. The results for four empirical datasets show that CA always performsbetter than LSA. Weighting the elements of the raw data matrix can improve CA; however,it is data dependent and the improvement is small. Adjusting the singular value weightingexponent often improves the performance of CA; however, the extent of the improvementdepends on the dataset and the number of dimensions. (PDF) Improving information retrieval through correspondence analysis instead of latent semantic analysis.
    Date
    15. 9.2023 12:28:29
    Source
    Journal of intelligent information systems [https://doi.org/10.1007/s10844-023-00815-y]
  10. Kwok, K.L.: ¬A network approach to probabilistic information retrieval (1995) 0.03
    0.030567968 = product of:
      0.07132526 = sum of:
        0.015377128 = weight(_text_:information in 5696) [ClassicSimilarity], result of:
          0.015377128 = score(doc=5696,freq=8.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.23274569 = fieldWeight in 5696, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=5696)
        0.045657367 = weight(_text_:retrieval in 5696) [ClassicSimilarity], result of:
          0.045657367 = score(doc=5696,freq=8.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.40105087 = fieldWeight in 5696, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=5696)
        0.010290766 = product of:
          0.030872298 = sum of:
            0.030872298 = weight(_text_:29 in 5696) [ClassicSimilarity], result of:
              0.030872298 = score(doc=5696,freq=2.0), product of:
                0.13239008 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037635546 = queryNorm
                0.23319192 = fieldWeight in 5696, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5696)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    Shows how probabilistic information retrieval based on document components may be implemented as a feedforward (feedbackward) artificial neural network. The network supports adaptation of connection weights as well as the growing of new edges between queries and terms based on user relevance feedback data for training, and it reflects query modification and expansion in information retrieval. A learning rule is applied that can also be viewed as supporting sequential learning using a harmonic sequence learning rate. Experimental results with 4 standard small collections and a large Wall Street Journal collection show that small query expansion levels of about 30 terms can achieve most of the gains at the low-recall high-precision region, while larger expansion levels continue to provide gains at the high-recall low-precision region of a precision recall curve
    Date
    29. 1.1996 18:42:14
    Source
    ACM transactions on information systems. 13(1995) no.3, S.324-353
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  11. Burgin, R.: ¬The retrieval effectiveness of 5 clustering algorithms as a function of indexing exhaustivity (1995) 0.03
    0.02944858 = product of:
      0.06871335 = sum of:
        0.006407136 = weight(_text_:information in 3365) [ClassicSimilarity], result of:
          0.006407136 = score(doc=3365,freq=2.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.09697737 = fieldWeight in 3365, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3365)
        0.05380772 = weight(_text_:retrieval in 3365) [ClassicSimilarity], result of:
          0.05380772 = score(doc=3365,freq=16.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.47264296 = fieldWeight in 3365, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3365)
        0.008498495 = product of:
          0.025495486 = sum of:
            0.025495486 = weight(_text_:22 in 3365) [ClassicSimilarity], result of:
              0.025495486 = score(doc=3365,freq=2.0), product of:
                0.13179328 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037635546 = queryNorm
                0.19345059 = fieldWeight in 3365, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3365)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    The retrieval effectiveness of 5 hierarchical clustering methods (single link, complete link, group average, Ward's method, and weighted average) is examined as a function of indexing exhaustivity with 4 test collections (CR, Cranfield, Medlars, and Time). Evaluations of retrieval effectiveness, based on 3 measures of optimal retrieval performance, confirm earlier findings that the performance of a retrieval system based on single link clustering varies as a function of indexing exhaustivity but fail ti find similar patterns for other clustering methods. The data also confirm earlier findings regarding the poor performance of single link clustering is a retrieval environment. The poor performance of single link clustering appears to derive from that method's tendency to produce a small number of large, ill defined document clusters. By contrast, the data examined here found the retrieval performance of the other clustering methods to be general comparable. The data presented also provides an opportunity to examine the theoretical limits of cluster based retrieval and to compare these theoretical limits to the effectiveness of operational implementations. Performance standards of the 4 document collections examined were found to vary widely, and the effectiveness of operational implementations were found to be in the range defined as unacceptable. Further improvements in search strategies and document representations warrant investigations
    Date
    22. 2.1996 11:20:06
    Source
    Journal of the American Society for Information Science. 46(1995) no.8, S.562-572
  12. Baeza-Yates, R.A.: Introduction to data structures and algorithms related to information retrieval (1992) 0.03
    0.029064046 = product of:
      0.101724155 = sum of:
        0.025628544 = weight(_text_:information in 3082) [ClassicSimilarity], result of:
          0.025628544 = score(doc=3082,freq=8.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.38790947 = fieldWeight in 3082, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=3082)
        0.07609561 = weight(_text_:retrieval in 3082) [ClassicSimilarity], result of:
          0.07609561 = score(doc=3082,freq=8.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.6684181 = fieldWeight in 3082, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.078125 = fieldNorm(doc=3082)
      0.2857143 = coord(2/7)
    
    Abstract
    In this chapter we review the main concepts and data structures used in information retrieval, and we classify information retrieval related algorithms
    Source
    Information retrieval: data structures and algorithms. Ed.: W.B. Frakes u. R. Baeza-Yates
  13. Ayadi, H.; Torjmen-Khemakhem, M.; Daoud, M.; Xiangji Huang, J.; Ben Jemaa, M.: MF-Re-Rank : a modality feature-based re-ranking model for medical image retrieval (2018) 0.03
    0.028966306 = product of:
      0.067588046 = sum of:
        0.010251419 = weight(_text_:information in 4459) [ClassicSimilarity], result of:
          0.010251419 = score(doc=4459,freq=8.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.1551638 = fieldWeight in 4459, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=4459)
        0.05047612 = weight(_text_:retrieval in 4459) [ClassicSimilarity], result of:
          0.05047612 = score(doc=4459,freq=22.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.44337842 = fieldWeight in 4459, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=4459)
        0.006860511 = product of:
          0.020581532 = sum of:
            0.020581532 = weight(_text_:29 in 4459) [ClassicSimilarity], result of:
              0.020581532 = score(doc=4459,freq=2.0), product of:
                0.13239008 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037635546 = queryNorm
                0.15546128 = fieldWeight in 4459, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4459)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    One of the main challenges in medical image retrieval is the increasing volume of image data, which render it difficult for domain experts to find relevant information from large data sets. Effective and efficient medical image retrieval systems are required to better manage medical image information. Text-based image retrieval (TBIR) was very successful in retrieving images with textual descriptions. Several TBIR approaches rely on models based on bag-of-words approaches, in which the image retrieval problem turns into one of standard text-based information retrieval; where the meanings and values of specific medical entities in the text and metadata are ignored in the image representation and retrieval process. However, we believe that TBIR should extract specific medical entities and terms and then exploit these elements to achieve better image retrieval results. Therefore, we propose a novel reranking method based on medical-image-dependent features. These features are manually selected by a medical expert from imaging modalities and medical terminology. First, we represent queries and images using only medical-image-dependent features such as image modality and image scale. Second, we exploit the defined features in a new reranking method for medical image retrieval. Our motivation is the large influence of image modality in medical image retrieval and its impact on image-relevance scores. To evaluate our approach, we performed a series of experiments on the medical ImageCLEF data sets from 2009 to 2013. The BM25 model, a language model, and an image-relevance feedback model are used as baselines to evaluate our approach. The experimental results show that compared to the BM25 model, the proposed model significantly enhances image retrieval performance. We also compared our approach with other state-of-the-art approaches and show that our approach performs comparably to those of the top three runs in the official ImageCLEF competition.
    Date
    29. 9.2018 11:43:31
    Source
    Journal of the Association for Information Science and Technology. 69(2018) no.9, S.1095-1108
  14. Chang, C.-H.; Hsu, C.-C.: Integrating query expansion and conceptual relevance feedback for personalized Web information retrieval (1998) 0.03
    0.028929979 = product of:
      0.06750328 = sum of:
        0.017939983 = weight(_text_:information in 1319) [ClassicSimilarity], result of:
          0.017939983 = score(doc=1319,freq=8.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.27153665 = fieldWeight in 1319, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1319)
        0.037665404 = weight(_text_:retrieval in 1319) [ClassicSimilarity], result of:
          0.037665404 = score(doc=1319,freq=4.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.33085006 = fieldWeight in 1319, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1319)
        0.011897894 = product of:
          0.03569368 = sum of:
            0.03569368 = weight(_text_:22 in 1319) [ClassicSimilarity], result of:
              0.03569368 = score(doc=1319,freq=2.0), product of:
                0.13179328 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037635546 = queryNorm
                0.2708308 = fieldWeight in 1319, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1319)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    Keyword based querying has been an immediate and efficient way to specify and retrieve related information that the user inquired. However, conventional document ranking based on an automatic assessment of document relevance to the query may not be the best approach when little information is given. Proposes an idea to integrate 2 existing techniques, query expansion and relevance feedback to achieve a concept-based information search for the Web
    Date
    1. 8.1996 22:08:06
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  15. Paris, L.A.H.; Tibbo, H.R.: Freestyle vs. Boolean : a comparison of partial and exact match retrieval systems (1998) 0.03
    0.028759884 = product of:
      0.067106396 = sum of:
        0.0089699915 = weight(_text_:information in 3329) [ClassicSimilarity], result of:
          0.0089699915 = score(doc=3329,freq=2.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.13576832 = fieldWeight in 3329, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3329)
        0.046130512 = weight(_text_:retrieval in 3329) [ClassicSimilarity], result of:
          0.046130512 = score(doc=3329,freq=6.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.40520695 = fieldWeight in 3329, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3329)
        0.0120058935 = product of:
          0.03601768 = sum of:
            0.03601768 = weight(_text_:29 in 3329) [ClassicSimilarity], result of:
              0.03601768 = score(doc=3329,freq=2.0), product of:
                0.13239008 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037635546 = queryNorm
                0.27205724 = fieldWeight in 3329, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3329)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    Compares the performance of partial match options, LEXIS/NEXIS's Freestyle, with that of traditional Boolean retrieval. Defines natural language and the natural language search engines currently available. Although the Boolean searches had better results more often than the Freestyle searches, neither mechanism demonstrated superior performance for every query. These results do not in any way prove the superiority of partial match techniques or exact match techniques, but they do suggest that different queries demand different techniques. Further study and analysis are needed to determine which elements of a query make it best suited for partial match or exact match retrieval
    Date
    12. 3.1999 10:29:27
    Source
    Information processing and management. 34(1998) nos.2/3, S.175-190
  16. Otterbacher, J.; Erkan, G.; Radev, D.R.: Biased LexRank : passage retrieval using random walks with question-based priors (2009) 0.03
    0.028759884 = product of:
      0.067106396 = sum of:
        0.0089699915 = weight(_text_:information in 2450) [ClassicSimilarity], result of:
          0.0089699915 = score(doc=2450,freq=2.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.13576832 = fieldWeight in 2450, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2450)
        0.046130512 = weight(_text_:retrieval in 2450) [ClassicSimilarity], result of:
          0.046130512 = score(doc=2450,freq=6.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.40520695 = fieldWeight in 2450, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2450)
        0.0120058935 = product of:
          0.03601768 = sum of:
            0.03601768 = weight(_text_:29 in 2450) [ClassicSimilarity], result of:
              0.03601768 = score(doc=2450,freq=2.0), product of:
                0.13239008 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037635546 = queryNorm
                0.27205724 = fieldWeight in 2450, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2450)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    We present Biased LexRank, a method for semi-supervised passage retrieval in the context of question answering. We represent a text as a graph of passages linked based on their pairwise lexical similarity. We use traditional passage retrieval techniques to identify passages that are likely to be relevant to a user's natural language question. We then perform a random walk on the lexical similarity graph in order to recursively retrieve additional passages that are similar to other relevant passages. We present results on several benchmarks that show the applicability of our work to question answering and topic-focused text summarization.
    Date
    22.11.2008 17:11:29
    Source
    Information processing and management. 45(2009) no.1, S.42-54
  17. Dominich, S.: Mathematical foundations of information retrieval (2001) 0.03
    0.028599195 = product of:
      0.06673145 = sum of:
        0.015694214 = weight(_text_:information in 1753) [ClassicSimilarity], result of:
          0.015694214 = score(doc=1753,freq=12.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.23754507 = fieldWeight in 1753, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1753)
        0.04253874 = weight(_text_:retrieval in 1753) [ClassicSimilarity], result of:
          0.04253874 = score(doc=1753,freq=10.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.37365708 = fieldWeight in 1753, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1753)
        0.008498495 = product of:
          0.025495486 = sum of:
            0.025495486 = weight(_text_:22 in 1753) [ClassicSimilarity], result of:
              0.025495486 = score(doc=1753,freq=2.0), product of:
                0.13179328 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037635546 = queryNorm
                0.19345059 = fieldWeight in 1753, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1753)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    This book offers a comprehensive and consistent mathematical approach to information retrieval (IR) without which no implementation is possible, and sheds an entirely new light upon the structure of IR models. It contains the descriptions of all IR models in a unified formal style and language, along with examples for each, thus offering a comprehensive overview of them. The book also creates mathematical foundations and a consistent mathematical theory (including all mathematical results achieved so far) of IR as a stand-alone mathematical discipline, which thus can be read and taught independently. Also, the book contains all necessary mathematical knowledge on which IR relies, to help the reader avoid searching different sources. The book will be of interest to computer or information scientists, librarians, mathematicians, undergraduate students and researchers whose work involves information retrieval.
    Date
    22. 3.2008 12:26:32
    LCSH
    Information storage and retrieval
    Subject
    Information storage and retrieval
  18. Witschel, H.F.: Global term weights in distributed environments (2008) 0.03
    0.02859807 = product of:
      0.06672883 = sum of:
        0.010873271 = weight(_text_:information in 2096) [ClassicSimilarity], result of:
          0.010873271 = score(doc=2096,freq=4.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.16457605 = fieldWeight in 2096, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2096)
        0.045657367 = weight(_text_:retrieval in 2096) [ClassicSimilarity], result of:
          0.045657367 = score(doc=2096,freq=8.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.40105087 = fieldWeight in 2096, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=2096)
        0.0101981945 = product of:
          0.030594582 = sum of:
            0.030594582 = weight(_text_:22 in 2096) [ClassicSimilarity], result of:
              0.030594582 = score(doc=2096,freq=2.0), product of:
                0.13179328 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037635546 = queryNorm
                0.23214069 = fieldWeight in 2096, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2096)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    This paper examines the estimation of global term weights (such as IDF) in information retrieval scenarios where a global view on the collection is not available. In particular, the two options of either sampling documents or of using a reference corpus independent of the target retrieval collection are compared using standard IR test collections. In addition, the possibility of pruning term lists based on frequency is evaluated. The results show that very good retrieval performance can be reached when just the most frequent terms of a collection - an "extended stop word list" - are known and all terms which are not in that list are treated equally. However, the list cannot always be fully estimated from a general-purpose reference corpus, but some "domain-specific stop words" need to be added. A good solution for achieving this is to mix estimates from small samples of the target retrieval collection with ones derived from a reference corpus.
    Date
    1. 8.2008 9:44:22
    Source
    Information processing and management. 44(2008) no.3, S.1049-1061
  19. Campos, L.M. de; Fernández-Luna, J.M.; Huete, J.F.: Implementing relevance feedback in the Bayesian network retrieval model (2003) 0.03
    0.027023837 = product of:
      0.06305562 = sum of:
        0.013316983 = weight(_text_:information in 825) [ClassicSimilarity], result of:
          0.013316983 = score(doc=825,freq=6.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.20156369 = fieldWeight in 825, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=825)
        0.03954044 = weight(_text_:retrieval in 825) [ClassicSimilarity], result of:
          0.03954044 = score(doc=825,freq=6.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.34732026 = fieldWeight in 825, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=825)
        0.0101981945 = product of:
          0.030594582 = sum of:
            0.030594582 = weight(_text_:22 in 825) [ClassicSimilarity], result of:
              0.030594582 = score(doc=825,freq=2.0), product of:
                0.13179328 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037635546 = queryNorm
                0.23214069 = fieldWeight in 825, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=825)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    Relevance Feedback consists in automatically formulating a new query according to the relevance judgments provided by the user after evaluating a set of retrieved documents. In this article, we introduce several relevance feedback methods for the Bayesian Network Retrieval ModeL The theoretical frame an which our methods are based uses the concept of partial evidences, which summarize the new pieces of information gathered after evaluating the results obtained by the original query. These partial evidences are inserted into the underlying Bayesian network and a new inference process (probabilities propagation) is run to compute the posterior relevance probabilities of the documents in the collection given the new query. The quality of the proposed methods is tested using a preliminary experimentation with different standard document collections.
    Date
    22. 3.2003 19:30:19
    Footnote
    Beitrag eines Themenheftes: Mathematical, logical, and formal methods in information retrieval
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.4, S.302-313
  20. Song, D.; Bruza, P.D.: Towards context sensitive information inference (2003) 0.03
    0.02687097 = product of:
      0.06269893 = sum of:
        0.021250067 = weight(_text_:information in 1428) [ClassicSimilarity], result of:
          0.021250067 = score(doc=1428,freq=22.0), product of:
            0.066068366 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.037635546 = queryNorm
            0.32163754 = fieldWeight in 1428, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1428)
        0.032950368 = weight(_text_:retrieval in 1428) [ClassicSimilarity], result of:
          0.032950368 = score(doc=1428,freq=6.0), product of:
            0.11384433 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.037635546 = queryNorm
            0.28943354 = fieldWeight in 1428, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1428)
        0.008498495 = product of:
          0.025495486 = sum of:
            0.025495486 = weight(_text_:22 in 1428) [ClassicSimilarity], result of:
              0.025495486 = score(doc=1428,freq=2.0), product of:
                0.13179328 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037635546 = queryNorm
                0.19345059 = fieldWeight in 1428, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1428)
          0.33333334 = coord(1/3)
      0.42857143 = coord(3/7)
    
    Abstract
    Humans can make hasty, but generally robust judgements about what a text fragment is, or is not, about. Such judgements are termed information inference. This article furnishes an account of information inference from a psychologistic stance. By drawing an theories from nonclassical logic and applied cognition, an information inference mechanism is proposed that makes inferences via computations of information flow through an approximation of a conceptual space. Within a conceptual space information is represented geometrically. In this article, geometric representations of words are realized as vectors in a high dimensional semantic space, which is automatically constructed from a text corpus. Two approaches were presented for priming vector representations according to context. The first approach uses a concept combination heuristic to adjust the vector representation of a concept in the light of the representation of another concept. The second approach computes a prototypical concept an the basis of exemplar trace texts and moves it in the dimensional space according to the context. Information inference is evaluated by measuring the effectiveness of query models derived by information flow computations. Results show that information flow contributes significantly to query model effectiveness, particularly with respect to precision. Moreover, retrieval effectiveness compares favorably with two probabilistic query models, and another based an semantic association. More generally, this article can be seen as a contribution towards realizing operational systems that mimic text-based human reasoning.
    Date
    22. 3.2003 19:35:46
    Footnote
    Beitrag eines Themenheftes: Mathematical, logical, and formal methods in information retrieval
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.4, S.321-334
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval

Languages

Types

  • a 307
  • m 12
  • el 6
  • s 5
  • r 4
  • p 2
  • x 2
  • d 1
  • More… Less…