Search (2 results, page 1 of 1)

  • × author_ss:"Smeaton, A.F."
  • × theme_ss:"Retrievalstudien"
  • × year_i:[1990 TO 2000}
  1. Kelledy, F.; Smeaton, A.F.: Thresholding the postings lists in information retrieval : experiments on TREC data (1995) 0.00
    4.604387E-4 = product of:
      0.00690658 = sum of:
        0.00690658 = product of:
          0.01381316 = sum of:
            0.01381316 = weight(_text_:information in 5804) [ClassicSimilarity], result of:
              0.01381316 = score(doc=5804,freq=8.0), product of:
                0.050870337 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.028978055 = queryNorm
                0.27153665 = fieldWeight in 5804, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5804)
          0.5 = coord(1/2)
      0.06666667 = coord(1/15)
    
    Abstract
    A variety of methods for speeding up the response time of information retrieval processes have been put forward, one of which is the idea of thresholding. Thresholding relies on the data in information retrieval storage structures being organised to allow cut-off points to be used during processing. These cut-off points or thresholds are designed and ised to reduce the amount of information processed and to maintain the quality or minimise the degradation of response to a user's query. TREC is an annual series of benchmarking exercises to compare indexing and retrieval techniques. Reports experiments with a portion of the TREC data where features are introduced into the retrieval process to improve response time. These features improve response time while maintaining the same level of retrieval effectiveness
  2. Smeaton, A.F.; Harman, D.: ¬The TREC experiments and their impact on Europe (1997) 0.00
    4.5571616E-4 = product of:
      0.006835742 = sum of:
        0.006835742 = product of:
          0.013671484 = sum of:
            0.013671484 = weight(_text_:information in 7702) [ClassicSimilarity], result of:
              0.013671484 = score(doc=7702,freq=6.0), product of:
                0.050870337 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.028978055 = queryNorm
                0.2687516 = fieldWeight in 7702, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0625 = fieldNorm(doc=7702)
          0.5 = coord(1/2)
      0.06666667 = coord(1/15)
    
    Abstract
    Reviews the overall results of the TREC experiments in information retrieval, which differed from other information retrieval research projects in that the document collections used in the research were massive, and the groups participating in the collaborative evaluation are among the main organizations in the field. Reviews the findings of TREC, the way in which it operates and the specialist 'tracks' it supports and concentrates on european involvement in TREC, examining the participants and the emergence of European TREC like exercises
    Source
    Journal of information science. 23(1997) no.2, S.169-174

Authors