Search (3 results, page 1 of 1)

  • × author_ss:"Fox, E."
  • × theme_ss:"Retrievalalgorithmen"
  1. Fox, E.; Betrabet, S.; Koushik, M.; Lee, W.: Extended Boolean models (1992) 0.00
    0.0016616598 = product of:
      0.0033233196 = sum of:
        0.0033233196 = product of:
          0.006646639 = sum of:
            0.006646639 = weight(_text_:a in 3512) [ClassicSimilarity], result of:
              0.006646639 = score(doc=3512,freq=8.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.15287387 = fieldWeight in 3512, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3512)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The classical interpretation of Boolean operators in an information retrieval system is in general too strict. A standard Boolean query rarely comes close to retrieving all and only those documents which are relevant to a query. Many models have been proposed with the aim of softening the interpretation of the Boolean operators in order to improve the precision and recall of the search results. This chapter discusses 3 such models: the Mixed Min and Max (MMM), the Paice, and the P-noem models. The MMM and Paice models are essentially variations of the classical fuzzy-set model, while the P-norm scheme is a distance-based approach. Our experimental results indicate that each of the above models provide better performance than the classical Boolean model in terms of retrieval effectiveness
    Type
    a
  2. Harman, D.; Fox, E.; Baeza-Yates, R.; Lee, W.: Inverted files (1992) 0.00
    0.0015666279 = product of:
      0.0031332558 = sum of:
        0.0031332558 = product of:
          0.0062665115 = sum of:
            0.0062665115 = weight(_text_:a in 3497) [ClassicSimilarity], result of:
              0.0062665115 = score(doc=3497,freq=4.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.14413087 = fieldWeight in 3497, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3497)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This chaper presents a survey of the various structures (techniques) that can be used in building inverted files, and gives the details for producing an inverted file using sorted arrays. The chapter ends with 2 modifications to this basic method that are affective for large data collections
    Type
    a
  3. Wartik, S.; Fox, E.; Heath, L.; Chen, Q.-F.: Hashing algorithms (1992) 0.00
    0.0015666279 = product of:
      0.0031332558 = sum of:
        0.0031332558 = product of:
          0.0062665115 = sum of:
            0.0062665115 = weight(_text_:a in 3510) [ClassicSimilarity], result of:
              0.0062665115 = score(doc=3510,freq=4.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.14413087 = fieldWeight in 3510, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3510)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Discusses hashing, an information storage and retrieval technique useful for implementing many of the other structures in this book. The concepts underlying hashing are presented, along with 2 implementation strategies. The chapter also contains an extensive discussion of perfect hashing, an important optimization in information retrieval, and an O(n) algorithm to find minimal perfect hash functions for a set of keys
    Type
    a