Search (8 results, page 1 of 1)

  • × author_ss:"Harman, D."
  • × language_ss:"e"
  • × year_i:[1990 TO 2000}
  1. Harman, D.: How effective is suffixing? (1991) 0.01
    0.0073656123 = product of:
      0.02946245 = sum of:
        0.02946245 = weight(_text_:for in 2106) [ClassicSimilarity], result of:
          0.02946245 = score(doc=2106,freq=2.0), product of:
            0.08876751 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.047278564 = queryNorm
            0.33190575 = fieldWeight in 2106, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.125 = fieldNorm(doc=2106)
      0.25 = coord(1/4)
    
    Source
    Journal of the American Society for Information Science. 42(1991) no.1, S.7-15
  2. Harman, D.: ¬The Text REtrieval Conferences (TRECs) : providing a test-bed for information retrieval systems (1998) 0.01
    0.0055814567 = product of:
      0.022325827 = sum of:
        0.022325827 = weight(_text_:for in 1314) [ClassicSimilarity], result of:
          0.022325827 = score(doc=1314,freq=6.0), product of:
            0.08876751 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.047278564 = queryNorm
            0.25150898 = fieldWeight in 1314, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1314)
      0.25 = coord(1/4)
    
    Abstract
    The Text REtrieval Conference (TREC) workshop series encourages research in information retrieval from large text applications by providing a large test collection, uniform scoring procedures and a forum for organizations interested in comparing their results. Now in its seventh year, the conference has become the major experimental effort in the field. Participants in the TREC conferences have examined a wide variety of retrieval techniques, including methods using automatic thesauri, sophisticated term weighting, natural language techniques, relevance feedback and advanced pattern matching. The TREC conference series is co-sponsored by the National Institute of Standards and Technology (NIST) and the Information Technology Office of the Defense Advanced Research Projects Agency (DARPA)
    Source
    Bulletin of the American Society for Information Science. 24(1998), April/May, S.11-13
  3. Harman, D.; Fox, E.; Baeza-Yates, R.; Lee, W.: Inverted files (1992) 0.01
    0.0052082743 = product of:
      0.020833097 = sum of:
        0.020833097 = weight(_text_:for in 3497) [ClassicSimilarity], result of:
          0.020833097 = score(doc=3497,freq=4.0), product of:
            0.08876751 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.047278564 = queryNorm
            0.23469281 = fieldWeight in 3497, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0625 = fieldNorm(doc=3497)
      0.25 = coord(1/4)
    
    Abstract
    This chaper presents a survey of the various structures (techniques) that can be used in building inverted files, and gives the details for producing an inverted file using sorted arrays. The chapter ends with 2 modifications to this basic method that are affective for large data collections
  4. Harman, D.: Relevance feedback and other query modification techniques (1992) 0.01
    0.0052082743 = product of:
      0.020833097 = sum of:
        0.020833097 = weight(_text_:for in 3508) [ClassicSimilarity], result of:
          0.020833097 = score(doc=3508,freq=4.0), product of:
            0.08876751 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.047278564 = queryNorm
            0.23469281 = fieldWeight in 3508, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0625 = fieldNorm(doc=3508)
      0.25 = coord(1/4)
    
    Abstract
    Presents a survey of relevance feedback techniques that have been used in past research, recommends various query modification approaches for use in different retrieval systems, and gives some guidelines for the efficient design of the relevance feedback component of a retrieval system
  5. Harman, D.: User-friendly systems instead of user-friendly front-ends (1992) 0.00
    0.00455724 = product of:
      0.01822896 = sum of:
        0.01822896 = weight(_text_:for in 3323) [ClassicSimilarity], result of:
          0.01822896 = score(doc=3323,freq=4.0), product of:
            0.08876751 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.047278564 = queryNorm
            0.20535621 = fieldWeight in 3323, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3323)
      0.25 = coord(1/4)
    
    Abstract
    Most commercial on-line information retrieval systems are not designed to service end users and, therefore, have often built front-ends to their systems specifically to serve the end user market. These front ends have not been well accepted, mostly because the underlying systems are still difficult for end users to use successfully in searching. New techniques, based on statistical methods, that allow natural language input and return lists of records in order of likely relevance, have long been available from research laboratories. Presents 4 prototype implementations of these statistical retrieval systems that demonstrate their potential as powerful and easily used retrieval systems able to service all users. The systems consist of: the PRISE system; the CITE system; the Muscat system; and the News Retrieval Tool
    Source
    Journal of the American Society for Information Science. 43(1992), S.164-174
  6. Harman, D.; McCoy, W.; Toense, R.: Prototyping a distributed information retrieval system that uses statistical ranking (1991) 0.00
    0.00455724 = product of:
      0.01822896 = sum of:
        0.01822896 = weight(_text_:for in 3844) [ClassicSimilarity], result of:
          0.01822896 = score(doc=3844,freq=4.0), product of:
            0.08876751 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.047278564 = queryNorm
            0.20535621 = fieldWeight in 3844, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3844)
      0.25 = coord(1/4)
    
    Abstract
    Centralised systems continue to dominate the information retrieval market, with increased competition from CD-ROM based systems. As more large organisations begin to implement office automation systems, however, many will find that neither of these types of retrieval systems will satisfy their requirements, especially those requirements involving easy integration into other systems and heavy usage by casual end users. A prototype distributed information retrieval system was designed and built using a distributed architecture and using statistical ranking techniques to help provide better service for the end user. The distributed architecture was shown to be a feasible alternative to centralised or CD-ROM information retrieval, and user testing of the ranking methodology showed both widespread user enthusiasm for this retrieval technique and very fast response times
  7. Harman, D.: Automatic indexing (1994) 0.00
    0.0036828062 = product of:
      0.014731225 = sum of:
        0.014731225 = weight(_text_:for in 7729) [ClassicSimilarity], result of:
          0.014731225 = score(doc=7729,freq=2.0), product of:
            0.08876751 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.047278564 = queryNorm
            0.16595288 = fieldWeight in 7729, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0625 = fieldNorm(doc=7729)
      0.25 = coord(1/4)
    
    Content
    Enthält die Abschnitte: What constitutes a record; What constitutes a word and what 'words' to index; Use of stop lists; Use of suffixing or stemming; Advanced automatic indexing techniques (term weighting, query expansion, the use of multiple-word phrases for indexing)
  8. Harman, D.: Ranking algorithms (1992) 0.00
    0.0036828062 = product of:
      0.014731225 = sum of:
        0.014731225 = weight(_text_:for in 3511) [ClassicSimilarity], result of:
          0.014731225 = score(doc=3511,freq=2.0), product of:
            0.08876751 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.047278564 = queryNorm
            0.16595288 = fieldWeight in 3511, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0625 = fieldNorm(doc=3511)
      0.25 = coord(1/4)
    
    Abstract
    Presents both a summary of past research done in the development of ranking algorithms and detailed instructions on implementing a ranking type of retrieval system. This type of retrieval system takes as input a natural language query without Boolean syntax and produces a list of records that 'answer' the query, with the records ranked in order of likely relevance. Ranking retrieval systems are particularly appropriate for end-users