Search (455 results, page 1 of 23)

  • × theme_ss:"Retrievalstudien"
  1. Van der Walt, H.E.A.; Brakel, P.A. van: Method for the evaluation of the retrieval effectiveness of a CD-ROM bibliographic database (1991) 0.07
    0.07369515 = product of:
      0.22108546 = sum of:
        0.043534026 = weight(_text_:retrieval in 3114) [ClassicSimilarity], result of:
          0.043534026 = score(doc=3114,freq=6.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.40520695 = fieldWeight in 3114, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3114)
        0.02015124 = weight(_text_:of in 3114) [ClassicSimilarity], result of:
          0.02015124 = score(doc=3114,freq=18.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.36282203 = fieldWeight in 3114, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3114)
        0.15740019 = weight(_text_:compact in 3114) [ClassicSimilarity], result of:
          0.15740019 = score(doc=3114,freq=2.0), product of:
            0.26885647 = queryWeight, product of:
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.035517205 = queryNorm
            0.5854432 = fieldWeight in 3114, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3114)
      0.33333334 = coord(3/9)
    
    Abstract
    Addresses the problem of how potential users of CD-ROM data bases can objectively establish which version of the same data base is best suited for a specific situation. The problem was solved by applying the retrieval effectiveness of current on-line data base search systems as a standard measurement. 5 search queries from the medical sciences were presented by experienced users of MEDLINE. Search strategies were written for both DIALOG and DATA-STAR. Search results were compared to create a recall base from documents present in both on-line searches. This recall base was then used to establish the retrieval and precision of 4 CD-ROM data bases: MEDLINE, Compact Cambrdge MEDLINE, DIALOG OnDisc, Comprehensive MEDLINE/EBSCO
    Source
    African journal of library and information science. 59(1991) no.1, S.32-42
  2. Sanderson, M.: ¬The Reuters test collection (1996) 0.05
    0.052676234 = product of:
      0.11852153 = sum of:
        0.057450015 = weight(_text_:retrieval in 6971) [ClassicSimilarity], result of:
          0.057450015 = score(doc=6971,freq=8.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.5347345 = fieldWeight in 6971, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=6971)
        0.029435357 = weight(_text_:use in 6971) [ClassicSimilarity], result of:
          0.029435357 = score(doc=6971,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.27065295 = fieldWeight in 6971, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0625 = fieldNorm(doc=6971)
        0.018803908 = weight(_text_:of in 6971) [ClassicSimilarity], result of:
          0.018803908 = score(doc=6971,freq=12.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.33856338 = fieldWeight in 6971, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=6971)
        0.012832243 = product of:
          0.03849673 = sum of:
            0.03849673 = weight(_text_:22 in 6971) [ClassicSimilarity], result of:
              0.03849673 = score(doc=6971,freq=2.0), product of:
                0.1243752 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035517205 = queryNorm
                0.30952093 = fieldWeight in 6971, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6971)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    Describes the Reuters test collection, which at 22.173 references is significantly larger than most traditional test collections. In addition, Reuters has none of the recall calculation problems normally associated with some of the larger test collections available. Explains the method derived by D.D. Lewis to perform retrieval experiments on the Reuters collection and illustrates the use of the Reuters collection using some simple retrieval experiments that compare the performance of stemming algorithms
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
  3. Hodges, P.R.: Keyword in title indexes : effectiveness of retrieval in computer searches (1983) 0.05
    0.045454558 = product of:
      0.10227276 = sum of:
        0.05026876 = weight(_text_:retrieval in 5001) [ClassicSimilarity], result of:
          0.05026876 = score(doc=5001,freq=8.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.46789268 = fieldWeight in 5001, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5001)
        0.025755936 = weight(_text_:use in 5001) [ClassicSimilarity], result of:
          0.025755936 = score(doc=5001,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.23682132 = fieldWeight in 5001, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5001)
        0.01501985 = weight(_text_:of in 5001) [ClassicSimilarity], result of:
          0.01501985 = score(doc=5001,freq=10.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.2704316 = fieldWeight in 5001, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5001)
        0.011228213 = product of:
          0.033684637 = sum of:
            0.033684637 = weight(_text_:22 in 5001) [ClassicSimilarity], result of:
              0.033684637 = score(doc=5001,freq=2.0), product of:
                0.1243752 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035517205 = queryNorm
                0.2708308 = fieldWeight in 5001, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5001)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    A study was done to test the effectiveness of retrieval using title word searching. It was based on actual search profiles used in the Mechanized Information Center at Ohio State University, in order ro replicate as closely as possible actual searching conditions. Fewer than 50% of the relevant titles were retrieved by keywords in titles. The low rate of retrieval can be attributes to three sources: titles themselves, user and information specialist ignorance of the subject vocabulary in use, and to general language problems. Across fields it was found that the social sciences had the best retrieval rate, with science having the next best, and arts and humanities the lowest. Ways to enhance and supplement keyword in title searching on the computer and in printed indexes are discussed.
    Date
    14. 3.1996 13:22:21
  4. Davis, M.W.: On the effective use of large parallel corpora in cross-language text retrieval (1998) 0.04
    0.0434326 = product of:
      0.1302978 = sum of:
        0.07462976 = weight(_text_:retrieval in 6302) [ClassicSimilarity], result of:
          0.07462976 = score(doc=6302,freq=6.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.6946405 = fieldWeight in 6302, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.09375 = fieldNorm(doc=6302)
        0.044153035 = weight(_text_:use in 6302) [ClassicSimilarity], result of:
          0.044153035 = score(doc=6302,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.40597942 = fieldWeight in 6302, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.09375 = fieldNorm(doc=6302)
        0.011514995 = weight(_text_:of in 6302) [ClassicSimilarity], result of:
          0.011514995 = score(doc=6302,freq=2.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.20732689 = fieldWeight in 6302, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.09375 = fieldNorm(doc=6302)
      0.33333334 = coord(3/9)
    
    Series
    The Kluwer International series on information retrieval
    Source
    Cross-language information retrieval. Ed.: G. Grefenstette
  5. MacCall, S.L.; Cleveland, A.D.; Gibson, I.E.: Outline and preliminary evaluation of the classical digital library model (1999) 0.04
    0.039267793 = product of:
      0.08835253 = sum of:
        0.031095734 = weight(_text_:retrieval in 6541) [ClassicSimilarity], result of:
          0.031095734 = score(doc=6541,freq=6.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.28943354 = fieldWeight in 6541, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6541)
        0.031864714 = weight(_text_:use in 6541) [ClassicSimilarity], result of:
          0.031864714 = score(doc=6541,freq=6.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.29299045 = fieldWeight in 6541, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6541)
        0.017299127 = weight(_text_:of in 6541) [ClassicSimilarity], result of:
          0.017299127 = score(doc=6541,freq=26.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.31146988 = fieldWeight in 6541, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6541)
        0.008092954 = product of:
          0.02427886 = sum of:
            0.02427886 = weight(_text_:29 in 6541) [ClassicSimilarity], result of:
              0.02427886 = score(doc=6541,freq=2.0), product of:
                0.12493842 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.035517205 = queryNorm
                0.19432661 = fieldWeight in 6541, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=6541)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    The growing number of networked information resources and services offers unprecedented opportunities for delivering high quality information to the computer desktop of a wide range of individuals. However, currently there is a reliance on a database retrieval model, in which endusers use keywords to search large collections of automatically indexed resources in order to find needed information. As an alternative to the database retrieval model, this paper outlines the classical digital library model, which is derived from traditional practices of library and information science professionals. These practices include the selection and organization of information resources for local populations of users and the integration of advanced information retrieval tools, such as databases and the Internet into these collections. To evaluate this model, library and information professionals and endusers involved with primary care medicine were asked to respond to a series of questions comparing their experiences with a digital library developed for the primary care population to their experiences with general Internet use. Preliminary results are reported
    Date
    29. 9.2001 20:12:49
    Series
    Proceedings of the American Society for Information Science; vol.36
    Source
    Knowledge: creation, organization and use. Proceedings of the 62nd Annual Meeting of the American Society for Information Science, 31.10.-4.11.1999. Ed.: L. Woods
  6. Petrelli, D.: On the role of user-centred evaluation in the advancement of interactive information retrieval (2008) 0.04
    0.039264627 = product of:
      0.08834541 = sum of:
        0.043976005 = weight(_text_:retrieval in 2026) [ClassicSimilarity], result of:
          0.043976005 = score(doc=2026,freq=12.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.40932083 = fieldWeight in 2026, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2026)
        0.018397098 = weight(_text_:use in 2026) [ClassicSimilarity], result of:
          0.018397098 = score(doc=2026,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.1691581 = fieldWeight in 2026, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2026)
        0.017952153 = weight(_text_:of in 2026) [ClassicSimilarity], result of:
          0.017952153 = score(doc=2026,freq=28.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.32322758 = fieldWeight in 2026, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2026)
        0.008020152 = product of:
          0.024060456 = sum of:
            0.024060456 = weight(_text_:22 in 2026) [ClassicSimilarity], result of:
              0.024060456 = score(doc=2026,freq=2.0), product of:
                0.1243752 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035517205 = queryNorm
                0.19345059 = fieldWeight in 2026, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2026)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    This paper discusses the role of user-centred evaluations as an essential method for researching interactive information retrieval. It draws mainly on the work carried out during the Clarity Project where different user-centred evaluations were run during the lifecycle of a cross-language information retrieval system. The iterative testing was not only instrumental to the development of a usable system, but it enhanced our knowledge of the potential, impact, and actual use of cross-language information retrieval technology. Indeed the role of the user evaluation was dual: by testing a specific prototype it was possible to gain a micro-view and assess the effectiveness of each component of the complex system; by cumulating the result of all the evaluations (in total 43 people were involved) it was possible to build a macro-view of how cross-language retrieval would impact on users and their tasks. By showing the richness of results that can be acquired, this paper aims at stimulating researchers into considering user-centred evaluations as a flexible, adaptable and comprehensive technique for investigating non-traditional information access systems.
    Footnote
    Beitrag eines Themenbereichs: Evaluation of Interactive Information Retrieval Systems
    Source
    Information processing and management. 44(2008) no.1, S.22-38
  7. TREC: experiment and evaluation in information retrieval (2005) 0.04
    0.035797123 = product of:
      0.080543526 = sum of:
        0.053106073 = weight(_text_:retrieval in 636) [ClassicSimilarity], result of:
          0.053106073 = score(doc=636,freq=70.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.49430186 = fieldWeight in 636, product of:
              8.3666 = tf(freq=70.0), with freq of:
                70.0 = termFreq=70.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.01953125 = fieldNorm(doc=636)
        0.009198549 = weight(_text_:use in 636) [ClassicSimilarity], result of:
          0.009198549 = score(doc=636,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.08457905 = fieldWeight in 636, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.01953125 = fieldNorm(doc=636)
        0.014192422 = weight(_text_:of in 636) [ClassicSimilarity], result of:
          0.014192422 = score(doc=636,freq=70.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.2555338 = fieldWeight in 636, product of:
              8.3666 = tf(freq=70.0), with freq of:
                70.0 = termFreq=70.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.01953125 = fieldNorm(doc=636)
        0.004046477 = product of:
          0.01213943 = sum of:
            0.01213943 = weight(_text_:29 in 636) [ClassicSimilarity], result of:
              0.01213943 = score(doc=636,freq=2.0), product of:
                0.12493842 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.035517205 = queryNorm
                0.097163305 = fieldWeight in 636, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=636)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    The Text REtrieval Conference (TREC), a yearly workshop hosted by the US government's National Institute of Standards and Technology, provides the infrastructure necessary for large-scale evaluation of text retrieval methodologies. With the goal of accelerating research in this area, TREC created the first large test collections of full-text documents and standardized retrieval evaluation. The impact has been significant; since TREC's beginning in 1992, retrieval effectiveness has approximately doubled. TREC has built a variety of large test collections, including collections for such specialized retrieval tasks as cross-language retrieval and retrieval of speech. Moreover, TREC has accelerated the transfer of research ideas into commercial systems, as demonstrated in the number of retrieval techniques developed in TREC that are now used in Web search engines. This book provides a comprehensive review of TREC research, summarizing the variety of TREC results, documenting the best practices in experimental information retrieval, and suggesting areas for further research. The first part of the book describes TREC's history, test collections, and retrieval methodology. Next, the book provides "track" reports -- describing the evaluations of specific tasks, including routing and filtering, interactive retrieval, and retrieving noisy text. The final part of the book offers perspectives on TREC from such participants as Microsoft Research, University of Massachusetts, Cornell University, University of Waterloo, City University of New York, and IBM. The book will be of interest to researchers in information retrieval and related technologies, including natural language processing.
    Content
    Enthält die Beiträge: 1. The Text REtrieval Conference - Ellen M. Voorhees and Donna K. Harman 2. The TREC Test Collections - Donna K. Harman 3. Retrieval System Evaluation - Chris Buckley and Ellen M. Voorhees 4. The TREC Ad Hoc Experiments - Donna K. Harman 5. Routing and Filtering - Stephen Robertson and Jamie Callan 6. The TREC Interactive Tracks: Putting the User into Search - Susan T. Dumais and Nicholas J. Belkin 7. Beyond English - Donna K. Harman 8. Retrieving Noisy Text - Ellen M. Voorhees and John S. Garofolo 9.The Very Large Collection and Web Tracks - David Hawking and Nick Craswell 10. Question Answering in TREC - Ellen M. Voorhees 11. The University of Massachusetts and a Dozen TRECs - James Allan, W. Bruce Croft and Jamie Callan 12. How Okapi Came to TREC - Stephen Robertson 13. The SMART Project at TREC - Chris Buckley 14. Ten Years of Ad Hoc Retrieval at TREC Using PIRCS - Kui-Lam Kwok 15. MultiText Experiments for TREC - Gordon V. Cormack, Charles L. A. Clarke, Christopher R. Palmer and Thomas R. Lynam 16. A Language-Modeling Approach to TREC - Djoerd Hiemstra and Wessel Kraaij 17. BM Research Activities at TREC - Eric W. Brown, David Carmel, Martin Franz, Abraham Ittycheriah, Tapas Kanungo, Yoelle Maarek, J. Scott McCarley, Robert L. Mack, John M. Prager, John R. Smith, Aya Soffer, Jason Y. Zien and Alan D. Marwick Epilogue: Metareflections on TREC - Karen Sparck Jones
    Date
    29. 3.1996 18:16:49
    Footnote
    Rez. in: JASIST 58(2007) no.6, S.910-911 (J.L. Vicedo u. J. Gomez): "The Text REtrieval Conference (TREC) is a yearly workshop hosted by the U.S. government's National Institute of Standards and Technology (NIST) that fosters and supports research in information retrieval as well as speeding the transfer of technology between research labs and industry. Since 1992, TREC has provided the infrastructure necessary for large-scale evaluations of different text retrieval methodologies. TREC impact has been very important and its success has been mainly supported by its continuous adaptation to the emerging information retrieval needs. Not in vain, TREC has built evaluation benchmarks for more than 20 different retrieval problems such as Web retrieval, speech retrieval, or question-answering. The large and intense trajectory of annual TREC conferences has resulted in an immense bulk of documents reflecting the different eval uation and research efforts developed. This situation makes it difficult sometimes to observe clearly how research in information retrieval (IR) has evolved over the course of TREC. TREC: Experiment and Evaluation in Information Retrieval succeeds in organizing and condensing all this research into a manageable volume that describes TREC history and summarizes the main lessons learned. The book is organized into three parts. The first part is devoted to the description of TREC's origin and history, the test collections, and the evaluation methodology developed. The second part describes a selection of the major evaluation exercises (tracks), and the third part contains contributions from research groups that had a large and remarkable participation in TREC. Finally, Karen Spark Jones, one of the main promoters of research in IR, closes the book with an epilogue that analyzes the impact of TREC on this research field.
    ... TREC: Experiment and Evaluation in Information Retrieval is a reliable and comprehensive review of the TREC program and has been adopted by NIST as the official history of TREC (see http://trec.nist.gov). We were favorably surprised by the book. Well structured and written, chapters are self-contained and the existence of references to specialized and more detailed publications is continuous, which makes it easier to expand into the different aspects analyzed in the text. This book succeeds in compiling TREC evolution from its inception in 1992 to 2003 in an adequate and manageable volume. Thanks to the impressive effort performed by the authors and their experience in the field, it can satiate the interests of a great variety of readers. While expert researchers in the IR field and IR-related industrial companies can use it as a reference manual, it seems especially useful for students and non-expert readers willing to approach this research area. Like NIST, we would recommend this reading to anyone who may be interested in textual information retrieval."
    LCSH
    Information storage and retrieval systems / Congresses
    Text REtrieval Conference
    RSWK
    Information Retrieval / Textverarbeitung / Aufsatzsammlung (BVB)
    Kongress / Information Retrieval / Kongress (GBV)
    Subject
    Information Retrieval / Textverarbeitung / Aufsatzsammlung (BVB)
    Kongress / Information Retrieval / Kongress (GBV)
    Information storage and retrieval systems / Congresses
    Text REtrieval Conference
  8. Keen, E.M.: Some aspects of proximity searching in text retrieval systems (1992) 0.03
    0.034654688 = product of:
      0.10396406 = sum of:
        0.040623292 = weight(_text_:retrieval in 6190) [ClassicSimilarity], result of:
          0.040623292 = score(doc=6190,freq=4.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.37811437 = fieldWeight in 6190, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=6190)
        0.041627884 = weight(_text_:use in 6190) [ClassicSimilarity], result of:
          0.041627884 = score(doc=6190,freq=4.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.3827611 = fieldWeight in 6190, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0625 = fieldNorm(doc=6190)
        0.021712884 = weight(_text_:of in 6190) [ClassicSimilarity], result of:
          0.021712884 = score(doc=6190,freq=16.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.39093933 = fieldWeight in 6190, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=6190)
      0.33333334 = coord(3/9)
    
    Abstract
    Describes and evaluates the proximity search facilities in external online systems and in-house retrieval software. Discusses and illustrates capabilities, syntax and circumstances of use. Presents measurements of the overheads required by proximity for storage, record input time and search time. The search strategy narrowing effect of proximity is illustrated by recall and precision test results. Usage and problems lead to a number of design ideas for better implementation: some based on existing Boolean strategies, one on the use of weighted proximity to automatically produce ranked output. A comparison of Boolean, quorum and proximate term pairs distance is included
    Source
    Journal of information science. 18(1992), S.89-98
  9. Blair, D.C.; Maron, M.E.: ¬An evaluation of retrieval effectiveness for a full-text document-retrieval system (1985) 0.03
    0.03385633 = product of:
      0.10156898 = sum of:
        0.07181252 = weight(_text_:retrieval in 1345) [ClassicSimilarity], result of:
          0.07181252 = score(doc=1345,freq=8.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.6684181 = fieldWeight in 1345, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.078125 = fieldNorm(doc=1345)
        0.013570553 = weight(_text_:of in 1345) [ClassicSimilarity], result of:
          0.013570553 = score(doc=1345,freq=4.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.24433708 = fieldWeight in 1345, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.078125 = fieldNorm(doc=1345)
        0.016185908 = product of:
          0.04855772 = sum of:
            0.04855772 = weight(_text_:29 in 1345) [ClassicSimilarity], result of:
              0.04855772 = score(doc=1345,freq=2.0), product of:
                0.12493842 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.035517205 = queryNorm
                0.38865322 = fieldWeight in 1345, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1345)
          0.33333334 = coord(1/3)
      0.33333334 = coord(3/9)
    
    Footnote
    Vgl. auch : Salton, G.: Another look ... Comm. ACM 29(1986) S.S.648-656; Blair, D.C.: Full text retrieval ... Int. Class. 13(1986) S.18-23: Blair, D.C., M.E. Maron: Full-text information retrieval ... Inf. proc. man. 26(1990) S.437-447.
    Source
    Communications of the Association for Computing Machinery. 28(1985), S.280-299
  10. Yerbury, H.; Parker, J.: Novice searchers' use of familiar structures in searching bibliographic information retrieval systems (1998) 0.03
    0.033521123 = product of:
      0.10056337 = sum of:
        0.03731488 = weight(_text_:retrieval in 2874) [ClassicSimilarity], result of:
          0.03731488 = score(doc=2874,freq=6.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.34732026 = fieldWeight in 2874, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=2874)
        0.044153035 = weight(_text_:use in 2874) [ClassicSimilarity], result of:
          0.044153035 = score(doc=2874,freq=8.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.40597942 = fieldWeight in 2874, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=2874)
        0.01909546 = weight(_text_:of in 2874) [ClassicSimilarity], result of:
          0.01909546 = score(doc=2874,freq=22.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.34381276 = fieldWeight in 2874, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2874)
      0.33333334 = coord(3/9)
    
    Abstract
    Reports results of a study of the use of metaphors as problem solving mechanisms by novice searchers of bibliographic databases. Metaphors provide a framework or 'familiar structure' of credible associations within which relationships in other domains may be considered. 28 students taking an undergraduate course in information retrieval at Sydney University of Technology, were recorded as they 'talked through' a search on a bibliographic retrieval system. The transcripts were analyzed using conventional methods and the NUDIST software package for qualitative research. A range of metaphors was apparent from the language use by students in the search process. Those which predominated were: a journey; human interaction; a building or matching process; a problem solving process, and a search for a quantity. Many of the studentes experiencing the interaction as a problem solving process or a search for quantity perceived the outcomes as successful. Concludes that when memory for operating methods and procedures is incomplete an unconscious approach through the use of a conceptual system which is consonant with the task at hand may also lead to success in bibliographic searching
    Source
    Journal of information science. 24(1998) no.4, S.207-214
  11. Wan, T.-L.; Evens, M.; Wan, Y.-W.; Pao, Y.-Y.: Experiments with automatic indexing and a relational thesaurus in a Chinese information retrieval system (1997) 0.03
    0.03280385 = product of:
      0.098411545 = sum of:
        0.056202184 = weight(_text_:retrieval in 956) [ClassicSimilarity], result of:
          0.056202184 = score(doc=956,freq=10.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.5231199 = fieldWeight in 956, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=956)
        0.025755936 = weight(_text_:use in 956) [ClassicSimilarity], result of:
          0.025755936 = score(doc=956,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.23682132 = fieldWeight in 956, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0546875 = fieldNorm(doc=956)
        0.016453419 = weight(_text_:of in 956) [ClassicSimilarity], result of:
          0.016453419 = score(doc=956,freq=12.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.29624295 = fieldWeight in 956, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=956)
      0.33333334 = coord(3/9)
    
    Abstract
    This article describes a series of experiments with an interactive Chinese information retrieval system named CIRS and an interactive relational thesaurus. 2 important issues have been explored: whether thesauri enhance the retrieval effectiveness of Chinese documents, and whether automatic indexing can complete with manual indexing in a Chinese information retrieval system. Recall and precision are used to measure and evaluate the effectiveness of the system. Statistical analysis of the recall and precision measures suggest that the use of the relational thesaurus does improve the retrieval effectiveness both in the automatic indexing environment and in the manual indexing environment and that automatic indexing is at least as good as manual indexing
    Source
    Journal of the American Society for Information Science. 48(1997) no.12, S.1086-1096
  12. Rijsbergen, C.J. van: ¬A test for the separation of relevant and non-relevant documents in experimental retrieval collections (1973) 0.03
    0.03248416 = product of:
      0.09745248 = sum of:
        0.028725008 = weight(_text_:retrieval in 5002) [ClassicSimilarity], result of:
          0.028725008 = score(doc=5002,freq=2.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.26736724 = fieldWeight in 5002, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=5002)
        0.017165542 = weight(_text_:of in 5002) [ClassicSimilarity], result of:
          0.017165542 = score(doc=5002,freq=10.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.3090647 = fieldWeight in 5002, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=5002)
        0.051561933 = product of:
          0.0773429 = sum of:
            0.038846172 = weight(_text_:29 in 5002) [ClassicSimilarity], result of:
              0.038846172 = score(doc=5002,freq=2.0), product of:
                0.12493842 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.035517205 = queryNorm
                0.31092256 = fieldWeight in 5002, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5002)
            0.03849673 = weight(_text_:22 in 5002) [ClassicSimilarity], result of:
              0.03849673 = score(doc=5002,freq=2.0), product of:
                0.1243752 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035517205 = queryNorm
                0.30952093 = fieldWeight in 5002, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5002)
          0.6666667 = coord(2/3)
      0.33333334 = coord(3/9)
    
    Abstract
    Many retrievalexperiments are intended to discover ways of improving performance, taking the results obtained with some particular technique as a baseline. The fact that substantial alterations to a system often have little or no effect on particular collections is puzzling. This may be due to the initially poor seperation of relevant and non-relevant documents. The paper presents a procedure for characterizing this seperation for a collection, which can be used to show whether proposed modifications of the base system are likely to be useful.
    Date
    19. 3.1996 11:22:12
    Source
    Journal of documentation. 29(1973) no.3, S.251-257
  13. Drabenstott, K.M.; Weller, M.S.: ¬A comparative approach to system evaluation : delegating control of retrieval tests to an experimental online system (1996) 0.03
    0.03167449 = product of:
      0.09502347 = sum of:
        0.05026876 = weight(_text_:retrieval in 7435) [ClassicSimilarity], result of:
          0.05026876 = score(doc=7435,freq=8.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.46789268 = fieldWeight in 7435, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=7435)
        0.025755936 = weight(_text_:use in 7435) [ClassicSimilarity], result of:
          0.025755936 = score(doc=7435,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.23682132 = fieldWeight in 7435, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0546875 = fieldNorm(doc=7435)
        0.018998774 = weight(_text_:of in 7435) [ClassicSimilarity], result of:
          0.018998774 = score(doc=7435,freq=16.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.34207192 = fieldWeight in 7435, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=7435)
      0.33333334 = coord(3/9)
    
    Abstract
    Describes the comparative approach to system evaluation used in this research project which delegated the administartion of an online retrieval test to an experimental online catalogue to produce data for evaluating the effectiveness of a new subject access design. Describes the methods enlisted to sort out problem test administration, e.g. to identify out-of-scope queries, incomplete system administration, and suspect post-search questionnaire responses. Covers how w the researchers handled problem search administrations and what actions they would use to reduce or eliminate the occurrence of such administrations in future online retrieval tests that delegate control of retrieval tests to online systems
    Source
    Global complexity: information, chaos and control. Proceedings of the 59th Annual Meeting of the American Society for Information Science, ASIS'96, Baltimore, Maryland, 21-24 Oct 1996. Ed.: S. Hardin
  14. Hersh, W.R.; Hickam, D.H.: ¬An evaluation of interactive Boolean and natural language searching with an online medical textbook (1995) 0.03
    0.031659424 = product of:
      0.09497827 = sum of:
        0.043534026 = weight(_text_:retrieval in 2651) [ClassicSimilarity], result of:
          0.043534026 = score(doc=2651,freq=6.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.40520695 = fieldWeight in 2651, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2651)
        0.0364244 = weight(_text_:use in 2651) [ClassicSimilarity], result of:
          0.0364244 = score(doc=2651,freq=4.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.33491597 = fieldWeight in 2651, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2651)
        0.01501985 = weight(_text_:of in 2651) [ClassicSimilarity], result of:
          0.01501985 = score(doc=2651,freq=10.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.2704316 = fieldWeight in 2651, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2651)
      0.33333334 = coord(3/9)
    
    Abstract
    Few studies have compared the interactive use of Boolean and natural language search systems. Studies the use of 3 retrieval systems by senior medical students searching on queries generated by actual physicians in a clinical setting. The searchers were randomized to search on 2 or 3 different retrieval systems: a Boolean system, a word-based natural language system, and a concept-based natural language system. Results showed no statistically significant differences in recall or precision among the 3 systems. Likewise, there is no user preference for any system over the other. The study revealed problems with traditional measures of retrieval evaluation when applied to the interactive search setting
    Source
    Journal of the American Society for Information Science. 46(1995) no.7, S.478-489
  15. Wildemuth, B.; Freund, L.; Toms, E.G.: Untangling search task complexity and difficulty in the context of interactive information retrieval studies (2014) 0.03
    0.031284023 = product of:
      0.070389055 = sum of:
        0.025389558 = weight(_text_:retrieval in 1786) [ClassicSimilarity], result of:
          0.025389558 = score(doc=1786,freq=4.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.23632148 = fieldWeight in 1786, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1786)
        0.018397098 = weight(_text_:use in 1786) [ClassicSimilarity], result of:
          0.018397098 = score(doc=1786,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.1691581 = fieldWeight in 1786, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1786)
        0.018582245 = weight(_text_:of in 1786) [ClassicSimilarity], result of:
          0.018582245 = score(doc=1786,freq=30.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.33457235 = fieldWeight in 1786, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1786)
        0.008020152 = product of:
          0.024060456 = sum of:
            0.024060456 = weight(_text_:22 in 1786) [ClassicSimilarity], result of:
              0.024060456 = score(doc=1786,freq=2.0), product of:
                0.1243752 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035517205 = queryNorm
                0.19345059 = fieldWeight in 1786, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1786)
          0.33333334 = coord(1/3)
      0.44444445 = coord(4/9)
    
    Abstract
    Purpose - One core element of interactive information retrieval (IIR) experiments is the assignment of search tasks. The purpose of this paper is to provide an analytical review of current practice in developing those search tasks to test, observe or control task complexity and difficulty. Design/methodology/approach - Over 100 prior studies of IIR were examined in terms of how each defined task complexity and/or difficulty (or related concepts) and subsequently interpreted those concepts in the development of the assigned search tasks. Findings - Search task complexity is found to include three dimensions: multiplicity of subtasks or steps, multiplicity of facets, and indeterminability. Search task difficulty is based on an interaction between the search task and the attributes of the searcher or the attributes of the search situation. The paper highlights the anomalies in our use of these two concepts, concluding with suggestions for future methodological research related to search task complexity and difficulty. Originality/value - By analyzing and synthesizing current practices, this paper provides guidance for future experiments in IIR that involve these two constructs.
    Content
    Beitrag in einem Special Issue: Festschrift in honour of Nigel Ford
    Date
    6. 4.2015 19:31:22
    Source
    Journal of documentation. 70(2014) no.6, S.1118-1140
  16. MacCall, S.L.; Cleveland, A.D.: ¬A relevance-based quantitative measure for Internet information retrieval evaluation (1999) 0.03
    0.031093273 = product of:
      0.093279816 = sum of:
        0.048173305 = weight(_text_:retrieval in 6689) [ClassicSimilarity], result of:
          0.048173305 = score(doc=6689,freq=10.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.44838852 = fieldWeight in 6689, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=6689)
        0.022076517 = weight(_text_:use in 6689) [ClassicSimilarity], result of:
          0.022076517 = score(doc=6689,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.20298971 = fieldWeight in 6689, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=6689)
        0.02302999 = weight(_text_:of in 6689) [ClassicSimilarity], result of:
          0.02302999 = score(doc=6689,freq=32.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.41465378 = fieldWeight in 6689, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=6689)
      0.33333334 = coord(3/9)
    
    Abstract
    An important indicator of a maturating Internet is the development of metrics for its evaluation as a practical tool for enduser information retrieval. However, the Internet presents specific problems for traditional IR measures, such as the need to deal with the variety of classes of retrieval tools. This paper presents a metric for comparing the performance of common classes of Internet information retrieval tool, including human indexed catalogs of web resources and automatically indexed databases of web pages. The metric uses a relevance-based quantitative measure to compare the performance of endusers using these Internet information retrieval tools. The benefit of the proposed metric is that it is relevance-based (using enduser relevance judgments), and it facilitates the comparison of the performance of different classes of IIR tools
    Series
    Proceedings of the American Society for Information Science; vol36
    Source
    Knowledge: creation, organization and use. Proceedings of the 62nd Annual Meeting of the American Society for Information Science, 31.10.-4.11.1999. Ed.: L. Woods
  17. Dunlop, M.D.; Johnson, C.W.; Reid, J.: Exploring the layers of information retrieval evaluation (1998) 0.03
    0.030826038 = product of:
      0.09247811 = sum of:
        0.05026876 = weight(_text_:retrieval in 3762) [ClassicSimilarity], result of:
          0.05026876 = score(doc=3762,freq=8.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.46789268 = fieldWeight in 3762, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3762)
        0.025755936 = weight(_text_:use in 3762) [ClassicSimilarity], result of:
          0.025755936 = score(doc=3762,freq=2.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.23682132 = fieldWeight in 3762, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3762)
        0.016453419 = weight(_text_:of in 3762) [ClassicSimilarity], result of:
          0.016453419 = score(doc=3762,freq=12.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.29624295 = fieldWeight in 3762, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3762)
      0.33333334 = coord(3/9)
    
    Abstract
    Presents current work on modelling interactive information retrieval systems and users' interactions with them. Analyzes the papers in this special issue in the context of evaluation in information retrieval (IR) by examining the different layers at which IR use could be evaluated. IR poses the double evaluation problem of evaluating both the underlying system effectiveness and the overall ability of the system to aid users. The papers look at different issues in combining human-computer interaction (HCI) research with IR research and provide insights into the problem of evaluating the information seeking process
    Footnote
    Contribution to a special section of articles related to human-computer interaction and information retrieval
  18. Kwok, K.-L.: Ten years of ad hoc retrieval at TREC using PIRCS (2005) 0.03
    0.030624343 = product of:
      0.09187303 = sum of:
        0.060934942 = weight(_text_:retrieval in 5090) [ClassicSimilarity], result of:
          0.060934942 = score(doc=5090,freq=4.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.5671716 = fieldWeight in 5090, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.09375 = fieldNorm(doc=5090)
        0.011514995 = weight(_text_:of in 5090) [ClassicSimilarity], result of:
          0.011514995 = score(doc=5090,freq=2.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.20732689 = fieldWeight in 5090, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.09375 = fieldNorm(doc=5090)
        0.019423088 = product of:
          0.058269262 = sum of:
            0.058269262 = weight(_text_:29 in 5090) [ClassicSimilarity], result of:
              0.058269262 = score(doc=5090,freq=2.0), product of:
                0.12493842 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.035517205 = queryNorm
                0.46638384 = fieldWeight in 5090, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=5090)
          0.33333334 = coord(1/3)
      0.33333334 = coord(3/9)
    
    Date
    29. 3.1996 18:16:49
    Source
    TREC: experiment and evaluation in information retrieval. Ed.: E.M. Voorhees, u. D.K. Harman
  19. Dalrymple, P.W.: Retrieval by reformulation in two library catalogs : toward a cognitive model of searching behavior (1990) 0.03
    0.030574655 = product of:
      0.09172396 = sum of:
        0.05026876 = weight(_text_:retrieval in 5089) [ClassicSimilarity], result of:
          0.05026876 = score(doc=5089,freq=2.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.46789268 = fieldWeight in 5089, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.109375 = fieldNorm(doc=5089)
        0.018998774 = weight(_text_:of in 5089) [ClassicSimilarity], result of:
          0.018998774 = score(doc=5089,freq=4.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.34207192 = fieldWeight in 5089, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.109375 = fieldNorm(doc=5089)
        0.022456426 = product of:
          0.067369275 = sum of:
            0.067369275 = weight(_text_:22 in 5089) [ClassicSimilarity], result of:
              0.067369275 = score(doc=5089,freq=2.0), product of:
                0.1243752 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035517205 = queryNorm
                0.5416616 = fieldWeight in 5089, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=5089)
          0.33333334 = coord(1/3)
      0.33333334 = coord(3/9)
    
    Date
    22. 7.2006 18:43:54
    Source
    Journal of the American Society for Information Science. 41(1990) no.4, S.272-281
  20. Aldous, K.J.: ¬A system for the automatic retrieval of information from a specialist database (1996) 0.03
    0.030278161 = product of:
      0.09083448 = sum of:
        0.03731488 = weight(_text_:retrieval in 4078) [ClassicSimilarity], result of:
          0.03731488 = score(doc=4078,freq=6.0), product of:
            0.10743652 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.035517205 = queryNorm
            0.34732026 = fieldWeight in 4078, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=4078)
        0.031220913 = weight(_text_:use in 4078) [ClassicSimilarity], result of:
          0.031220913 = score(doc=4078,freq=4.0), product of:
            0.10875683 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.035517205 = queryNorm
            0.2870708 = fieldWeight in 4078, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=4078)
        0.022298694 = weight(_text_:of in 4078) [ClassicSimilarity], result of:
          0.022298694 = score(doc=4078,freq=30.0), product of:
            0.05554029 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035517205 = queryNorm
            0.4014868 = fieldWeight in 4078, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4078)
      0.33333334 = coord(3/9)
    
    Abstract
    Accessing useful information from a complex database requires knowledge of the structure of the database and an understanding of the methods of information retrieval. A means of overcoming this knowledge barrier to the use of narrow domain databases is proposed in which the user is required to enter only a series of terms which identify the required material. Describes a method which classifies terms according to their meaning in the context of the database and which uses this classification to access and execute models of code stored in the database to effect retrieval. Presents an implementation of the method using a database of technical information on the nature and use of fungicides. Initial results of trials with potential users indicate that the system can produce relevant resposes to queries expressed in this style. Since the code modules are part of the database, extensions may be easily implemented to handle most queries which users are likely to pose

Years

Languages

Types

  • a 419
  • s 15
  • m 10
  • el 8
  • r 8
  • x 3
  • d 1
  • p 1
  • More… Less…