Search (170 results, page 1 of 9)

  • × theme_ss:"Retrievalstudien"
  1. Lazonder, A.W.; Biemans, H.J.A.; Wopereis, I.G.J.H.: Differences between novice and experienced users in searching information on the World Wide Web (2000) 0.09
    0.09045595 = product of:
      0.1507599 = sum of:
        0.060152818 = weight(_text_:wide in 4598) [ClassicSimilarity], result of:
          0.060152818 = score(doc=4598,freq=2.0), product of:
            0.20479609 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046221454 = queryNorm
            0.29372054 = fieldWeight in 4598, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046875 = fieldNorm(doc=4598)
        0.072971776 = weight(_text_:web in 4598) [ClassicSimilarity], result of:
          0.072971776 = score(doc=4598,freq=10.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.48375595 = fieldWeight in 4598, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4598)
        0.017635323 = product of:
          0.035270646 = sum of:
            0.035270646 = weight(_text_:research in 4598) [ClassicSimilarity], result of:
              0.035270646 = score(doc=4598,freq=4.0), product of:
                0.13186905 = queryWeight, product of:
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046221454 = queryNorm
                0.2674672 = fieldWeight in 4598, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4598)
          0.5 = coord(1/2)
      0.6 = coord(3/5)
    
    Abstract
    Searching for information on the WWW basically comes down to locating an appropriate Web site and to retrieving relevant information from that site. This study examined the effect of a user's WWW experience on both phases of the search process. 35 students from 2 schools for Dutch pre-university education were observed while performing 3 search tasks. The results indicate that subjects with WWW-experience are more proficient in locating Web sites than are novice WWW-users. The observed differences were ascribed to the experts' superior skills in operating Web search engines. However, on tasks that required subjects to locate information on specific Web sites, the performance of experienced and novice users was equivalent - a result that is in line with hypertext research. Based on these findings, implications for training and supporting students in searching for information on the WWW are identified. Finally, the role of the subjects' level of domain expertise is discussed and directions for future research are proposed
  2. Agata, T.: ¬A measure for evaluating search engines on the World Wide Web : retrieval test with ESL (Expected Search Length) (1997) 0.07
    0.074229434 = product of:
      0.18557358 = sum of:
        0.120305635 = weight(_text_:wide in 3892) [ClassicSimilarity], result of:
          0.120305635 = score(doc=3892,freq=2.0), product of:
            0.20479609 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046221454 = queryNorm
            0.5874411 = fieldWeight in 3892, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.09375 = fieldNorm(doc=3892)
        0.065267935 = weight(_text_:web in 3892) [ClassicSimilarity], result of:
          0.065267935 = score(doc=3892,freq=2.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.43268442 = fieldWeight in 3892, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.09375 = fieldNorm(doc=3892)
      0.4 = coord(2/5)
    
  3. ¬The Eleventh Text Retrieval Conference, TREC 2002 (2003) 0.05
    0.050745755 = product of:
      0.12686439 = sum of:
        0.04351196 = weight(_text_:web in 4049) [ClassicSimilarity], result of:
          0.04351196 = score(doc=4049,freq=2.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.2884563 = fieldWeight in 4049, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0625 = fieldNorm(doc=4049)
        0.08335243 = sum of:
          0.033253483 = weight(_text_:research in 4049) [ClassicSimilarity], result of:
            0.033253483 = score(doc=4049,freq=2.0), product of:
              0.13186905 = queryWeight, product of:
                2.8529835 = idf(docFreq=6931, maxDocs=44218)
                0.046221454 = queryNorm
              0.2521705 = fieldWeight in 4049, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.8529835 = idf(docFreq=6931, maxDocs=44218)
                0.0625 = fieldNorm(doc=4049)
          0.050098952 = weight(_text_:22 in 4049) [ClassicSimilarity], result of:
            0.050098952 = score(doc=4049,freq=2.0), product of:
              0.16185966 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046221454 = queryNorm
              0.30952093 = fieldWeight in 4049, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=4049)
      0.4 = coord(2/5)
    
    Abstract
    Proceedings of the llth TREC-conference held in Gaithersburg, Maryland (USA), November 19-22, 2002. Aim of the conference was discussion an retrieval and related information-seeking tasks for large test collection. 93 research groups used different techniques, for information retrieval from the same large database. This procedure makes it possible to compare the results. The tasks are: Cross-language searching, filtering, interactive searching, searching for novelty, question answering, searching for video shots, and Web searching.
  4. Wu, C.-J.: Experiments on using the Dublin Core to reduce the retrieval error ratio (1998) 0.04
    0.043300506 = product of:
      0.10825126 = sum of:
        0.07017829 = weight(_text_:wide in 5201) [ClassicSimilarity], result of:
          0.07017829 = score(doc=5201,freq=2.0), product of:
            0.20479609 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046221454 = queryNorm
            0.342674 = fieldWeight in 5201, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5201)
        0.038072966 = weight(_text_:web in 5201) [ClassicSimilarity], result of:
          0.038072966 = score(doc=5201,freq=2.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.25239927 = fieldWeight in 5201, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5201)
      0.4 = coord(2/5)
    
    Abstract
    In order to test the power of metadata on information retrieval, an experiment was designed and conducted on a group of 7 graduate students using the Dublin Core as the cataloguing metadata. Results show that, on average, the retrieval error rate is only 2.9 per cent for the MES system (http://140.136.85.194), which utilizes the Dublin Core to describe the documents on the World Wide Web, in contrast to 20.7 per cent for the 7 famous search engines including HOTBOT, GAIS, LYCOS, EXCITE, INFOSEEK, YAHOO, and OCTOPUS. The very low error rate indicates that the users can use the information of the Dublin Core to decide whether to retrieve the documents or not
  5. Bar-Ilan, J.: ¬The Web as an information source on informetrics? : A content analysis (2000) 0.04
    0.041909147 = product of:
      0.104772866 = sum of:
        0.09230281 = weight(_text_:web in 4587) [ClassicSimilarity], result of:
          0.09230281 = score(doc=4587,freq=16.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.6119082 = fieldWeight in 4587, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4587)
        0.012470056 = product of:
          0.024940113 = sum of:
            0.024940113 = weight(_text_:research in 4587) [ClassicSimilarity], result of:
              0.024940113 = score(doc=4587,freq=2.0), product of:
                0.13186905 = queryWeight, product of:
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046221454 = queryNorm
                0.18912788 = fieldWeight in 4587, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4587)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    This article addresses the question of whether the Web can serve as an information source for research. Specifically, it analyzes by way of content analysis the Web pages retrieved by the major search engines on a particular date (June 7, 1998), as a result of the query 'informetrics OR informetric'. In 807 out of the 942 retrieved pages, the search terms were mentioned in the context of information science. Over 70% of the pages contained only indirect information on the topic, in the form of hypertext links and bibliographical references without annotation. The bibliographical references extracted from the Web pages were analyzed, and lists of most productive authors, most cited authors, works, and sources were compiled. The list of reference obtained from the Web was also compared to data retrieved from commercial databases. For most cases, the list of references extracted from the Web outperformed the commercial, bibliographic databases. The results of these comparisons indicate that valuable, freely available data is hidden in the Web waiting to be extracted from the millions of Web pages
  6. Voorhees, E.M.; Harman, D.K.: ¬The Text REtrieval Conference (2005) 0.04
    0.04003495 = product of:
      0.06672491 = sum of:
        0.035089146 = weight(_text_:wide in 5082) [ClassicSimilarity], result of:
          0.035089146 = score(doc=5082,freq=2.0), product of:
            0.20479609 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046221454 = queryNorm
            0.171337 = fieldWeight in 5082, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5082)
        0.019036483 = weight(_text_:web in 5082) [ClassicSimilarity], result of:
          0.019036483 = score(doc=5082,freq=2.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.12619963 = fieldWeight in 5082, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5082)
        0.012599284 = product of:
          0.025198568 = sum of:
            0.025198568 = weight(_text_:research in 5082) [ClassicSimilarity], result of:
              0.025198568 = score(doc=5082,freq=6.0), product of:
                0.13186905 = queryWeight, product of:
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046221454 = queryNorm
                0.19108781 = fieldWeight in 5082, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=5082)
          0.5 = coord(1/2)
      0.6 = coord(3/5)
    
    Abstract
    Text retrieval technology targets a problem that is all too familiar: finding relevant information in large stores of electronic documents. The problem is an old one, with the first research conference devoted to the subject held in 1958 [11]. Since then the problem has continued to grow as more information is created in electronic form and more people gain electronic access. The advent of the World Wide Web, where anyone can publish so everyone must search, is a graphic illustration of the need for effective retrieval technology. The Text REtrieval Conference (TREC) is a workshop series designed to build the infrastructure necessary for the large-scale evaluation of text retrieval technology, thereby accelerating its transfer into the commercial sector. The series is sponsored by the U.S. National Institute of Standards and Technology (NIST) and the U.S. Department of Defense. At the time of this writing, there have been twelve TREC workshops and preparations for the thirteenth workshop are under way. Participants in the workshops have been drawn from the academic, commercial, and government sectors, and have included representatives from more than twenty different countries. These collective efforts have accomplished a great deal: a variety of large test collections have been built for both traditional ad hoc retrieval and related tasks such as cross-language retrieval, speech retrieval, and question answering; retrieval effectiveness has approximately doubled; and many commercial retrieval systems now contain technology first developed in TREC.
    This book chronicles the evolution of retrieval systems over the course of TREC. To be sure, there has already been a wealth of information written about TREC. Each conference has produced a proceedings containing general overviews of the various tasks, papers written by the individual participants, and evaluation results.1 Reports on expanded versions of TREC experiments frequently appear in the wider information retrieval literature. There also have been special issues of journals devoted to particular TRECs [3; 13] and particular TREC tasks [6; 4]. No single volume could hope to be a comprehensive record of all TREC-related research. Instead, this book looks to distill the overabundance of detail into a manageable whole that summarizes the main lessons learned from TREC. The book consists of three main parts. The first part contains introductory and descriptive chapters on TREC's history, the major products of TREC (the test collections), and the retrieval evaluation methodology. Part II includes chapters describing the major TREC ''tracks,'' evaluations of special subtopics such as cross-language retrieval and question answering. Part III contains contributions from research groups that have participated in TREC. The epilogue to the book is written by Karen Sparck Jones, who reflects on the impact TREC has had on the information retrieval field. The structure of this introductory chapter is similar to that of the book as a whole. The chapter begins with a short history of TREC; expanded descriptions of specific aspects of the history are included in subsequent chapters to make those chapters self-contained. Section 1.2 describes TREC's track structure, which has been responsible for the growth of TREC and allows TREC to adapt to changing needs. The final section lists both the major accomplishments of TREC and some remaining challenges.
  7. MacFarlane, A.: Evaluation of web search for the information practitioner (2007) 0.04
    0.039524574 = product of:
      0.09881143 = sum of:
        0.08634137 = weight(_text_:web in 817) [ClassicSimilarity], result of:
          0.08634137 = score(doc=817,freq=14.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.57238775 = fieldWeight in 817, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=817)
        0.012470056 = product of:
          0.024940113 = sum of:
            0.024940113 = weight(_text_:research in 817) [ClassicSimilarity], result of:
              0.024940113 = score(doc=817,freq=2.0), product of:
                0.13186905 = queryWeight, product of:
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046221454 = queryNorm
                0.18912788 = fieldWeight in 817, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046875 = fieldNorm(doc=817)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Purpose - The aim of the paper is to put forward a structured mechanism for web search evaluation. The paper seeks to point to useful scientific research and show how information practitioners can use these methods in evaluation of search on the web for their users. Design/methodology/approach - The paper puts forward an approach which utilizes traditional laboratory-based evaluation measures such as average precision/precision at N documents, augmented with diagnostic measures such as link broken, etc., which are used to show why precision measures are depressed as well as the quality of the search engines crawling mechanism. Findings - The paper shows how to use diagnostic measures in conjunction with precision in order to evaluate web search. Practical implications - The methodology presented in this paper will be useful to any information professional who regularly uses web search as part of their information seeking and needs to evaluate web search services. Originality/value - The paper argues that the use of diagnostic measures is essential in web search, as precision measures on their own do not allow a searcher to understand why search results differ between search engines.
  8. Khan, K.; Locatis, C.: Searching through cyberspace : the effects of link display and link density on information retrieval from hypertext on the World Wide Web (1998) 0.04
    0.037114717 = product of:
      0.09278679 = sum of:
        0.060152818 = weight(_text_:wide in 446) [ClassicSimilarity], result of:
          0.060152818 = score(doc=446,freq=2.0), product of:
            0.20479609 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046221454 = queryNorm
            0.29372054 = fieldWeight in 446, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046875 = fieldNorm(doc=446)
        0.032633968 = weight(_text_:web in 446) [ClassicSimilarity], result of:
          0.032633968 = score(doc=446,freq=2.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.21634221 = fieldWeight in 446, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=446)
      0.4 = coord(2/5)
    
  9. Griesbaum, J.: Evaluierung hybrider Suchsysteme im WWW (2000) 0.04
    0.037114717 = product of:
      0.09278679 = sum of:
        0.060152818 = weight(_text_:wide in 2482) [ClassicSimilarity], result of:
          0.060152818 = score(doc=2482,freq=2.0), product of:
            0.20479609 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046221454 = queryNorm
            0.29372054 = fieldWeight in 2482, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046875 = fieldNorm(doc=2482)
        0.032633968 = weight(_text_:web in 2482) [ClassicSimilarity], result of:
          0.032633968 = score(doc=2482,freq=2.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.21634221 = fieldWeight in 2482, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=2482)
      0.4 = coord(2/5)
    
    Abstract
    Der Ausgangspunkt dieser Arbeit ist die Suchproblematik im World Wide Web. Suchmaschinen sind einerseits unverzichtbar für erfolgreiches Information Retrieval, andererseits wird ihnen eine mäßige Leistungsfähigkeit vorgeworfen. Das Thema dieser Arbeit ist die Untersuchung der Retrievaleffektivität deutschsprachiger Suchmaschinen. Es soll festgestellt werden, welche Retrievaleffektivität Nutzer derzeit erwarten können. Ein Ansatz, um die Retrievaleffektivität von Suchmaschinen zu erhöhen besteht darin, redaktionell von Menschen erstellte und automatisch generierte Suchergebnisse in einer Trefferliste zu vermengen. Ziel dieser Arbeit ist es, die Retrievaleffektivität solcher hybrider Systeme im Vergleich zu rein roboterbasierten Suchmaschinen zu evaluieren. Zunächst werden hierzu die grundlegenden Problembereiche bei der Evaluation von Retrievalsystemen analysiert. In Anlehnung an die von Tague-Sutcliff vorgeschlagene Methodik wird unter Beachtung der webspezifischen Besonderheiten eine mögliche Vorgehensweise erschlossen. Darauf aufbauend wird das konkrete Setting für die Durchführung der Evaluation erarbeitet und ein Retrievaleffektivitätstest bei den Suchmaschinen Lycos.de, AItaVista.de und QualiGo durchgeführt.
  10. Harman, D.: ¬The Text REtrieval Conferences (TRECs) : providing a test-bed for information retrieval systems (1998) 0.04
    0.036301136 = product of:
      0.09075284 = sum of:
        0.07017829 = weight(_text_:wide in 1314) [ClassicSimilarity], result of:
          0.07017829 = score(doc=1314,freq=2.0), product of:
            0.20479609 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046221454 = queryNorm
            0.342674 = fieldWeight in 1314, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1314)
        0.020574544 = product of:
          0.041149087 = sum of:
            0.041149087 = weight(_text_:research in 1314) [ClassicSimilarity], result of:
              0.041149087 = score(doc=1314,freq=4.0), product of:
                0.13186905 = queryWeight, product of:
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046221454 = queryNorm
                0.31204507 = fieldWeight in 1314, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1314)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The Text REtrieval Conference (TREC) workshop series encourages research in information retrieval from large text applications by providing a large test collection, uniform scoring procedures and a forum for organizations interested in comparing their results. Now in its seventh year, the conference has become the major experimental effort in the field. Participants in the TREC conferences have examined a wide variety of retrieval techniques, including methods using automatic thesauri, sophisticated term weighting, natural language techniques, relevance feedback and advanced pattern matching. The TREC conference series is co-sponsored by the National Institute of Standards and Technology (NIST) and the Information Technology Office of the Defense Advanced Research Projects Agency (DARPA)
  11. Dresel, R.; Hörnig, D.; Kaluza, H.; Peter, A.; Roßmann, A.; Sieber, W.: Evaluation deutscher Web-Suchwerkzeuge : Ein vergleichender Retrievaltest (2001) 0.03
    0.03463387 = product of:
      0.08658468 = sum of:
        0.061535202 = weight(_text_:web in 261) [ClassicSimilarity], result of:
          0.061535202 = score(doc=261,freq=4.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.4079388 = fieldWeight in 261, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0625 = fieldNorm(doc=261)
        0.025049476 = product of:
          0.050098952 = sum of:
            0.050098952 = weight(_text_:22 in 261) [ClassicSimilarity], result of:
              0.050098952 = score(doc=261,freq=2.0), product of:
                0.16185966 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046221454 = queryNorm
                0.30952093 = fieldWeight in 261, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=261)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Die deutschen Suchmaschinen, Abacho, Acoon, Fireball und Lycos sowie die Web-Kataloge Web.de und Yahoo! werden einem Qualitätstest nach relativem Recall, Precision und Availability unterzogen. Die Methoden der Retrievaltests werden vorgestellt. Im Durchschnitt werden bei einem Cut-Off-Wert von 25 ein Recall von rund 22%, eine Precision von knapp 19% und eine Verfügbarkeit von 24% erreicht
  12. Kantor, P.; Kim, M.H.; Ibraev, U.; Atasoy, K.: Estimating the number of relevant documents in enormous collections (1999) 0.03
    0.03092893 = product of:
      0.07732233 = sum of:
        0.05012735 = weight(_text_:wide in 6690) [ClassicSimilarity], result of:
          0.05012735 = score(doc=6690,freq=2.0), product of:
            0.20479609 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046221454 = queryNorm
            0.24476713 = fieldWeight in 6690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6690)
        0.027194975 = weight(_text_:web in 6690) [ClassicSimilarity], result of:
          0.027194975 = score(doc=6690,freq=2.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.18028519 = fieldWeight in 6690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6690)
      0.4 = coord(2/5)
    
    Abstract
    In assessing information retrieval systems, it is important to know not only the precision of the retrieved set, but also to compare the number of retrieved relevant items to the total number of relevant items. For large collections, such as the TREC test collections, or the World Wide Web, it is not possible to enumerate the entire set of relevant documents. If the retrieved documents are evaluated, a variant of the statistical "capture-recapture" method can be used to estimate the total number of relevant documents, providing the several retrieval systems used are sufficiently independent. We show that the underlying signal detection model supporting such an analysis can be extended in two ways. First, assuming that there are two distinct performance characteristics (corresponding to the chance of retrieving a relevant, and retrieving a given non-relevant document), we show that if there are three or more independent systems available it is possible to estimate the number of relevant documents without actually having to decide whether each individual document is relevant. We report applications of this 3-system method to the TREC data, leading to the conclusion that the independence assumptions are not satisfied. We then extend the model to a multi-system, multi-problem model, and show that it is possible to include statistical dependencies of all orders in the model, and determine the number of relevant documents for each of the problems in the set. Application to the TREC setting will be presented
  13. Mansourian, Y.; Ford, N.: Search persistence and failure on the web : a "bounded rationality" and "satisficing" analysis (2007) 0.03
    0.027967116 = product of:
      0.06991779 = sum of:
        0.05329105 = weight(_text_:web in 841) [ClassicSimilarity], result of:
          0.05329105 = score(doc=841,freq=12.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.35328537 = fieldWeight in 841, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=841)
        0.016626742 = product of:
          0.033253483 = sum of:
            0.033253483 = weight(_text_:research in 841) [ClassicSimilarity], result of:
              0.033253483 = score(doc=841,freq=8.0), product of:
                0.13186905 = queryWeight, product of:
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046221454 = queryNorm
                0.2521705 = fieldWeight in 841, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.03125 = fieldNorm(doc=841)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Purpose - This paper aims to examine our current knowledge of how searchers perceive and react to the possibility of missing potentially important information whilst searching the web is limited. The study reported here seeks to investigate such perceptions and reactions, and to explore the extent to which Simon's "bounded rationality" theory is useful in illuminating these issues. Design/methodology/approach - Totally 37 academic staff, research staff and research students in three university departments were interviewed about their web searching. The open-ended, semi-structured interviews were inductively analysed. Emergence of the concept of "good enough" searching prompted a further analysis to explore the extent to which the data could be interpreted in terms of Simon's concepts of "bounded rationality" and "satisficing". Findings - The results indicate that the risk of missing potentially important information was a matter of concern to the interviewees. Their estimations of the likely extent and importance of missed information affected decisions by individuals as to when to stop searching - decisions based on very different criteria, which map well onto Simon's concepts. On the basis of the interview data, the authors propose tentative categorizations of perceptions of the risk of missing information including "inconsequential" "tolerable" "damaging" and "disastrous" and search strategies including "perfunctory" "minimalist" "nervous" and "extensive". It is concluded that there is at least a prima facie case for bounded rationality and satisficing being considered as potentially useful concepts in our quest better to understand aspects of human information behaviour. Research limitations/implications - Although the findings are based on a relatively small sample and an exploratory qualitative analysis, it is argued that the study raises a number of interesting questions, and has implications for both the development of theory and practice in the areas of web searching and information literacy. Originality/value - The paper focuses on an aspect of web searching which has not to date been well explored. Whilst research has done much to illuminate searchers' perceptions of what they find on the web, we know relatively little of their perceptions of, and reactions to information that they fail to find. The study reported here provides some tentative models, based on empirical evidence, of these phenomena.
  14. Airio, E.: Who benefits from CLIR in web retrieval? (2008) 0.03
    0.027597502 = product of:
      0.068993755 = sum of:
        0.056523696 = weight(_text_:web in 2342) [ClassicSimilarity], result of:
          0.056523696 = score(doc=2342,freq=6.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.37471575 = fieldWeight in 2342, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=2342)
        0.012470056 = product of:
          0.024940113 = sum of:
            0.024940113 = weight(_text_:research in 2342) [ClassicSimilarity], result of:
              0.024940113 = score(doc=2342,freq=2.0), product of:
                0.13186905 = queryWeight, product of:
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046221454 = queryNorm
                0.18912788 = fieldWeight in 2342, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2342)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Purpose - The aim of the current paper is to test whether query translation is beneficial in web retrieval. Design/methodology/approach - The language pairs were Finnish-Swedish, English-German and Finnish-French. A total of 12-18 participants were recruited for each language pair. Each participant performed four retrieval tasks. The author's aim was to compare the performance of the translated queries with that of the target language queries. Thus, the author asked participants to formulate a source language query and a target language query for each task. The source language queries were translated into the target language utilizing a dictionary-based system. In English-German, also machine translation was utilized. The author used Google as the search engine. Findings - The results differed depending on the language pair. The author concluded that the dictionary coverage had an effect on the results. On average, the results of query-translation were better than in the traditional laboratory tests. Originality/value - This research shows that query translation in web is beneficial especially for users with moderate and non-active language skills. This is valuable information for developers of cross-language information retrieval systems.
  15. Pemberton, J.K.; Ojala, M.; Garman, N.: Head to head : searching the Web versus traditional services (1998) 0.03
    0.027424574 = product of:
      0.068561435 = sum of:
        0.04351196 = weight(_text_:web in 3572) [ClassicSimilarity], result of:
          0.04351196 = score(doc=3572,freq=2.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.2884563 = fieldWeight in 3572, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0625 = fieldNorm(doc=3572)
        0.025049476 = product of:
          0.050098952 = sum of:
            0.050098952 = weight(_text_:22 in 3572) [ClassicSimilarity], result of:
              0.050098952 = score(doc=3572,freq=2.0), product of:
                0.16185966 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046221454 = queryNorm
                0.30952093 = fieldWeight in 3572, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3572)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Source
    Online. 22(1998) no.3, S.24-26,28
  16. Wolff, C.: Leistungsvergleich der Retrievaloberflächen zwischen Web und klassischen Expertensystemen (2001) 0.03
    0.027356682 = product of:
      0.0683917 = sum of:
        0.053843305 = weight(_text_:web in 5870) [ClassicSimilarity], result of:
          0.053843305 = score(doc=5870,freq=4.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.35694647 = fieldWeight in 5870, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5870)
        0.014548399 = product of:
          0.029096797 = sum of:
            0.029096797 = weight(_text_:research in 5870) [ClassicSimilarity], result of:
              0.029096797 = score(doc=5870,freq=2.0), product of:
                0.13186905 = queryWeight, product of:
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046221454 = queryNorm
                0.22064918 = fieldWeight in 5870, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5870)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Die meisten Web-Auftritte der Hosts waren bisher für den Retrieval-Laien gedacht. Im Hintergrund steht dabei das Ziel: mehr Nutzung durch einfacheres Retrieval. Dieser Ansatz steht aber im Konflikt mit der wachsenden Datenmenge und Dokumentgröße, die eigentlich ein immer ausgefeilteres Retrieval verlangen. Häufig wird von Information Professionals die Kritik geäußert, dass die Webanwendungen einen Verlust an Relevanz bringen. Wie weit der Nutzer tatsächlich einen Kompromiss zwischen Relevanz und Vollständigkeit eingehen muss, soll in diesem Beitrag anhand verschiedener Host-Rechner quantifiziert werden
    Source
    Information Research & Content Management: Orientierung, Ordnung und Organisation im Wissensmarkt; 23. DGI-Online-Tagung der DGI und 53. Jahrestagung der Deutschen Gesellschaft für Informationswissenschaft und Informationspraxis e.V. DGI, Frankfurt am Main, 8.-10.5.2001. Proceedings. Hrsg.: R. Schmidt
  17. Mansourian, Y.; Ford, N.: Web searchers' attributions of success and failure: an empirical study (2007) 0.02
    0.022107538 = product of:
      0.055268843 = sum of:
        0.04351196 = weight(_text_:web in 840) [ClassicSimilarity], result of:
          0.04351196 = score(doc=840,freq=8.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.2884563 = fieldWeight in 840, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=840)
        0.011756882 = product of:
          0.023513764 = sum of:
            0.023513764 = weight(_text_:research in 840) [ClassicSimilarity], result of:
              0.023513764 = score(doc=840,freq=4.0), product of:
                0.13186905 = queryWeight, product of:
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046221454 = queryNorm
                0.17831147 = fieldWeight in 840, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.03125 = fieldNorm(doc=840)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Purpose - This paper reports the findings of a study designed to explore web searchers' perceptions of the causes of their search failure and success. In particular, it seeks to discover the extent to which the constructs locus of control and attribution theory might provide useful frameworks for understanding searchers' perceptions. Design/methodology/approach - A combination of inductive and deductive approaches were employed. Perceptions of failed and successful searches were derived from the inductive analysis of using open-ended qualitative interviews with a sample of 37 biologists at the University of Sheffield. These perceptions were classified into "internal" and "external" attributions, and the relationships between these categories and "successful" and "failed" searches were analysed deductively to test the extent to which they might be explainable using locus of control and attribution theory interpretive frameworks. Findings - All searchers were readily able to recall "successful" and "unsuccessful" searches. In a large majority of cases (82.4 per cent), they clearly attributed each search to either internal (e.g. ability or effort) or external (e.g. luck or information not being available) factors. The pattern of such relationships was analysed, and mapped onto those that would be predicted by locus of control and attribution theory. The authors conclude that the potential of these theoretical frameworks to illuminate one's understanding of web searching, and associated training, merits further systematic study. Research limitations/implications - The findings are based on a relatively small sample of academic and research staff in a particular subject area. Importantly, also, the study can at best provide a prima facie case for further systematic study since, although the patterns of attribution behaviour accord with those predictable by locus of control and attribution theory, data relating to the predictive elements of these theories (e.g. levels of confidence and achievement) were not available. This issue is discussed, and recommendations made for further work. Originality/value - The findings provide some empirical support for the notion that locus of control and attribution theory might - subject to the limitations noted above - be potentially useful theoretical frameworks for helping us better understand web-based information seeking. If so, they could have implications particularly for better understanding of searchers' motivations, and for the design and development of more effective search training programmes.
  18. Behnert, C.; Lewandowski, D.: ¬A framework for designing retrieval effectiveness studies of library information systems using human relevance assessments (2017) 0.02
    0.019540487 = product of:
      0.048851214 = sum of:
        0.038459502 = weight(_text_:web in 3700) [ClassicSimilarity], result of:
          0.038459502 = score(doc=3700,freq=4.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.25496176 = fieldWeight in 3700, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3700)
        0.010391714 = product of:
          0.020783428 = sum of:
            0.020783428 = weight(_text_:research in 3700) [ClassicSimilarity], result of:
              0.020783428 = score(doc=3700,freq=2.0), product of:
                0.13186905 = queryWeight, product of:
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046221454 = queryNorm
                0.15760657 = fieldWeight in 3700, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3700)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Purpose This paper demonstrates how to apply traditional information retrieval evaluation methods based on standards from the Text REtrieval Conference (TREC) and web search evaluation to all types of modern library information systems including online public access catalogs, discovery systems, and digital libraries that provide web search features to gather information from heterogeneous sources. Design/methodology/approach We apply conventional procedures from information retrieval evaluation to the library information system context considering the specific characteristics of modern library materials. Findings We introduce a framework consisting of five parts: (1) search queries, (2) search results, (3) assessors, (4) testing, and (5) data analysis. We show how to deal with comparability problems resulting from diverse document types, e.g., electronic articles vs. printed monographs and what issues need to be considered for retrieval tests in the library context. Practical implications The framework can be used as a guideline for conducting retrieval effectiveness studies in the library context. Originality/value Although a considerable amount of research has been done on information retrieval evaluation, and standards for conducting retrieval effectiveness studies do exist, to our knowledge this is the first attempt to provide a systematic framework for evaluating the retrieval effectiveness of twenty-first-century library information systems. We demonstrate which issues must be considered and what decisions must be made by researchers prior to a retrieval test.
  19. Sanderson, M.: ¬The Reuters test collection (1996) 0.02
    0.019425297 = product of:
      0.097126484 = sum of:
        0.097126484 = sum of:
          0.04702753 = weight(_text_:research in 6971) [ClassicSimilarity], result of:
            0.04702753 = score(doc=6971,freq=4.0), product of:
              0.13186905 = queryWeight, product of:
                2.8529835 = idf(docFreq=6931, maxDocs=44218)
                0.046221454 = queryNorm
              0.35662293 = fieldWeight in 6971, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                2.8529835 = idf(docFreq=6931, maxDocs=44218)
                0.0625 = fieldNorm(doc=6971)
          0.050098952 = weight(_text_:22 in 6971) [ClassicSimilarity], result of:
            0.050098952 = score(doc=6971,freq=2.0), product of:
              0.16185966 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046221454 = queryNorm
              0.30952093 = fieldWeight in 6971, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=6971)
      0.2 = coord(1/5)
    
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
  20. TREC: experiment and evaluation in information retrieval (2005) 0.02
    0.017470002 = product of:
      0.043675005 = sum of:
        0.023551539 = weight(_text_:web in 636) [ClassicSimilarity], result of:
          0.023551539 = score(doc=636,freq=6.0), product of:
            0.1508442 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046221454 = queryNorm
            0.15613155 = fieldWeight in 636, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.01953125 = fieldNorm(doc=636)
        0.020123467 = product of:
          0.040246934 = sum of:
            0.040246934 = weight(_text_:research in 636) [ClassicSimilarity], result of:
              0.040246934 = score(doc=636,freq=30.0), product of:
                0.13186905 = queryWeight, product of:
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.046221454 = queryNorm
                0.3052038 = fieldWeight in 636, product of:
                  5.477226 = tf(freq=30.0), with freq of:
                    30.0 = termFreq=30.0
                  2.8529835 = idf(docFreq=6931, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=636)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The Text REtrieval Conference (TREC), a yearly workshop hosted by the US government's National Institute of Standards and Technology, provides the infrastructure necessary for large-scale evaluation of text retrieval methodologies. With the goal of accelerating research in this area, TREC created the first large test collections of full-text documents and standardized retrieval evaluation. The impact has been significant; since TREC's beginning in 1992, retrieval effectiveness has approximately doubled. TREC has built a variety of large test collections, including collections for such specialized retrieval tasks as cross-language retrieval and retrieval of speech. Moreover, TREC has accelerated the transfer of research ideas into commercial systems, as demonstrated in the number of retrieval techniques developed in TREC that are now used in Web search engines. This book provides a comprehensive review of TREC research, summarizing the variety of TREC results, documenting the best practices in experimental information retrieval, and suggesting areas for further research. The first part of the book describes TREC's history, test collections, and retrieval methodology. Next, the book provides "track" reports -- describing the evaluations of specific tasks, including routing and filtering, interactive retrieval, and retrieving noisy text. The final part of the book offers perspectives on TREC from such participants as Microsoft Research, University of Massachusetts, Cornell University, University of Waterloo, City University of New York, and IBM. The book will be of interest to researchers in information retrieval and related technologies, including natural language processing.
    Content
    Enthält die Beiträge: 1. The Text REtrieval Conference - Ellen M. Voorhees and Donna K. Harman 2. The TREC Test Collections - Donna K. Harman 3. Retrieval System Evaluation - Chris Buckley and Ellen M. Voorhees 4. The TREC Ad Hoc Experiments - Donna K. Harman 5. Routing and Filtering - Stephen Robertson and Jamie Callan 6. The TREC Interactive Tracks: Putting the User into Search - Susan T. Dumais and Nicholas J. Belkin 7. Beyond English - Donna K. Harman 8. Retrieving Noisy Text - Ellen M. Voorhees and John S. Garofolo 9.The Very Large Collection and Web Tracks - David Hawking and Nick Craswell 10. Question Answering in TREC - Ellen M. Voorhees 11. The University of Massachusetts and a Dozen TRECs - James Allan, W. Bruce Croft and Jamie Callan 12. How Okapi Came to TREC - Stephen Robertson 13. The SMART Project at TREC - Chris Buckley 14. Ten Years of Ad Hoc Retrieval at TREC Using PIRCS - Kui-Lam Kwok 15. MultiText Experiments for TREC - Gordon V. Cormack, Charles L. A. Clarke, Christopher R. Palmer and Thomas R. Lynam 16. A Language-Modeling Approach to TREC - Djoerd Hiemstra and Wessel Kraaij 17. BM Research Activities at TREC - Eric W. Brown, David Carmel, Martin Franz, Abraham Ittycheriah, Tapas Kanungo, Yoelle Maarek, J. Scott McCarley, Robert L. Mack, John M. Prager, John R. Smith, Aya Soffer, Jason Y. Zien and Alan D. Marwick Epilogue: Metareflections on TREC - Karen Sparck Jones
    Footnote
    Rez. in: JASIST 58(2007) no.6, S.910-911 (J.L. Vicedo u. J. Gomez): "The Text REtrieval Conference (TREC) is a yearly workshop hosted by the U.S. government's National Institute of Standards and Technology (NIST) that fosters and supports research in information retrieval as well as speeding the transfer of technology between research labs and industry. Since 1992, TREC has provided the infrastructure necessary for large-scale evaluations of different text retrieval methodologies. TREC impact has been very important and its success has been mainly supported by its continuous adaptation to the emerging information retrieval needs. Not in vain, TREC has built evaluation benchmarks for more than 20 different retrieval problems such as Web retrieval, speech retrieval, or question-answering. The large and intense trajectory of annual TREC conferences has resulted in an immense bulk of documents reflecting the different eval uation and research efforts developed. This situation makes it difficult sometimes to observe clearly how research in information retrieval (IR) has evolved over the course of TREC. TREC: Experiment and Evaluation in Information Retrieval succeeds in organizing and condensing all this research into a manageable volume that describes TREC history and summarizes the main lessons learned. The book is organized into three parts. The first part is devoted to the description of TREC's origin and history, the test collections, and the evaluation methodology developed. The second part describes a selection of the major evaluation exercises (tracks), and the third part contains contributions from research groups that had a large and remarkable participation in TREC. Finally, Karen Spark Jones, one of the main promoters of research in IR, closes the book with an epilogue that analyzes the impact of TREC on this research field.
    ... TREC: Experiment and Evaluation in Information Retrieval is a reliable and comprehensive review of the TREC program and has been adopted by NIST as the official history of TREC (see http://trec.nist.gov). We were favorably surprised by the book. Well structured and written, chapters are self-contained and the existence of references to specialized and more detailed publications is continuous, which makes it easier to expand into the different aspects analyzed in the text. This book succeeds in compiling TREC evolution from its inception in 1992 to 2003 in an adequate and manageable volume. Thanks to the impressive effort performed by the authors and their experience in the field, it can satiate the interests of a great variety of readers. While expert researchers in the IR field and IR-related industrial companies can use it as a reference manual, it seems especially useful for students and non-expert readers willing to approach this research area. Like NIST, we would recommend this reading to anyone who may be interested in textual information retrieval."

Languages

Types

  • a 156
  • s 8
  • m 5
  • el 3
  • r 3
  • p 1
  • x 1
  • More… Less…