Search (453 results, page 1 of 23)

  • × theme_ss:"Retrievalstudien"
  1. Agata, T.: ¬A measure for evaluating search engines on the World Wide Web : retrieval test with ESL (Expected Search Length) (1997) 0.18
    0.18477678 = product of:
      0.29564285 = sum of:
        0.08285994 = weight(_text_:world in 3892) [ClassicSimilarity], result of:
          0.08285994 = score(doc=3892,freq=2.0), product of:
            0.16259687 = queryWeight, product of:
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.042302497 = queryNorm
            0.50960356 = fieldWeight in 3892, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.09375 = fieldNorm(doc=3892)
        0.11010534 = weight(_text_:wide in 3892) [ClassicSimilarity], result of:
          0.11010534 = score(doc=3892,freq=2.0), product of:
            0.18743214 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.042302497 = queryNorm
            0.5874411 = fieldWeight in 3892, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.09375 = fieldNorm(doc=3892)
        0.05973409 = weight(_text_:web in 3892) [ClassicSimilarity], result of:
          0.05973409 = score(doc=3892,freq=2.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.43268442 = fieldWeight in 3892, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.09375 = fieldNorm(doc=3892)
        0.01728395 = weight(_text_:information in 3892) [ClassicSimilarity], result of:
          0.01728395 = score(doc=3892,freq=2.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.23274569 = fieldWeight in 3892, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=3892)
        0.025659526 = product of:
          0.05131905 = sum of:
            0.05131905 = weight(_text_:retrieval in 3892) [ClassicSimilarity], result of:
              0.05131905 = score(doc=3892,freq=2.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.40105087 = fieldWeight in 3892, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3892)
          0.5 = coord(1/2)
      0.625 = coord(5/8)
    
    Source
    Library and information science. 1997, no.37, S.1-11
  2. Wu, C.-J.: Experiments on using the Dublin Core to reduce the retrieval error ratio (1998) 0.12
    0.119247794 = product of:
      0.19079646 = sum of:
        0.048334967 = weight(_text_:world in 5201) [ClassicSimilarity], result of:
          0.048334967 = score(doc=5201,freq=2.0), product of:
            0.16259687 = queryWeight, product of:
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.042302497 = queryNorm
            0.29726875 = fieldWeight in 5201, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5201)
        0.06422812 = weight(_text_:wide in 5201) [ClassicSimilarity], result of:
          0.06422812 = score(doc=5201,freq=2.0), product of:
            0.18743214 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.042302497 = queryNorm
            0.342674 = fieldWeight in 5201, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5201)
        0.03484489 = weight(_text_:web in 5201) [ClassicSimilarity], result of:
          0.03484489 = score(doc=5201,freq=2.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.25239927 = fieldWeight in 5201, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5201)
        0.017463064 = weight(_text_:information in 5201) [ClassicSimilarity], result of:
          0.017463064 = score(doc=5201,freq=6.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.23515764 = fieldWeight in 5201, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5201)
        0.025925435 = product of:
          0.05185087 = sum of:
            0.05185087 = weight(_text_:retrieval in 5201) [ClassicSimilarity], result of:
              0.05185087 = score(doc=5201,freq=6.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.40520695 = fieldWeight in 5201, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5201)
          0.5 = coord(1/2)
      0.625 = coord(5/8)
    
    Abstract
    In order to test the power of metadata on information retrieval, an experiment was designed and conducted on a group of 7 graduate students using the Dublin Core as the cataloguing metadata. Results show that, on average, the retrieval error rate is only 2.9 per cent for the MES system (http://140.136.85.194), which utilizes the Dublin Core to describe the documents on the World Wide Web, in contrast to 20.7 per cent for the 7 famous search engines including HOTBOT, GAIS, LYCOS, EXCITE, INFOSEEK, YAHOO, and OCTOPUS. The very low error rate indicates that the users can use the information of the Dublin Core to decide whether to retrieve the documents or not
    Source
    Journal of library and information science. 24(1998) no.1, S.50-64
  3. Khan, K.; Locatis, C.: Searching through cyberspace : the effects of link display and link density on information retrieval from hypertext on the World Wide Web (1998) 0.10
    0.09966378 = product of:
      0.15946205 = sum of:
        0.04142997 = weight(_text_:world in 446) [ClassicSimilarity], result of:
          0.04142997 = score(doc=446,freq=2.0), product of:
            0.16259687 = queryWeight, product of:
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.042302497 = queryNorm
            0.25480178 = fieldWeight in 446, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.046875 = fieldNorm(doc=446)
        0.05505267 = weight(_text_:wide in 446) [ClassicSimilarity], result of:
          0.05505267 = score(doc=446,freq=2.0), product of:
            0.18743214 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.042302497 = queryNorm
            0.29372054 = fieldWeight in 446, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046875 = fieldNorm(doc=446)
        0.029867046 = weight(_text_:web in 446) [ClassicSimilarity], result of:
          0.029867046 = score(doc=446,freq=2.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.21634221 = fieldWeight in 446, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=446)
        0.014968341 = weight(_text_:information in 446) [ClassicSimilarity], result of:
          0.014968341 = score(doc=446,freq=6.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.20156369 = fieldWeight in 446, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=446)
        0.018144025 = product of:
          0.03628805 = sum of:
            0.03628805 = weight(_text_:retrieval in 446) [ClassicSimilarity], result of:
              0.03628805 = score(doc=446,freq=4.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.2835858 = fieldWeight in 446, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.046875 = fieldNorm(doc=446)
          0.5 = coord(1/2)
      0.625 = coord(5/8)
    
    Abstract
    This study investigated information retrieval from hypertext on the WWW. Significant main and interaction effects were found for both link density (number of links per display) and display format (in paragraphs or lists) on search performance. Low link densities displayed in list format produced the best overall results, in terms of search accuracy, search time, number of links explored, and search task prioritization. Lower densities affected user ability to prioritize search tasks and produced more accurate searches, while list displays positively affected all aspects of searching except task prioritization. The performance of novices and experts, in terms of their previous experience browsing hypertext on the WWW, was compared. Experts performed better, mostly because of their superior task prioritization
    Source
    Journal of the American Society for Information Science. 49(1998) no.2, S.176-182
  4. Griesbaum, J.: Evaluierung hybrider Suchsysteme im WWW (2000) 0.09
    0.09238839 = product of:
      0.14782143 = sum of:
        0.04142997 = weight(_text_:world in 2482) [ClassicSimilarity], result of:
          0.04142997 = score(doc=2482,freq=2.0), product of:
            0.16259687 = queryWeight, product of:
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.042302497 = queryNorm
            0.25480178 = fieldWeight in 2482, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.046875 = fieldNorm(doc=2482)
        0.05505267 = weight(_text_:wide in 2482) [ClassicSimilarity], result of:
          0.05505267 = score(doc=2482,freq=2.0), product of:
            0.18743214 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.042302497 = queryNorm
            0.29372054 = fieldWeight in 2482, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046875 = fieldNorm(doc=2482)
        0.029867046 = weight(_text_:web in 2482) [ClassicSimilarity], result of:
          0.029867046 = score(doc=2482,freq=2.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.21634221 = fieldWeight in 2482, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=2482)
        0.008641975 = weight(_text_:information in 2482) [ClassicSimilarity], result of:
          0.008641975 = score(doc=2482,freq=2.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.116372846 = fieldWeight in 2482, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2482)
        0.012829763 = product of:
          0.025659526 = sum of:
            0.025659526 = weight(_text_:retrieval in 2482) [ClassicSimilarity], result of:
              0.025659526 = score(doc=2482,freq=2.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.20052543 = fieldWeight in 2482, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2482)
          0.5 = coord(1/2)
      0.625 = coord(5/8)
    
    Abstract
    Der Ausgangspunkt dieser Arbeit ist die Suchproblematik im World Wide Web. Suchmaschinen sind einerseits unverzichtbar für erfolgreiches Information Retrieval, andererseits wird ihnen eine mäßige Leistungsfähigkeit vorgeworfen. Das Thema dieser Arbeit ist die Untersuchung der Retrievaleffektivität deutschsprachiger Suchmaschinen. Es soll festgestellt werden, welche Retrievaleffektivität Nutzer derzeit erwarten können. Ein Ansatz, um die Retrievaleffektivität von Suchmaschinen zu erhöhen besteht darin, redaktionell von Menschen erstellte und automatisch generierte Suchergebnisse in einer Trefferliste zu vermengen. Ziel dieser Arbeit ist es, die Retrievaleffektivität solcher hybrider Systeme im Vergleich zu rein roboterbasierten Suchmaschinen zu evaluieren. Zunächst werden hierzu die grundlegenden Problembereiche bei der Evaluation von Retrievalsystemen analysiert. In Anlehnung an die von Tague-Sutcliff vorgeschlagene Methodik wird unter Beachtung der webspezifischen Besonderheiten eine mögliche Vorgehensweise erschlossen. Darauf aufbauend wird das konkrete Setting für die Durchführung der Evaluation erarbeitet und ein Retrievaleffektivitätstest bei den Suchmaschinen Lycos.de, AItaVista.de und QualiGo durchgeführt.
  5. Lazonder, A.W.; Biemans, H.J.A.; Wopereis, I.G.J.H.: Differences between novice and experienced users in searching information on the World Wide Web (2000) 0.09
    0.092217915 = product of:
      0.18443583 = sum of:
        0.04142997 = weight(_text_:world in 4598) [ClassicSimilarity], result of:
          0.04142997 = score(doc=4598,freq=2.0), product of:
            0.16259687 = queryWeight, product of:
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.042302497 = queryNorm
            0.25480178 = fieldWeight in 4598, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.046875 = fieldNorm(doc=4598)
        0.05505267 = weight(_text_:wide in 4598) [ClassicSimilarity], result of:
          0.05505267 = score(doc=4598,freq=2.0), product of:
            0.18743214 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.042302497 = queryNorm
            0.29372054 = fieldWeight in 4598, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046875 = fieldNorm(doc=4598)
        0.066784754 = weight(_text_:web in 4598) [ClassicSimilarity], result of:
          0.066784754 = score(doc=4598,freq=10.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.48375595 = fieldWeight in 4598, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4598)
        0.02116843 = weight(_text_:information in 4598) [ClassicSimilarity], result of:
          0.02116843 = score(doc=4598,freq=12.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.2850541 = fieldWeight in 4598, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=4598)
      0.5 = coord(4/8)
    
    Abstract
    Searching for information on the WWW basically comes down to locating an appropriate Web site and to retrieving relevant information from that site. This study examined the effect of a user's WWW experience on both phases of the search process. 35 students from 2 schools for Dutch pre-university education were observed while performing 3 search tasks. The results indicate that subjects with WWW-experience are more proficient in locating Web sites than are novice WWW-users. The observed differences were ascribed to the experts' superior skills in operating Web search engines. However, on tasks that required subjects to locate information on specific Web sites, the performance of experienced and novice users was equivalent - a result that is in line with hypertext research. Based on these findings, implications for training and supporting students in searching for information on the WWW are identified. Finally, the role of the subjects' level of domain expertise is discussed and directions for future research are proposed
    Source
    Journal of the American Society for Information Science. 51(2000) no.6, S.576-581
  6. Kantor, P.; Kim, M.H.; Ibraev, U.; Atasoy, K.: Estimating the number of relevant documents in enormous collections (1999) 0.08
    0.084259205 = product of:
      0.13481472 = sum of:
        0.034524977 = weight(_text_:world in 6690) [ClassicSimilarity], result of:
          0.034524977 = score(doc=6690,freq=2.0), product of:
            0.16259687 = queryWeight, product of:
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.042302497 = queryNorm
            0.21233483 = fieldWeight in 6690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6690)
        0.045877226 = weight(_text_:wide in 6690) [ClassicSimilarity], result of:
          0.045877226 = score(doc=6690,freq=2.0), product of:
            0.18743214 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.042302497 = queryNorm
            0.24476713 = fieldWeight in 6690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6690)
        0.024889207 = weight(_text_:web in 6690) [ClassicSimilarity], result of:
          0.024889207 = score(doc=6690,freq=2.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.18028519 = fieldWeight in 6690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6690)
        0.014403292 = weight(_text_:information in 6690) [ClassicSimilarity], result of:
          0.014403292 = score(doc=6690,freq=8.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.19395474 = fieldWeight in 6690, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6690)
        0.01512002 = product of:
          0.03024004 = sum of:
            0.03024004 = weight(_text_:retrieval in 6690) [ClassicSimilarity], result of:
              0.03024004 = score(doc=6690,freq=4.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.23632148 = fieldWeight in 6690, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=6690)
          0.5 = coord(1/2)
      0.625 = coord(5/8)
    
    Abstract
    In assessing information retrieval systems, it is important to know not only the precision of the retrieved set, but also to compare the number of retrieved relevant items to the total number of relevant items. For large collections, such as the TREC test collections, or the World Wide Web, it is not possible to enumerate the entire set of relevant documents. If the retrieved documents are evaluated, a variant of the statistical "capture-recapture" method can be used to estimate the total number of relevant documents, providing the several retrieval systems used are sufficiently independent. We show that the underlying signal detection model supporting such an analysis can be extended in two ways. First, assuming that there are two distinct performance characteristics (corresponding to the chance of retrieving a relevant, and retrieving a given non-relevant document), we show that if there are three or more independent systems available it is possible to estimate the number of relevant documents without actually having to decide whether each individual document is relevant. We report applications of this 3-system method to the TREC data, leading to the conclusion that the independence assumptions are not satisfied. We then extend the model to a multi-system, multi-problem model, and show that it is possible to include statistical dependencies of all orders in the model, and determine the number of relevant documents for each of the problems in the set. Application to the TREC setting will be presented
    Imprint
    Medford, NJ : Information Today
    Series
    Proceedings of the American Society for Information Science; vol.36
    Source
    Knowledge: creation, organization and use. Proceedings of the 62nd Annual Meeting of the American Society for Information Science, 31.10.-4.11.1999. Ed.: L. Woods
  7. Voorhees, E.M.; Harman, D.K.: ¬The Text REtrieval Conference (2005) 0.07
    0.07249272 = product of:
      0.11598835 = sum of:
        0.024167484 = weight(_text_:world in 5082) [ClassicSimilarity], result of:
          0.024167484 = score(doc=5082,freq=2.0), product of:
            0.16259687 = queryWeight, product of:
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.042302497 = queryNorm
            0.14863437 = fieldWeight in 5082, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5082)
        0.03211406 = weight(_text_:wide in 5082) [ClassicSimilarity], result of:
          0.03211406 = score(doc=5082,freq=2.0), product of:
            0.18743214 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.042302497 = queryNorm
            0.171337 = fieldWeight in 5082, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5082)
        0.017422445 = weight(_text_:web in 5082) [ClassicSimilarity], result of:
          0.017422445 = score(doc=5082,freq=2.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.12619963 = fieldWeight in 5082, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5082)
        0.0123482505 = weight(_text_:information in 5082) [ClassicSimilarity], result of:
          0.0123482505 = score(doc=5082,freq=12.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.16628155 = fieldWeight in 5082, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5082)
        0.029936114 = product of:
          0.05987223 = sum of:
            0.05987223 = weight(_text_:retrieval in 5082) [ClassicSimilarity], result of:
              0.05987223 = score(doc=5082,freq=32.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.46789268 = fieldWeight in 5082, product of:
                  5.656854 = tf(freq=32.0), with freq of:
                    32.0 = termFreq=32.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=5082)
          0.5 = coord(1/2)
      0.625 = coord(5/8)
    
    Abstract
    Text retrieval technology targets a problem that is all too familiar: finding relevant information in large stores of electronic documents. The problem is an old one, with the first research conference devoted to the subject held in 1958 [11]. Since then the problem has continued to grow as more information is created in electronic form and more people gain electronic access. The advent of the World Wide Web, where anyone can publish so everyone must search, is a graphic illustration of the need for effective retrieval technology. The Text REtrieval Conference (TREC) is a workshop series designed to build the infrastructure necessary for the large-scale evaluation of text retrieval technology, thereby accelerating its transfer into the commercial sector. The series is sponsored by the U.S. National Institute of Standards and Technology (NIST) and the U.S. Department of Defense. At the time of this writing, there have been twelve TREC workshops and preparations for the thirteenth workshop are under way. Participants in the workshops have been drawn from the academic, commercial, and government sectors, and have included representatives from more than twenty different countries. These collective efforts have accomplished a great deal: a variety of large test collections have been built for both traditional ad hoc retrieval and related tasks such as cross-language retrieval, speech retrieval, and question answering; retrieval effectiveness has approximately doubled; and many commercial retrieval systems now contain technology first developed in TREC.
    This book chronicles the evolution of retrieval systems over the course of TREC. To be sure, there has already been a wealth of information written about TREC. Each conference has produced a proceedings containing general overviews of the various tasks, papers written by the individual participants, and evaluation results.1 Reports on expanded versions of TREC experiments frequently appear in the wider information retrieval literature. There also have been special issues of journals devoted to particular TRECs [3; 13] and particular TREC tasks [6; 4]. No single volume could hope to be a comprehensive record of all TREC-related research. Instead, this book looks to distill the overabundance of detail into a manageable whole that summarizes the main lessons learned from TREC. The book consists of three main parts. The first part contains introductory and descriptive chapters on TREC's history, the major products of TREC (the test collections), and the retrieval evaluation methodology. Part II includes chapters describing the major TREC ''tracks,'' evaluations of special subtopics such as cross-language retrieval and question answering. Part III contains contributions from research groups that have participated in TREC. The epilogue to the book is written by Karen Sparck Jones, who reflects on the impact TREC has had on the information retrieval field. The structure of this introductory chapter is similar to that of the book as a whole. The chapter begins with a short history of TREC; expanded descriptions of specific aspects of the history are included in subsequent chapters to make those chapters self-contained. Section 1.2 describes TREC's track structure, which has been responsible for the growth of TREC and allows TREC to adapt to changing needs. The final section lists both the major accomplishments of TREC and some remaining challenges.
    Source
    TREC: experiment and evaluation in information retrieval. Ed.: E.M. Voorhees, u. D.K. Harman
  8. Crestani, F.; Rijsbergen, C.J. van: Information retrieval by imaging (1996) 0.07
    0.06757201 = product of:
      0.18019202 = sum of:
        0.05859083 = weight(_text_:world in 6967) [ClassicSimilarity], result of:
          0.05859083 = score(doc=6967,freq=4.0), product of:
            0.16259687 = queryWeight, product of:
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.042302497 = queryNorm
            0.36034414 = fieldWeight in 6967, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.046875 = fieldNorm(doc=6967)
        0.019324046 = weight(_text_:information in 6967) [ClassicSimilarity], result of:
          0.019324046 = score(doc=6967,freq=10.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.2602176 = fieldWeight in 6967, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=6967)
        0.102277145 = sum of:
          0.06788872 = weight(_text_:retrieval in 6967) [ClassicSimilarity], result of:
            0.06788872 = score(doc=6967,freq=14.0), product of:
              0.12796146 = queryWeight, product of:
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.042302497 = queryNorm
              0.5305404 = fieldWeight in 6967, product of:
                3.7416575 = tf(freq=14.0), with freq of:
                  14.0 = termFreq=14.0
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.046875 = fieldNorm(doc=6967)
          0.034388427 = weight(_text_:22 in 6967) [ClassicSimilarity], result of:
            0.034388427 = score(doc=6967,freq=2.0), product of:
              0.14813614 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.042302497 = queryNorm
              0.23214069 = fieldWeight in 6967, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=6967)
      0.375 = coord(3/8)
    
    Abstract
    Explains briefly what constitutes the imaging process and explains how imaging can be used in information retrieval. Proposes an approach based on the concept of: 'a term is a possible world'; which enables the exploitation of term to term relationships which are estimated using an information theoretic measure. Reports results of an evaluation exercise to compare the performance of imaging retrieval, using possible world semantics, with a benchmark and using the Cranfield 2 document collection to measure precision and recall. Initially, the performance imaging retrieval was seen to be better but statistical analysis proved that the difference was not significant. The problem with imaging retrieval lies in the amount of computations needed to be performed at run time and a later experiement investigated the possibility of reducing this amount. Notes lines of further investigation
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
  9. ¬The Eleventh Text Retrieval Conference, TREC 2002 (2003) 0.06
    0.061833713 = product of:
      0.1648899 = sum of:
        0.039822727 = weight(_text_:web in 4049) [ClassicSimilarity], result of:
          0.039822727 = score(doc=4049,freq=2.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.2884563 = fieldWeight in 4049, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0625 = fieldNorm(doc=4049)
        0.019957788 = weight(_text_:information in 4049) [ClassicSimilarity], result of:
          0.019957788 = score(doc=4049,freq=6.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.2687516 = fieldWeight in 4049, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=4049)
        0.10510938 = sum of:
          0.05925814 = weight(_text_:retrieval in 4049) [ClassicSimilarity], result of:
            0.05925814 = score(doc=4049,freq=6.0), product of:
              0.12796146 = queryWeight, product of:
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.042302497 = queryNorm
              0.46309367 = fieldWeight in 4049, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.0625 = fieldNorm(doc=4049)
          0.045851234 = weight(_text_:22 in 4049) [ClassicSimilarity], result of:
            0.045851234 = score(doc=4049,freq=2.0), product of:
              0.14813614 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.042302497 = queryNorm
              0.30952093 = fieldWeight in 4049, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=4049)
      0.375 = coord(3/8)
    
    Abstract
    Proceedings of the llth TREC-conference held in Gaithersburg, Maryland (USA), November 19-22, 2002. Aim of the conference was discussion an retrieval and related information-seeking tasks for large test collection. 93 research groups used different techniques, for information retrieval from the same large database. This procedure makes it possible to compare the results. The tasks are: Cross-language searching, filtering, interactive searching, searching for novelty, question answering, searching for video shots, and Web searching.
    Imprint
    Gaithersburg, MD : National Institute of Standards / Information Technology Laboratory
  10. Harter, S.P.: Variations in relevance assessments and the measurement of retrieval effectiveness (1996) 0.06
    0.059532233 = product of:
      0.119064465 = sum of:
        0.034524977 = weight(_text_:world in 3004) [ClassicSimilarity], result of:
          0.034524977 = score(doc=3004,freq=2.0), product of:
            0.16259687 = queryWeight, product of:
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.042302497 = queryNorm
            0.21233483 = fieldWeight in 3004, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3004)
        0.045877226 = weight(_text_:wide in 3004) [ClassicSimilarity], result of:
          0.045877226 = score(doc=3004,freq=2.0), product of:
            0.18743214 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.042302497 = queryNorm
            0.24476713 = fieldWeight in 3004, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3004)
        0.012473618 = weight(_text_:information in 3004) [ClassicSimilarity], result of:
          0.012473618 = score(doc=3004,freq=6.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.16796975 = fieldWeight in 3004, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3004)
        0.026188646 = product of:
          0.05237729 = sum of:
            0.05237729 = weight(_text_:retrieval in 3004) [ClassicSimilarity], result of:
              0.05237729 = score(doc=3004,freq=12.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.40932083 = fieldWeight in 3004, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3004)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    The purpose of this article is to bring attention to the problem of variations in relevance assessments and the effects that these may have on measures of retrieval effectiveness. Through an analytical review of the literature, I show that despite known wide variations in relevance assessments in experimental test collections, their effects on the measurement of retrieval performance are almost completely unstudied. I will further argue that what we know about tha many variables that have been found to affect relevance assessments under experimental conditions, as well as our new understanding of psychological, situational, user-based relevance, point to a single conclusion. We can no longer rest the evaluation of information retrieval systems on the assumption that such variations do not significantly affect the measurement of information retrieval performance. A series of thourough, rigorous, and extensive tests is needed, of precisely how, and under what conditions, variations in relevance assessments do, and do not, affect measures of retrieval performance. We need to develop approaches to evaluation that are sensitive to these variations and to human factors and individual differences more generally. Our approaches to evaluation must reflect the real world of real users
    Source
    Journal of the American Society for Information Science. 47(1996) no.1, S.37-49
  11. Breuer, T.; Tavakolpoursaleh, N.; Schaer, P.; Hienert, D.; Schaible, J.; Castro, L.J.: Online Information Retrieval Evaluation using the STELLA Framework (2022) 0.05
    0.053362496 = product of:
      0.10672499 = sum of:
        0.04142997 = weight(_text_:world in 640) [ClassicSimilarity], result of:
          0.04142997 = score(doc=640,freq=2.0), product of:
            0.16259687 = queryWeight, product of:
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.042302497 = queryNorm
            0.25480178 = fieldWeight in 640, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.8436708 = idf(docFreq=2573, maxDocs=44218)
              0.046875 = fieldNorm(doc=640)
        0.029867046 = weight(_text_:web in 640) [ClassicSimilarity], result of:
          0.029867046 = score(doc=640,freq=2.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.21634221 = fieldWeight in 640, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=640)
        0.01728395 = weight(_text_:information in 640) [ClassicSimilarity], result of:
          0.01728395 = score(doc=640,freq=8.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.23274569 = fieldWeight in 640, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=640)
        0.018144025 = product of:
          0.03628805 = sum of:
            0.03628805 = weight(_text_:retrieval in 640) [ClassicSimilarity], result of:
              0.03628805 = score(doc=640,freq=4.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.2835858 = fieldWeight in 640, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.046875 = fieldNorm(doc=640)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    Involving users in early phases of software development has become a common strategy as it enables developers to consider user needs from the beginning. Once a system is in production, new opportunities to observe, evaluate and learn from users emerge as more information becomes available. Gathering information from users to continuously evaluate their behavior is a common practice for commercial software, while the Cranfield paradigm remains the preferred option for Information Retrieval (IR) and recommendation systems in the academic world. Here we introduce the Infrastructures for Living Labs STELLA project which aims to create an evaluation infrastructure allowing experimental systems to run along production web-based academic search systems with real users. STELLA combines user interactions and log files analyses to enable large-scale A/B experiments for academic search.
  12. Ellis, D.: Progress and problems in information retrieval (1996) 0.04
    0.044296816 = product of:
      0.17718726 = sum of:
        0.0345679 = weight(_text_:information in 789) [ClassicSimilarity], result of:
          0.0345679 = score(doc=789,freq=18.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.46549135 = fieldWeight in 789, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=789)
        0.14261937 = sum of:
          0.09676813 = weight(_text_:retrieval in 789) [ClassicSimilarity], result of:
            0.09676813 = score(doc=789,freq=16.0), product of:
              0.12796146 = queryWeight, product of:
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.042302497 = queryNorm
              0.75622874 = fieldWeight in 789, product of:
                4.0 = tf(freq=16.0), with freq of:
                  16.0 = termFreq=16.0
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.0625 = fieldNorm(doc=789)
          0.045851234 = weight(_text_:22 in 789) [ClassicSimilarity], result of:
            0.045851234 = score(doc=789,freq=2.0), product of:
              0.14813614 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.042302497 = queryNorm
              0.30952093 = fieldWeight in 789, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=789)
      0.25 = coord(2/8)
    
    Abstract
    An introduction to the principal generic approaches to information retrieval research with their associated concepts, models and systems, this text is designed to keep the information professional up to date with the major themes and developments that have preoccupied researchers in recent month in relation to textual and documentary retrieval systems.
    COMPASS
    Information retrieval
    Content
    First published 1991 as New horizons in information retrieval
    Date
    26. 7.2002 20:22:46
    Footnote
    Rez. in: Managing information 3(1996) no.10, S.49 (D. Bawden); Program 32(1998) no.2, S.190-192 (C. Revie)
    LCSH
    Information retrieval
    Subject
    Information retrieval
    Information retrieval
  13. Harman, D.: ¬The Text REtrieval Conferences (TRECs) : providing a test-bed for information retrieval systems (1998) 0.04
    0.04419837 = product of:
      0.11786232 = sum of:
        0.06422812 = weight(_text_:wide in 1314) [ClassicSimilarity], result of:
          0.06422812 = score(doc=1314,freq=2.0), product of:
            0.18743214 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.042302497 = queryNorm
            0.342674 = fieldWeight in 1314, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1314)
        0.02016461 = weight(_text_:information in 1314) [ClassicSimilarity], result of:
          0.02016461 = score(doc=1314,freq=8.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.27153665 = fieldWeight in 1314, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1314)
        0.033469595 = product of:
          0.06693919 = sum of:
            0.06693919 = weight(_text_:retrieval in 1314) [ClassicSimilarity], result of:
              0.06693919 = score(doc=1314,freq=10.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.5231199 = fieldWeight in 1314, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1314)
          0.5 = coord(1/2)
      0.375 = coord(3/8)
    
    Abstract
    The Text REtrieval Conference (TREC) workshop series encourages research in information retrieval from large text applications by providing a large test collection, uniform scoring procedures and a forum for organizations interested in comparing their results. Now in its seventh year, the conference has become the major experimental effort in the field. Participants in the TREC conferences have examined a wide variety of retrieval techniques, including methods using automatic thesauri, sophisticated term weighting, natural language techniques, relevance feedback and advanced pattern matching. The TREC conference series is co-sponsored by the National Institute of Standards and Technology (NIST) and the Information Technology Office of the Defense Advanced Research Projects Agency (DARPA)
    Source
    Bulletin of the American Society for Information Science. 24(1998), April/May, S.11-13
  14. Frei, H.P.; Meienberg, S.; Schäuble, P.: ¬The perils of interpreting recall and precision values (1991) 0.04
    0.04408252 = product of:
      0.11755338 = sum of:
        0.07340356 = weight(_text_:wide in 786) [ClassicSimilarity], result of:
          0.07340356 = score(doc=786,freq=2.0), product of:
            0.18743214 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.042302497 = queryNorm
            0.3916274 = fieldWeight in 786, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0625 = fieldNorm(doc=786)
        0.019957788 = weight(_text_:information in 786) [ClassicSimilarity], result of:
          0.019957788 = score(doc=786,freq=6.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.2687516 = fieldWeight in 786, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=786)
        0.024192033 = product of:
          0.048384067 = sum of:
            0.048384067 = weight(_text_:retrieval in 786) [ClassicSimilarity], result of:
              0.048384067 = score(doc=786,freq=4.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.37811437 = fieldWeight in 786, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0625 = fieldNorm(doc=786)
          0.5 = coord(1/2)
      0.375 = coord(3/8)
    
    Abstract
    The traditional recall and precision measure is inappropriate when retrieval algorithms that retrieve information from Wide Area Networks are evaluated. The principle reason is that information available in WANs is dynamic and its size os orders of magnitude greater than the size of the usual test collections. To overcome these problems, a new efffectiveness measure has been developed, which we call the 'usefulness measure'
    Source
    Information retrieval: GI/GMD-Workshop, Darmstadt, 23.-24.6.1991: Proceedings. Ed.: N. Fuhr
  15. Voorhees, E.M.; Harman, D.: Overview of the Sixth Text REtrieval Conference (TREC-6) (2000) 0.04
    0.040069126 = product of:
      0.1602765 = sum of:
        0.02016461 = weight(_text_:information in 6438) [ClassicSimilarity], result of:
          0.02016461 = score(doc=6438,freq=2.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.27153665 = fieldWeight in 6438, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=6438)
        0.1401119 = sum of:
          0.05987223 = weight(_text_:retrieval in 6438) [ClassicSimilarity], result of:
            0.05987223 = score(doc=6438,freq=2.0), product of:
              0.12796146 = queryWeight, product of:
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.042302497 = queryNorm
              0.46789268 = fieldWeight in 6438, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.109375 = fieldNorm(doc=6438)
          0.08023966 = weight(_text_:22 in 6438) [ClassicSimilarity], result of:
            0.08023966 = score(doc=6438,freq=2.0), product of:
              0.14813614 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.042302497 = queryNorm
              0.5416616 = fieldWeight in 6438, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.109375 = fieldNorm(doc=6438)
      0.25 = coord(2/8)
    
    Date
    11. 8.2001 16:22:19
    Source
    Information processing and management. 36(2000) no.1, S.3-36
  16. Dalrymple, P.W.: Retrieval by reformulation in two library catalogs : toward a cognitive model of searching behavior (1990) 0.04
    0.040069126 = product of:
      0.1602765 = sum of:
        0.02016461 = weight(_text_:information in 5089) [ClassicSimilarity], result of:
          0.02016461 = score(doc=5089,freq=2.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.27153665 = fieldWeight in 5089, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=5089)
        0.1401119 = sum of:
          0.05987223 = weight(_text_:retrieval in 5089) [ClassicSimilarity], result of:
            0.05987223 = score(doc=5089,freq=2.0), product of:
              0.12796146 = queryWeight, product of:
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.042302497 = queryNorm
              0.46789268 = fieldWeight in 5089, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.109375 = fieldNorm(doc=5089)
          0.08023966 = weight(_text_:22 in 5089) [ClassicSimilarity], result of:
            0.08023966 = score(doc=5089,freq=2.0), product of:
              0.14813614 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.042302497 = queryNorm
              0.5416616 = fieldWeight in 5089, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.109375 = fieldNorm(doc=5089)
      0.25 = coord(2/8)
    
    Date
    22. 7.2006 18:43:54
    Source
    Journal of the American Society for Information Science. 41(1990) no.4, S.272-281
  17. Mandl, T.: Web- und Multimedia-Dokumente : Neuere Entwicklungen bei der Evaluierung von Information Retrieval Systemen (2003) 0.04
    0.038939655 = product of:
      0.103839085 = sum of:
        0.039822727 = weight(_text_:web in 1734) [ClassicSimilarity], result of:
          0.039822727 = score(doc=1734,freq=2.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.2884563 = fieldWeight in 1734, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0625 = fieldNorm(doc=1734)
        0.025765393 = weight(_text_:information in 1734) [ClassicSimilarity], result of:
          0.025765393 = score(doc=1734,freq=10.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.3469568 = fieldWeight in 1734, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=1734)
        0.038250964 = product of:
          0.07650193 = sum of:
            0.07650193 = weight(_text_:retrieval in 1734) [ClassicSimilarity], result of:
              0.07650193 = score(doc=1734,freq=10.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.59785134 = fieldWeight in 1734, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1734)
          0.5 = coord(1/2)
      0.375 = coord(3/8)
    
    Abstract
    Die Menge an Daten im Internet steigt weiter rapide an. Damit wächst auch der Bedarf an qualitativ hochwertigen Information Retrieval Diensten zur Orientierung und problemorientierten Suche. Die Entscheidung für die Benutzung oder Beschaffung von Information Retrieval Software erfordert aussagekräftige Evaluierungsergebnisse. Dieser Beitrag stellt neuere Entwicklungen bei der Evaluierung von Information Retrieval Systemen vor und zeigt den Trend zu Spezialisierung und Diversifizierung von Evaluierungsstudien, die den Realitätsgrad derErgebnisse erhöhen. DerSchwerpunkt liegt auf dem Retrieval von Fachtexten, Internet-Seiten und Multimedia-Objekten.
    Source
    Information - Wissenschaft und Praxis. 54(2003) H.4, S.203-210
  18. Ravana, S.D.; Taheri, M.S.; Rajagopal, P.: Document-based approach to improve the accuracy of pairwise comparison in evaluating information retrieval systems (2015) 0.04
    0.038646072 = product of:
      0.10305619 = sum of:
        0.024889207 = weight(_text_:web in 2587) [ClassicSimilarity], result of:
          0.024889207 = score(doc=2587,freq=2.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.18028519 = fieldWeight in 2587, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2587)
        0.012473618 = weight(_text_:information in 2587) [ClassicSimilarity], result of:
          0.012473618 = score(doc=2587,freq=6.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.16796975 = fieldWeight in 2587, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2587)
        0.06569336 = sum of:
          0.037036337 = weight(_text_:retrieval in 2587) [ClassicSimilarity], result of:
            0.037036337 = score(doc=2587,freq=6.0), product of:
              0.12796146 = queryWeight, product of:
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.042302497 = queryNorm
              0.28943354 = fieldWeight in 2587, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2587)
          0.028657023 = weight(_text_:22 in 2587) [ClassicSimilarity], result of:
            0.028657023 = score(doc=2587,freq=2.0), product of:
              0.14813614 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.042302497 = queryNorm
              0.19345059 = fieldWeight in 2587, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2587)
      0.375 = coord(3/8)
    
    Abstract
    Purpose The purpose of this paper is to propose a method to have more accurate results in comparing performance of the paired information retrieval (IR) systems with reference to the current method, which is based on the mean effectiveness scores of the systems across a set of identified topics/queries. Design/methodology/approach Based on the proposed approach, instead of the classic method of using a set of topic scores, the documents level scores are considered as the evaluation unit. These document scores are the defined document's weight, which play the role of the mean average precision (MAP) score of the systems as a significance test's statics. The experiments were conducted using the TREC 9 Web track collection. Findings The p-values generated through the two types of significance tests, namely the Student's t-test and Mann-Whitney show that by using the document level scores as an evaluation unit, the difference between IR systems is more significant compared with utilizing topic scores. Originality/value Utilizing a suitable test collection is a primary prerequisite for IR systems comparative evaluation. However, in addition to reusable test collections, having an accurate statistical testing is a necessity for these evaluations. The findings of this study will assist IR researchers to evaluate their retrieval systems and algorithms more accurately.
    Date
    20. 1.2015 18:30:22
    Source
    Aslib journal of information management. 67(2015) no.4, S.408-421
  19. Hawking, D.; Craswell, N.: ¬The very large collection and Web tracks (2005) 0.04
    0.038504086 = product of:
      0.10267756 = sum of:
        0.05973409 = weight(_text_:web in 5085) [ClassicSimilarity], result of:
          0.05973409 = score(doc=5085,freq=2.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.43268442 = fieldWeight in 5085, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.09375 = fieldNorm(doc=5085)
        0.01728395 = weight(_text_:information in 5085) [ClassicSimilarity], result of:
          0.01728395 = score(doc=5085,freq=2.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.23274569 = fieldWeight in 5085, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=5085)
        0.025659526 = product of:
          0.05131905 = sum of:
            0.05131905 = weight(_text_:retrieval in 5085) [ClassicSimilarity], result of:
              0.05131905 = score(doc=5085,freq=2.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.40105087 = fieldWeight in 5085, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.09375 = fieldNorm(doc=5085)
          0.5 = coord(1/2)
      0.375 = coord(3/8)
    
    Source
    TREC: experiment and evaluation in information retrieval. Ed.: E.M. Voorhees, u. D.K. Harman
  20. MacCall, S.L.; Cleveland, A.D.: ¬A relevance-based quantitative measure for Internet information retrieval evaluation (1999) 0.04
    0.035171673 = product of:
      0.09379113 = sum of:
        0.042238384 = weight(_text_:web in 6689) [ClassicSimilarity], result of:
          0.042238384 = score(doc=6689,freq=4.0), product of:
            0.13805464 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.042302497 = queryNorm
            0.3059541 = fieldWeight in 6689, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=6689)
        0.022864517 = weight(_text_:information in 6689) [ClassicSimilarity], result of:
          0.022864517 = score(doc=6689,freq=14.0), product of:
            0.0742611 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.042302497 = queryNorm
            0.3078936 = fieldWeight in 6689, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=6689)
        0.028688224 = product of:
          0.057376448 = sum of:
            0.057376448 = weight(_text_:retrieval in 6689) [ClassicSimilarity], result of:
              0.057376448 = score(doc=6689,freq=10.0), product of:
                0.12796146 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.042302497 = queryNorm
                0.44838852 = fieldWeight in 6689, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.046875 = fieldNorm(doc=6689)
          0.5 = coord(1/2)
      0.375 = coord(3/8)
    
    Abstract
    An important indicator of a maturating Internet is the development of metrics for its evaluation as a practical tool for enduser information retrieval. However, the Internet presents specific problems for traditional IR measures, such as the need to deal with the variety of classes of retrieval tools. This paper presents a metric for comparing the performance of common classes of Internet information retrieval tool, including human indexed catalogs of web resources and automatically indexed databases of web pages. The metric uses a relevance-based quantitative measure to compare the performance of endusers using these Internet information retrieval tools. The benefit of the proposed metric is that it is relevance-based (using enduser relevance judgments), and it facilitates the comparison of the performance of different classes of IIR tools
    Imprint
    Medford, NJ : Information Today
    Series
    Proceedings of the American Society for Information Science; vol36
    Source
    Knowledge: creation, organization and use. Proceedings of the 62nd Annual Meeting of the American Society for Information Science, 31.10.-4.11.1999. Ed.: L. Woods

Languages

Types

  • a 421
  • s 15
  • el 10
  • m 9
  • r 5
  • x 3
  • d 1
  • p 1
  • More… Less…