Search (90 results, page 1 of 5)

  • × language_ss:"e"
  • × theme_ss:"Retrievalstudien"
  • × year_i:[2000 TO 2010}
  1. Leininger, K.: Interindexer consistency in PsychINFO (2000) 0.06
    0.056897886 = product of:
      0.085346825 = sum of:
        0.01891706 = weight(_text_:of in 2552) [ClassicSimilarity], result of:
          0.01891706 = score(doc=2552,freq=10.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.23179851 = fieldWeight in 2552, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2552)
        0.066429764 = sum of:
          0.02400495 = weight(_text_:science in 2552) [ClassicSimilarity], result of:
            0.02400495 = score(doc=2552,freq=2.0), product of:
              0.13747036 = queryWeight, product of:
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.05218836 = queryNorm
              0.17461908 = fieldWeight in 2552, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.046875 = fieldNorm(doc=2552)
          0.042424813 = weight(_text_:22 in 2552) [ClassicSimilarity], result of:
            0.042424813 = score(doc=2552,freq=2.0), product of:
              0.18275474 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05218836 = queryNorm
              0.23214069 = fieldWeight in 2552, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2552)
      0.6666667 = coord(2/3)
    
    Abstract
    Reports results of a study to examine interindexer consistency (the degree to which indexers, when assigning terms to a chosen record, will choose the same terms to reflect that record) in the PsycINFO database using 60 records that were inadvertently processed twice between 1996 and 1998. Five aspects of interindexer consistency were analysed. Two methods were used to calculate interindexer consistency: one posited by Hooper (1965) and the other by Rollin (1981). Aspects analysed were: checktag consistency (66.24% using Hooper's calculation and 77.17% using Rollin's); major-to-all term consistency (49.31% and 62.59% respectively); overall indexing consistency (49.02% and 63.32%); classification code consistency (44.17% and 45.00%); and major-to-major term consistency (43.24% and 56.09%). The average consistency across all categories was 50.4% using Hooper's method and 60.83% using Rollin's. Although comparison with previous studies is difficult due to methodological variations in the overall study of indexing consistency and the specific characteristics of the database, results generally support previous findings when trends and similar studies are analysed.
    Date
    9. 2.1997 18:44:22
    Source
    Journal of librarianship and information science. 32(2000) no.1, S.4-8
  2. King, D.W.: Blazing new trails : in celebration of an audacious career (2000) 0.05
    0.054491144 = product of:
      0.08173671 = sum of:
        0.02637858 = weight(_text_:of in 1184) [ClassicSimilarity], result of:
          0.02637858 = score(doc=1184,freq=28.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.32322758 = fieldWeight in 1184, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1184)
        0.055358134 = sum of:
          0.020004123 = weight(_text_:science in 1184) [ClassicSimilarity], result of:
            0.020004123 = score(doc=1184,freq=2.0), product of:
              0.13747036 = queryWeight, product of:
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.05218836 = queryNorm
              0.1455159 = fieldWeight in 1184, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1184)
          0.03535401 = weight(_text_:22 in 1184) [ClassicSimilarity], result of:
            0.03535401 = score(doc=1184,freq=2.0), product of:
              0.18275474 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05218836 = queryNorm
              0.19345059 = fieldWeight in 1184, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1184)
      0.6666667 = coord(2/3)
    
    Abstract
    I had the distinct pleasure of working with Pauline Atherton (Cochrane) during the 1960s, a period that can be considered the heyday of automated information system design and evaluation in the United States. I first met Pauline at the 1962 American Documentation Institute annual meeting in North Hollywood, Florida. My company, Westat Research Analysts, had recently been awarded a contract by the U.S. Patent Office to provide statistical support for the design of experiments with automated information retrieval systems. I was asked to attend the meeting to learn more about information retrieval systems and to begin informing others of U.S. Patent Office activities in this area. At one session, Pauline and I questioned a speaker about the research that he presented. Pauline's questions concerned the logic of their approach and mine, the statistical aspects. After the session, she came over to talk to me and we began a professional and personal friendship that continues to this day. During the 1960s, Pauline was involved in several important information-retrieval projects including a series of studies for the American Institute of Physics, a dissertation examining the relevance of retrieved documents, and development and evaluation of an online information-retrieval system. I had the opportunity to work with Pauline and her colleagues an four of those projects and will briefly describe her work in the 1960s.
    Date
    22. 9.1997 19:16:05
    Imprint
    Urbana-Champaign, IL : Illinois University at Urbana-Champaign, Graduate School of Library and Information Science
    Source
    Saving the time of the library user through subject access innovation: Papers in honor of Pauline Atherton Cochrane. Ed.: W.J. Wheeler
  3. Voorhees, E.M.; Harman, D.: Overview of the Sixth Text REtrieval Conference (TREC-6) (2000) 0.05
    0.046157025 = product of:
      0.06923553 = sum of:
        0.01973992 = weight(_text_:of in 6438) [ClassicSimilarity], result of:
          0.01973992 = score(doc=6438,freq=2.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.24188137 = fieldWeight in 6438, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.109375 = fieldNorm(doc=6438)
        0.049495615 = product of:
          0.09899123 = sum of:
            0.09899123 = weight(_text_:22 in 6438) [ClassicSimilarity], result of:
              0.09899123 = score(doc=6438,freq=2.0), product of:
                0.18275474 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05218836 = queryNorm
                0.5416616 = fieldWeight in 6438, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6438)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Date
    11. 8.2001 16:22:19
  4. Larsen, B.; Ingwersen, P.; Lund, B.: Data fusion according to the principle of polyrepresentation (2009) 0.04
    0.044564277 = product of:
      0.066846415 = sum of:
        0.02255991 = weight(_text_:of in 2752) [ClassicSimilarity], result of:
          0.02255991 = score(doc=2752,freq=32.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.27643585 = fieldWeight in 2752, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=2752)
        0.044286508 = sum of:
          0.0160033 = weight(_text_:science in 2752) [ClassicSimilarity], result of:
            0.0160033 = score(doc=2752,freq=2.0), product of:
              0.13747036 = queryWeight, product of:
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.05218836 = queryNorm
              0.11641272 = fieldWeight in 2752, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.03125 = fieldNorm(doc=2752)
          0.028283209 = weight(_text_:22 in 2752) [ClassicSimilarity], result of:
            0.028283209 = score(doc=2752,freq=2.0), product of:
              0.18275474 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05218836 = queryNorm
              0.15476047 = fieldWeight in 2752, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=2752)
      0.6666667 = coord(2/3)
    
    Abstract
    We report data fusion experiments carried out on the four best-performing retrieval models from TREC 5. Three were conceptually/algorithmically very different from one another; one was algorithmically similar to one of the former. The objective of the test was to observe the performance of the 11 logical data fusion combinations compared to the performance of the four individual models and their intermediate fusions when following the principle of polyrepresentation. This principle is based on cognitive IR perspective (Ingwersen & Järvelin, 2005) and implies that each retrieval model is regarded as a representation of a unique interpretation of information retrieval (IR). It predicts that only fusions of very different, but equally good, IR models may outperform each constituent as well as their intermediate fusions. Two kinds of experiments were carried out. One tested restricted fusions, which entails that only the inner disjoint overlap documents between fused models are ranked. The second set of experiments was based on traditional data fusion methods. The experiments involved the 30 TREC 5 topics that contain more than 44 relevant documents. In all tests, the Borda and CombSUM scoring methods were used. Performance was measured by precision and recall, with document cutoff values (DCVs) at 100 and 15 documents, respectively. Results show that restricted fusions made of two, three, or four cognitively/algorithmically very different retrieval models perform significantly better than do the individual models at DCV100. At DCV15, however, the results of polyrepresentative fusion were less predictable. The traditional fusion method based on polyrepresentation principles demonstrates a clear picture of performance at both DCV levels and verifies the polyrepresentation predictions for data fusion in IR. Data fusion improves retrieval performance over their constituent IR models only if the models all are quite conceptually/algorithmically dissimilar and equally and well performing, in that order of importance.
    Date
    22. 3.2009 18:48:28
    Source
    Journal of the American Society for Information Science and Technology. 60(2009) no.4, S.646-654
  5. Voiskunskii, V.G.: Evaluation of search results (2000) 0.04
    0.037281495 = product of:
      0.05592224 = sum of:
        0.027916465 = weight(_text_:of in 4670) [ClassicSimilarity], result of:
          0.027916465 = score(doc=4670,freq=4.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.34207192 = fieldWeight in 4670, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.109375 = fieldNorm(doc=4670)
        0.028005775 = product of:
          0.05601155 = sum of:
            0.05601155 = weight(_text_:science in 4670) [ClassicSimilarity], result of:
              0.05601155 = score(doc=4670,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.40744454 = fieldWeight in 4670, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4670)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Encyclopedia of library and information science. Vol.66, [=Suppl.29]
  6. Tombros, T.; Crestani, F.: Users' perception of relevance of spoken documents (2000) 0.03
    0.033059694 = product of:
      0.049589537 = sum of:
        0.035586648 = weight(_text_:of in 4996) [ClassicSimilarity], result of:
          0.035586648 = score(doc=4996,freq=26.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.43605784 = fieldWeight in 4996, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4996)
        0.0140028875 = product of:
          0.028005775 = sum of:
            0.028005775 = weight(_text_:science in 4996) [ClassicSimilarity], result of:
              0.028005775 = score(doc=4996,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.20372227 = fieldWeight in 4996, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4996)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    We present the results of a study of user's perception of relevance of documents. The aim is to study experimentally how users' perception varies depending on the form that retrieved documents are presented. Documents retrieved in response to a query are presented to users in a variety of ways, from full text to a machine spoken query-biased automatically-generated summary, and the difference in users' perception of relevance is studied. The experimental results suggest that the effectiveness of advanced multimedia Information Retrieval applications may be affected by the low level of users' perception of relevance of retrieved documents
    Source
    Journal of the American Society for Information Science. 51(2000) no.10, S.929-939
  7. Kilgour, F.: ¬An experiment using coordinate title word searches (2004) 0.03
    0.032128956 = product of:
      0.048193432 = sum of:
        0.034190547 = weight(_text_:of in 2065) [ClassicSimilarity], result of:
          0.034190547 = score(doc=2065,freq=24.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.41895083 = fieldWeight in 2065, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2065)
        0.0140028875 = product of:
          0.028005775 = sum of:
            0.028005775 = weight(_text_:science in 2065) [ClassicSimilarity], result of:
              0.028005775 = score(doc=2065,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.20372227 = fieldWeight in 2065, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2065)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This study, the fourth and last of a series designed to produce new information to improve retrievability of books in libraries, explores the effectiveness of retrieving a known-item book using words from titles only. From daily printouts of circulation records at the Walter Royal Davis Library of the University of North Carolina at Chapel Hill, 749 titles were taken and then searched an the 4-million entry catalog at the library of the University of Michigan. The principal finding was that searches produced titles having personal authors 81.4% of the time and anonymous titles 91.5% of the time; these figures are 15 and 5%, respectively, lower than the lowest findings presented in the previous three articles of this series (Kilgour, 1995; 1997; 2001).
    Source
    Journal of the American Society for Information Science and technology. 55(2004) no.1, S.74-80
  8. Wilbur, W.J.: Global term weights for document retrieval learned from TREC data (2001) 0.03
    0.031830467 = product of:
      0.047745697 = sum of:
        0.01973992 = weight(_text_:of in 2647) [ClassicSimilarity], result of:
          0.01973992 = score(doc=2647,freq=2.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.24188137 = fieldWeight in 2647, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.109375 = fieldNorm(doc=2647)
        0.028005775 = product of:
          0.05601155 = sum of:
            0.05601155 = weight(_text_:science in 2647) [ClassicSimilarity], result of:
              0.05601155 = score(doc=2647,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.40744454 = fieldWeight in 2647, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2647)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Journal of information science. 27(2001) no.5, S.303-310
  9. Beaulieu, M.: Approaches to user-based studies in information seeking and retrieval : a Sheffield perspective (2003) 0.03
    0.031830467 = product of:
      0.047745697 = sum of:
        0.01973992 = weight(_text_:of in 4692) [ClassicSimilarity], result of:
          0.01973992 = score(doc=4692,freq=2.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.24188137 = fieldWeight in 4692, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.109375 = fieldNorm(doc=4692)
        0.028005775 = product of:
          0.05601155 = sum of:
            0.05601155 = weight(_text_:science in 4692) [ClassicSimilarity], result of:
              0.05601155 = score(doc=4692,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.40744454 = fieldWeight in 4692, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4692)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Journal of information science. 29(2003) no.4, S.239-248
  10. Bar-Ilan, J.: ¬The Web as an information source on informetrics? : A content analysis (2000) 0.03
    0.03165127 = product of:
      0.047476903 = sum of:
        0.03050284 = weight(_text_:of in 4587) [ClassicSimilarity], result of:
          0.03050284 = score(doc=4587,freq=26.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.37376386 = fieldWeight in 4587, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4587)
        0.016974064 = product of:
          0.033948127 = sum of:
            0.033948127 = weight(_text_:science in 4587) [ClassicSimilarity], result of:
              0.033948127 = score(doc=4587,freq=4.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.24694869 = fieldWeight in 4587, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4587)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This article addresses the question of whether the Web can serve as an information source for research. Specifically, it analyzes by way of content analysis the Web pages retrieved by the major search engines on a particular date (June 7, 1998), as a result of the query 'informetrics OR informetric'. In 807 out of the 942 retrieved pages, the search terms were mentioned in the context of information science. Over 70% of the pages contained only indirect information on the topic, in the form of hypertext links and bibliographical references without annotation. The bibliographical references extracted from the Web pages were analyzed, and lists of most productive authors, most cited authors, works, and sources were compiled. The list of reference obtained from the Web was also compared to data retrieved from commercial databases. For most cases, the list of references extracted from the Web outperformed the commercial, bibliographic databases. The results of these comparisons indicate that valuable, freely available data is hidden in the Web waiting to be extracted from the millions of Web pages
    Source
    Journal of the American Society for Information Science. 51(2000) no.5, S.432-443
  11. Petrelli, D.: On the role of user-centred evaluation in the advancement of interactive information retrieval (2008) 0.03
    0.02937039 = product of:
      0.044055585 = sum of:
        0.02637858 = weight(_text_:of in 2026) [ClassicSimilarity], result of:
          0.02637858 = score(doc=2026,freq=28.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.32322758 = fieldWeight in 2026, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2026)
        0.017677005 = product of:
          0.03535401 = sum of:
            0.03535401 = weight(_text_:22 in 2026) [ClassicSimilarity], result of:
              0.03535401 = score(doc=2026,freq=2.0), product of:
                0.18275474 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05218836 = queryNorm
                0.19345059 = fieldWeight in 2026, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2026)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This paper discusses the role of user-centred evaluations as an essential method for researching interactive information retrieval. It draws mainly on the work carried out during the Clarity Project where different user-centred evaluations were run during the lifecycle of a cross-language information retrieval system. The iterative testing was not only instrumental to the development of a usable system, but it enhanced our knowledge of the potential, impact, and actual use of cross-language information retrieval technology. Indeed the role of the user evaluation was dual: by testing a specific prototype it was possible to gain a micro-view and assess the effectiveness of each component of the complex system; by cumulating the result of all the evaluations (in total 43 people were involved) it was possible to build a macro-view of how cross-language retrieval would impact on users and their tasks. By showing the richness of results that can be acquired, this paper aims at stimulating researchers into considering user-centred evaluations as a flexible, adaptable and comprehensive technique for investigating non-traditional information access systems.
    Footnote
    Beitrag eines Themenbereichs: Evaluation of Interactive Information Retrieval Systems
    Source
    Information processing and management. 44(2008) no.1, S.22-38
  12. Robins, D.: Shifts of focus on various aspects of user information problems during interactive information retrieval (2000) 0.03
    0.029104514 = product of:
      0.04365677 = sum of:
        0.031654295 = weight(_text_:of in 4995) [ClassicSimilarity], result of:
          0.031654295 = score(doc=4995,freq=28.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.38787308 = fieldWeight in 4995, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4995)
        0.012002475 = product of:
          0.02400495 = sum of:
            0.02400495 = weight(_text_:science in 4995) [ClassicSimilarity], result of:
              0.02400495 = score(doc=4995,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.17461908 = fieldWeight in 4995, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4995)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The author presents the results of additional analyses of shifts of focus in IR interaction. Results indicate that users and search intermediaries work toward search goals in nonlinear fashion. Twenty interactions between 20 different users and one of four different search intermediaries were examined. Analysis of discourse between the two parties during interactive information retrieval (IR) shows changes in topic occurs, on average, every seven utterances. These twenty interactions included some 9,858 utterances and 1,439 foci. Utterances are defined as any uninterrupted sound, statement, gesture, etc., made by a participant in the discourse dyad. These utterances are segmented by the researcher according to their intentional focus, i.e., the topic on which the conversation between the user and search intermediary focus until the focus changes (i.e., shifts of focus). In all but two of the 20 interactions, the search intermediary initiated a majority of shifts of focus. Six focus categories were observed. These were foci dealing with: documents; evaluation of search results; search strategies; IR system; topic of the search; and information about the user
    Source
    Journal of the American Society for Information Science. 51(2000) no.10, S.913-928
  13. Pirkola, A.; Järvelin, K.: Employing the resolution power of search keys (2001) 0.03
    0.02907518 = product of:
      0.043612767 = sum of:
        0.02960988 = weight(_text_:of in 5907) [ClassicSimilarity], result of:
          0.02960988 = score(doc=5907,freq=18.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.36282203 = fieldWeight in 5907, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5907)
        0.0140028875 = product of:
          0.028005775 = sum of:
            0.028005775 = weight(_text_:science in 5907) [ClassicSimilarity], result of:
              0.028005775 = score(doc=5907,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.20372227 = fieldWeight in 5907, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5907)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Search key resolution power is analyzed in the context of a request, i.e., among the set of search keys for the request. Methods of characterizing the resolution power of keys automatically are studied, and the effects search keys of varying resolution power have on retrieval effectiveness are analyzed. It is shown that it often is possible to identify the best key of a query while the discrimination between the remaining keys presents problems. It is also shown that query performance is improved by suitably using the best key in a structured query. The tests were run with InQuery in a subcollection of the TREC collection, which contained some 515,000 documents
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.7, S.575-583
  14. Huang, M.-h.; Wang, H.-y.: ¬The influence of document presentation order and number of documents judged an users' judgments of relevance (2004) 0.03
    0.027946234 = product of:
      0.04191935 = sum of:
        0.027916465 = weight(_text_:of in 2885) [ClassicSimilarity], result of:
          0.027916465 = score(doc=2885,freq=16.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.34207192 = fieldWeight in 2885, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2885)
        0.0140028875 = product of:
          0.028005775 = sum of:
            0.028005775 = weight(_text_:science in 2885) [ClassicSimilarity], result of:
              0.028005775 = score(doc=2885,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.20372227 = fieldWeight in 2885, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2885)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This article attempts to verify the hypothesis of the document presentation order by an empirical, two-stage experiment. It aims to identify the relationship between number of documents judged and order effect. The results indicate that significant order effect takes place when 15 and 30 documents are presented. Sets with 45 and 60 documents still reveal the order effect. However, subjects are not influenced by order of presentation when the set of documents has 5 and 75 members, respectively.
    Source
    Journal of the American Society for Information Science and Technology. 55(2004) no.11, S.970-979
  15. Vakkari, P.; Sormunen, E.: ¬The influence of relevance levels an the effectiveness of interactive information retrieval (2004) 0.03
    0.027539104 = product of:
      0.041308656 = sum of:
        0.029306183 = weight(_text_:of in 2884) [ClassicSimilarity], result of:
          0.029306183 = score(doc=2884,freq=24.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.3591007 = fieldWeight in 2884, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2884)
        0.012002475 = product of:
          0.02400495 = sum of:
            0.02400495 = weight(_text_:science in 2884) [ClassicSimilarity], result of:
              0.02400495 = score(doc=2884,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.17461908 = fieldWeight in 2884, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2884)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    In this paper, we focus an the effect of graded relevance an the results of interactive information retrieval (IR) experiments based an assigned search tasks in a test collection. A group of 26 subjects searched for four Text REtrieval Conference (TREC) topics using automatic and interactive query expansion based an relevance feedback. The TREC- and user-suggested pools of relevant documents were reassessed an a four-level relevance scale. The results show that the users could identify nearly all highly relevant documents and about half of the marginal ones. Users also selected a fair number of irrelevant documents for query expansion. The findings suggest that the effectiveness of query expansion is closely related to the searchers' success in retrieving and identifying highly relevant documents for feedback. The implications of the results an interpreting the findings of past experiments with liberal relevance thresholds are also discussed.
    Source
    Journal of the American Society for Information Science and Technology. 55(2004) no.11, S.963-969
  16. Alemayehu, N.: Analysis of performance variation using quey expansion (2003) 0.03
    0.02670734 = product of:
      0.04006101 = sum of:
        0.028058534 = weight(_text_:of in 1454) [ClassicSimilarity], result of:
          0.028058534 = score(doc=1454,freq=22.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.34381276 = fieldWeight in 1454, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=1454)
        0.012002475 = product of:
          0.02400495 = sum of:
            0.02400495 = weight(_text_:science in 1454) [ClassicSimilarity], result of:
              0.02400495 = score(doc=1454,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.17461908 = fieldWeight in 1454, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1454)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Information retrieval performance evaluation is commonly made based an the classical recall and precision based figures or graphs. However, important information indicating causes for variation may remain hidden under the average recall and precision figures. Identifying significant causes for variation can help researchers and developers to focus an opportunities for improvement that underlay the averages. This article presents a case study showing the potential of a statistical repeated measures analysis of variance for testing the significance of factors in retrieval performance variation. The TREC-9 Query Track performance data is used as a case study and the factors studied are retrieval method, topic, and their interaction. The results show that retrieval method, topic, and their interaction are all significant. A topic level analysis is also made to see the nature of variation in the performance of retrieval methods across topics. The observed retrieval performances of expansion runs are truly significant improvements for most of the topics. Analyses of the effect of query expansion an document ranking confirm that expansion affects ranking positively.
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.5, S.379-391
  17. Kilgour, F.G.; Moran, B.B.: Surname plus recallable title word searches for known items by scholars (2000) 0.03
    0.025708806 = product of:
      0.038563207 = sum of:
        0.02255991 = weight(_text_:of in 4296) [ClassicSimilarity], result of:
          0.02255991 = score(doc=4296,freq=8.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.27643585 = fieldWeight in 4296, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=4296)
        0.0160033 = product of:
          0.0320066 = sum of:
            0.0320066 = weight(_text_:science in 4296) [ClassicSimilarity], result of:
              0.0320066 = score(doc=4296,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.23282544 = fieldWeight in 4296, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4296)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This experiment searches an online library catalog employing author surnames, plus title words of books in citations of 8 scholarly works whose authors selected the title words used as being recallable. Searches comprising surname together with two recallable title words, or one if only one was available, yielded a single-screen miniature catalog (minicat) 99.0% of the time
    Source
    Journal of the American Society for Information Science. 51(2000) no.1, S.83-89
  18. Bar-Ilan, J.: Methods for measuring search engine performance over time (2002) 0.03
    0.025708806 = product of:
      0.038563207 = sum of:
        0.02255991 = weight(_text_:of in 305) [ClassicSimilarity], result of:
          0.02255991 = score(doc=305,freq=8.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.27643585 = fieldWeight in 305, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=305)
        0.0160033 = product of:
          0.0320066 = sum of:
            0.0320066 = weight(_text_:science in 305) [ClassicSimilarity], result of:
              0.0320066 = score(doc=305,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.23282544 = fieldWeight in 305, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0625 = fieldNorm(doc=305)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This study introduces methods for evaluating search engine performance over a time period. Several measures are defined, which as a whole describe search engine functionality over time. The necessary setup for such studies is described, and the use of these measures is illustrated through a specific example. The set of measures introduced here may serve as a guideline for the search engines for testing and improving their functionality. We recommend setting up a standard suite of measures for evaluating search engine performance.
    Source
    Journal of the American Society for Information Science and technology. 53(2002) no.4, S.308-319
  19. Robertson, S.: On the history of evaluation in IR (2009) 0.03
    0.025708806 = product of:
      0.038563207 = sum of:
        0.02255991 = weight(_text_:of in 3653) [ClassicSimilarity], result of:
          0.02255991 = score(doc=3653,freq=8.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.27643585 = fieldWeight in 3653, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=3653)
        0.0160033 = product of:
          0.0320066 = sum of:
            0.0320066 = weight(_text_:science in 3653) [ClassicSimilarity], result of:
              0.0320066 = score(doc=3653,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.23282544 = fieldWeight in 3653, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3653)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This paper is a personal take on the history of evaluation experiments in information retrieval. It describes some of the early experiments that were formative in our understanding, and goes on to discuss the current dominance of TREC (the Text REtrieval Conference) and to assess its impact.
    Source
    Information science in transition, Ed.: A. Gilchrist
  20. Hemminger, B.M.; Saelim, B.; Sullivan, P.F.; Vision, T.J.: Comparison of full-text searching to metadata searching for genes in two biomedical literature cohorts (2007) 0.03
    0.025467968 = product of:
      0.03820195 = sum of:
        0.028199887 = weight(_text_:of in 1327) [ClassicSimilarity], result of:
          0.028199887 = score(doc=1327,freq=32.0), product of:
            0.08160993 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.05218836 = queryNorm
            0.34554482 = fieldWeight in 1327, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1327)
        0.010002062 = product of:
          0.020004123 = sum of:
            0.020004123 = weight(_text_:science in 1327) [ClassicSimilarity], result of:
              0.020004123 = score(doc=1327,freq=2.0), product of:
                0.13747036 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.05218836 = queryNorm
                0.1455159 = fieldWeight in 1327, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1327)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Researchers have traditionally used bibliographic databases to search out information. Today, the full-text of resources is increasingly available for searching, and more researchers are performing full-text searches. This study compares differences in the number of articles discovered between metadata and full-text searches of the same literature cohort when searching for gene names in two biomedical literature domains. Three reviewers additionally ranked 100 articles in each domain. Significantly more articles were discovered via full-text searching; however, the precision of full-text searching also is significantly lower than that of metadata searching. Certain features of articles correlated with higher relevance ratings. A significant feature measured was the number of matches of the search term in the full-text of the article, with a larger number of matches having a statistically significant higher usefulness (i.e., relevance) rating. By using the number of hits of the search term in the full-text to rank the importance of the article, performance of full-text searching was improved so that both recall and precision were as good as or better than that for metadata searching. This suggests that full-text searching alone may be sufficient, and that metadata searching as a surrogate is not necessary.
    Source
    Journal of the American Society for Information Science and Technology. 58(2007) no.14, S.2341-2352

Types

  • a 88
  • m 2
  • el 1
  • s 1
  • More… Less…