Search (9 results, page 1 of 1)

  • × theme_ss:"Retrievalstudien"
  • × year_i:[2000 TO 2010}
  1. Abdou, S.; Savoy, J.: Searching in Medline : query expansion and manual indexing evaluation (2008) 0.01
    0.014269808 = product of:
      0.042809423 = sum of:
        0.042809423 = product of:
          0.085618846 = sum of:
            0.085618846 = weight(_text_:van in 2062) [ClassicSimilarity], result of:
              0.085618846 = score(doc=2062,freq=2.0), product of:
                0.23160313 = queryWeight, product of:
                  5.5765896 = idf(docFreq=454, maxDocs=44218)
                  0.04153132 = queryNorm
                0.36967915 = fieldWeight in 2062, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.5765896 = idf(docFreq=454, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2062)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Based on a relatively large subset representing one third of the Medline collection, this paper evaluates ten different IR models, including recent developments in both probabilistic and language models. We show that the best performing IR models is a probabilistic model developed within the Divergence from Randomness framework [Amati, G., & van Rijsbergen, C.J. (2002) Probabilistic models of information retrieval based on measuring the divergence from randomness. ACM-Transactions on Information Systems 20(4), 357-389], which result in 170% enhancements in mean average precision when compared to the classical tf idf vector-space model. This paper also reports on our impact evaluations on the retrieval effectiveness of manually assigned descriptors (MeSH or Medical Subject Headings), showing that by including these terms retrieval performance can improve from 2.4% to 13.5%, depending on the underling IR model. Finally, we design a new general blind-query expansion approach showing improved retrieval performances compared to those obtained using the Rocchio approach.
  2. Ruthven, I.; Lalmas, M.; Rijsbergen, K. van: Combining and selecting characteristics of information use (2002) 0.01
    0.013453705 = product of:
      0.040361114 = sum of:
        0.040361114 = product of:
          0.08072223 = sum of:
            0.08072223 = weight(_text_:van in 5208) [ClassicSimilarity], result of:
              0.08072223 = score(doc=5208,freq=4.0), product of:
                0.23160313 = queryWeight, product of:
                  5.5765896 = idf(docFreq=454, maxDocs=44218)
                  0.04153132 = queryNorm
                0.34853685 = fieldWeight in 5208, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.5765896 = idf(docFreq=454, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5208)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Ruthven, Lalmas, and van Rijsbergen use traditional term importance measures like inverse document frequency, noise, based upon in-document frequency, and term frequency supplemented by theme value which is calculated from differences of expected positions of words in a text from their actual positions, on the assumption that even distribution indicates term association with a main topic, and context, which is based on a query term's distance from the nearest other query term relative to the average expected distribution of all query terms in the document. They then define document characteristics like specificity, the sum of all idf values in a document over the total terms in the document, or document complexity, measured by the documents average idf value; and information to noise ratio, info-noise, tokens after stopping and stemming over tokens before these processes, measuring the ratio of useful and non-useful information in a document. Retrieval tests are then carried out using each characteristic, combinations of the characteristics, and relevance feedback to determine the correct combination of characteristics. A file ranks independently of query terms by both specificity and info-noise, but if presence of a query term is required unique rankings are generated. Tested on five standard collections the traditional characteristics out preformed the new characteristics, which did, however, out preform random retrieval. All possible combinations of characteristics were also tested both with and without a set of scaling weights applied. All characteristics can benefit by combination with another characteristic or set of characteristics and performance as a single characteristic is a good indicator of performance in combination. Larger combinations tended to be more effective than smaller ones and weighting increased precision measures of middle ranking combinations but decreased the ranking of poorer combinations. The best combinations vary for each collection, and in some collections with the addition of weighting. Finally, with all documents ranked by the all characteristics combination, they take the top 30 documents and calculate the characteristic scores for each term in both the relevant and the non-relevant sets. Then taking for each query term the characteristics whose average was higher for relevant than non-relevant documents the documents are re-ranked. The relevance feedback method of selecting characteristics can select a good set of characteristics for query terms.
  3. Voorhees, E.M.; Harman, D.: Overview of the Sixth Text REtrieval Conference (TREC-6) (2000) 0.01
    0.013129482 = product of:
      0.039388444 = sum of:
        0.039388444 = product of:
          0.07877689 = sum of:
            0.07877689 = weight(_text_:22 in 6438) [ClassicSimilarity], result of:
              0.07877689 = score(doc=6438,freq=2.0), product of:
                0.14543562 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04153132 = queryNorm
                0.5416616 = fieldWeight in 6438, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6438)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    11. 8.2001 16:22:19
  4. Dresel, R.; Hörnig, D.; Kaluza, H.; Peter, A.; Roßmann, A.; Sieber, W.: Evaluation deutscher Web-Suchwerkzeuge : Ein vergleichender Retrievaltest (2001) 0.01
    0.0075025614 = product of:
      0.022507684 = sum of:
        0.022507684 = product of:
          0.04501537 = sum of:
            0.04501537 = weight(_text_:22 in 261) [ClassicSimilarity], result of:
              0.04501537 = score(doc=261,freq=2.0), product of:
                0.14543562 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04153132 = queryNorm
                0.30952093 = fieldWeight in 261, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=261)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Die deutschen Suchmaschinen, Abacho, Acoon, Fireball und Lycos sowie die Web-Kataloge Web.de und Yahoo! werden einem Qualitätstest nach relativem Recall, Precision und Availability unterzogen. Die Methoden der Retrievaltests werden vorgestellt. Im Durchschnitt werden bei einem Cut-Off-Wert von 25 ein Recall von rund 22%, eine Precision von knapp 19% und eine Verfügbarkeit von 24% erreicht
  5. ¬The Eleventh Text Retrieval Conference, TREC 2002 (2003) 0.01
    0.0075025614 = product of:
      0.022507684 = sum of:
        0.022507684 = product of:
          0.04501537 = sum of:
            0.04501537 = weight(_text_:22 in 4049) [ClassicSimilarity], result of:
              0.04501537 = score(doc=4049,freq=2.0), product of:
                0.14543562 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04153132 = queryNorm
                0.30952093 = fieldWeight in 4049, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4049)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Proceedings of the llth TREC-conference held in Gaithersburg, Maryland (USA), November 19-22, 2002. Aim of the conference was discussion an retrieval and related information-seeking tasks for large test collection. 93 research groups used different techniques, for information retrieval from the same large database. This procedure makes it possible to compare the results. The tasks are: Cross-language searching, filtering, interactive searching, searching for novelty, question answering, searching for video shots, and Web searching.
  6. Leininger, K.: Interindexer consistency in PsychINFO (2000) 0.01
    0.0056269206 = product of:
      0.016880762 = sum of:
        0.016880762 = product of:
          0.033761524 = sum of:
            0.033761524 = weight(_text_:22 in 2552) [ClassicSimilarity], result of:
              0.033761524 = score(doc=2552,freq=2.0), product of:
                0.14543562 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04153132 = queryNorm
                0.23214069 = fieldWeight in 2552, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2552)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    9. 2.1997 18:44:22
  7. King, D.W.: Blazing new trails : in celebration of an audacious career (2000) 0.00
    0.004689101 = product of:
      0.014067302 = sum of:
        0.014067302 = product of:
          0.028134605 = sum of:
            0.028134605 = weight(_text_:22 in 1184) [ClassicSimilarity], result of:
              0.028134605 = score(doc=1184,freq=2.0), product of:
                0.14543562 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04153132 = queryNorm
                0.19345059 = fieldWeight in 1184, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1184)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 9.1997 19:16:05
  8. Petrelli, D.: On the role of user-centred evaluation in the advancement of interactive information retrieval (2008) 0.00
    0.004689101 = product of:
      0.014067302 = sum of:
        0.014067302 = product of:
          0.028134605 = sum of:
            0.028134605 = weight(_text_:22 in 2026) [ClassicSimilarity], result of:
              0.028134605 = score(doc=2026,freq=2.0), product of:
                0.14543562 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04153132 = queryNorm
                0.19345059 = fieldWeight in 2026, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2026)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    Information processing and management. 44(2008) no.1, S.22-38
  9. Larsen, B.; Ingwersen, P.; Lund, B.: Data fusion according to the principle of polyrepresentation (2009) 0.00
    0.0037512807 = product of:
      0.011253842 = sum of:
        0.011253842 = product of:
          0.022507684 = sum of:
            0.022507684 = weight(_text_:22 in 2752) [ClassicSimilarity], result of:
              0.022507684 = score(doc=2752,freq=2.0), product of:
                0.14543562 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04153132 = queryNorm
                0.15476047 = fieldWeight in 2752, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2752)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 3.2009 18:48:28