Search (463 results, page 1 of 24)

  • × theme_ss:"Retrievalstudien"
  1. Chevallet, J.-P.; Bruandet, M.F.: Impact de l'utilisation de multi terms sur la qualité des résponses dùn système de recherche d'information a indexation automatique (1999) 0.15
    0.14763463 = product of:
      0.19684616 = sum of:
        0.007327141 = weight(_text_:a in 6253) [ClassicSimilarity], result of:
          0.007327141 = score(doc=6253,freq=4.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.14413087 = fieldWeight in 6253, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=6253)
        0.14859463 = weight(_text_:et in 6253) [ClassicSimilarity], result of:
          0.14859463 = score(doc=6253,freq=6.0), product of:
            0.20686594 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.044089027 = queryNorm
            0.7183137 = fieldWeight in 6253, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0625 = fieldNorm(doc=6253)
        0.040924385 = product of:
          0.08184877 = sum of:
            0.08184877 = weight(_text_:al in 6253) [ClassicSimilarity], result of:
              0.08184877 = score(doc=6253,freq=2.0), product of:
                0.20205697 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.044089027 = queryNorm
                0.4050777 = fieldWeight in 6253, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6253)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Series
    Collection travaux et recherches; UL3
    Source
    Organisation des connaissances en vue de leur intégration dans les systèmes de représentation et de recherche d'information. Ed.: J. Maniez, et al
    Type
    a
  2. Belkin, N.J.: ¬An overview of results from Rutgers' investigations of interactive information retrieval (1998) 0.10
    0.104416266 = product of:
      0.13922168 = sum of:
        0.004579463 = weight(_text_:a in 2339) [ClassicSimilarity], result of:
          0.004579463 = score(doc=2339,freq=4.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.090081796 = fieldWeight in 2339, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2339)
        0.053619467 = weight(_text_:et in 2339) [ClassicSimilarity], result of:
          0.053619467 = score(doc=2339,freq=2.0), product of:
            0.20686594 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.044089027 = queryNorm
            0.2591991 = fieldWeight in 2339, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2339)
        0.081022754 = sum of:
          0.051155485 = weight(_text_:al in 2339) [ClassicSimilarity], result of:
            0.051155485 = score(doc=2339,freq=2.0), product of:
              0.20205697 = queryWeight, product of:
                4.582931 = idf(docFreq=1228, maxDocs=44218)
                0.044089027 = queryNorm
              0.25317356 = fieldWeight in 2339, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.582931 = idf(docFreq=1228, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2339)
          0.029867273 = weight(_text_:22 in 2339) [ClassicSimilarity], result of:
            0.029867273 = score(doc=2339,freq=2.0), product of:
              0.15439226 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.044089027 = queryNorm
              0.19345059 = fieldWeight in 2339, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2339)
      0.75 = coord(3/4)
    
    Abstract
    Over the last 4 years, the Information Interaction Laboratory at Rutgers' School of communication, Information and Library Studies has performed a series of investigations concerned with various aspects of people's interactions with advanced information retrieval (IR) systems. We have benn especially concerned with understanding not just what people do, and why, and with what effect, but also with what they would like to do, and how they attempt to accomplish it, and with what difficulties. These investigations have led to some quite interesting conclusions about the nature and structure of people's interactions with information, about support for cooperative human-computer interaction in query reformulation, and about the value of visualization of search results for supporting various forms of interaction with information. In this discussion, I give an overview of the research program and its projects, present representative results from the projects, and discuss some implications of these results for support of subject searching in information retrieval systems
    Date
    22. 9.1997 19:16:05
    Source
    Visualizing subject access for 21st century information resources: Papers presented at the 1997 Clinic on Library Applications of Data Processing, 2-4 Mar 1997, Graduate School of Library and Information Science, University of Illinois at Urbana-Champaign. Ed.: P.A. Cochrane et al
    Type
    a
  3. Gilchrist, A.: Research and consultancy (1998) 0.10
    0.10053201 = product of:
      0.13404268 = sum of:
        0.007327141 = weight(_text_:a in 1394) [ClassicSimilarity], result of:
          0.007327141 = score(doc=1394,freq=4.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.14413087 = fieldWeight in 1394, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=1394)
        0.08579115 = weight(_text_:et in 1394) [ClassicSimilarity], result of:
          0.08579115 = score(doc=1394,freq=2.0), product of:
            0.20686594 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.044089027 = queryNorm
            0.41471857 = fieldWeight in 1394, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0625 = fieldNorm(doc=1394)
        0.040924385 = product of:
          0.08184877 = sum of:
            0.08184877 = weight(_text_:al in 1394) [ClassicSimilarity], result of:
              0.08184877 = score(doc=1394,freq=2.0), product of:
                0.20205697 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.044089027 = queryNorm
                0.4050777 = fieldWeight in 1394, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1394)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Source
    Library and information work worldwide 1998. Ed.: M.B. Line et al
    Type
    a
  4. Hider, P.: ¬The search value added by professional indexing to a bibliographic database (2017) 0.09
    0.09205617 = product of:
      0.122741565 = sum of:
        0.010739793 = weight(_text_:a in 3868) [ClassicSimilarity], result of:
          0.010739793 = score(doc=3868,freq=22.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.21126054 = fieldWeight in 3868, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3868)
        0.07582938 = weight(_text_:et in 3868) [ClassicSimilarity], result of:
          0.07582938 = score(doc=3868,freq=4.0), product of:
            0.20686594 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.044089027 = queryNorm
            0.3665629 = fieldWeight in 3868, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3868)
        0.03617239 = product of:
          0.07234478 = sum of:
            0.07234478 = weight(_text_:al in 3868) [ClassicSimilarity], result of:
              0.07234478 = score(doc=3868,freq=4.0), product of:
                0.20205697 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.044089027 = queryNorm
                0.3580415 = fieldWeight in 3868, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3868)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Abstract
    Gross et al. (2015) have demonstrated that about a quarter of hits would typically be lost to keyword searchers if contemporary academic library catalogs dropped their controlled subject headings. This paper reports on an analysis of the loss levels that would result if a bibliographic database, namely the Australian Education Index (AEI), were missing the subject descriptors and identifiers assigned by its professional indexers, employing the methodology developed by Gross and Taylor (2005), and later by Gross et al. (2015). The results indicate that AEI users would lose a similar proportion of hits per query to that experienced by library catalog users: on average, 27% of the resources found by a sample of keyword queries on the AEI database would not have been found without the subject indexing, based on the Australian Thesaurus of Education Descriptors (ATED). The paper also discusses the methodological limitations of these studies, pointing out that real-life users might still find some of the resources missed by a particular query through follow-up searches, while additional resources might also be found through iterative searching on the subject vocabulary. The paper goes on to describe a new research design, based on a before - and - after experiment, which addresses some of these limitations. It is argued that this alternative design will provide a more realistic picture of the value that professionally assigned subject indexing and controlled subject vocabularies can add to literature searching of a more scholarly and thorough kind.
    Type
    a
  5. Hider, P.: ¬The search value added by professional indexing to a bibliographic database (2018) 0.09
    0.09128721 = product of:
      0.121716276 = sum of:
        0.0097145075 = weight(_text_:a in 4300) [ClassicSimilarity], result of:
          0.0097145075 = score(doc=4300,freq=18.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.19109234 = fieldWeight in 4300, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4300)
        0.07582938 = weight(_text_:et in 4300) [ClassicSimilarity], result of:
          0.07582938 = score(doc=4300,freq=4.0), product of:
            0.20686594 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.044089027 = queryNorm
            0.3665629 = fieldWeight in 4300, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4300)
        0.03617239 = product of:
          0.07234478 = sum of:
            0.07234478 = weight(_text_:al in 4300) [ClassicSimilarity], result of:
              0.07234478 = score(doc=4300,freq=4.0), product of:
                0.20205697 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.044089027 = queryNorm
                0.3580415 = fieldWeight in 4300, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4300)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Abstract
    Gross et al. (2015) have demonstrated that about a quarter of hits would typically be lost to keyword searchers if contemporary academic library catalogs dropped their controlled subject headings. This article reports on an investigation of the search value that subject descriptors and identifiers assigned by professional indexers add to a bibliographic database, namely the Australian Education Index (AEI). First, a similar methodology to that developed by Gross et al. (2015) was applied, with keyword searches representing a range of educational topics run on the AEI database with and without its subject indexing. The results indicated that AEI users would also lose, on average, about a quarter of hits per query. Second, an alternative research design was applied in which an experienced literature searcher was asked to find resources on a set of educational topics on an AEI database stripped of its subject indexing and then asked to search for additional resources on the same topics after the subject indexing had been reinserted. In this study, the proportion of additional resources that would have been lost had it not been for the subject indexing was again found to be about a quarter of the total resources found for each topic, on average.
    Type
    a
  6. Boros, E.; Kantor, P.B.; Neu, D.J.: Pheromonic representation of user quests by digital structures (1999) 0.09
    0.09042687 = product of:
      0.12056916 = sum of:
        0.008567391 = weight(_text_:a in 6684) [ClassicSimilarity], result of:
          0.008567391 = score(doc=6684,freq=14.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.1685276 = fieldWeight in 6684, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6684)
        0.07582938 = weight(_text_:et in 6684) [ClassicSimilarity], result of:
          0.07582938 = score(doc=6684,freq=4.0), product of:
            0.20686594 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.044089027 = queryNorm
            0.3665629 = fieldWeight in 6684, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6684)
        0.03617239 = product of:
          0.07234478 = sum of:
            0.07234478 = weight(_text_:al in 6684) [ClassicSimilarity], result of:
              0.07234478 = score(doc=6684,freq=4.0), product of:
                0.20205697 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.044089027 = queryNorm
                0.3580415 = fieldWeight in 6684, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=6684)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Abstract
    In a novel approach to information finding in networked environments, each user's specific purpose or "quest" can be represented in numerous ways. The most familiar is a list of keywords, or a natural language sentence or paragraph. More effective is an extended text that has been judged as to relevance. This forms the basis of relevance feedback, as it is used in information retrieval. In the "Ant World" project (Ant World, 1999; Kantor et al., 1999b; Kantor et al., 1999a), the items to be retrieved are not documents, but rather quests, represented by entire collections of judged documents. In order to save space and time we have developed methods for representing these complex entities in a short string of about 1,000 bytes, which we call a "Digital Information Pheromone" (DIP). The principles for determining the DIP for a given quest, and for matching DIPs to each other are presented. The effectiveness of this scheme is explored with some applications to the large judged collections of TREC documents
    Type
    a
  7. Cross-language information retrieval (1998) 0.08
    0.075842336 = product of:
      0.10112311 = sum of:
        0.009848507 = weight(_text_:a in 6299) [ClassicSimilarity], result of:
          0.009848507 = score(doc=6299,freq=74.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.19372822 = fieldWeight in 6299, product of:
              8.602325 = tf(freq=74.0), with freq of:
                74.0 = termFreq=74.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.01953125 = fieldNorm(doc=6299)
        0.05994839 = weight(_text_:et in 6299) [ClassicSimilarity], result of:
          0.05994839 = score(doc=6299,freq=10.0), product of:
            0.20686594 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.044089027 = queryNorm
            0.28979343 = fieldWeight in 6299, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.01953125 = fieldNorm(doc=6299)
        0.03132621 = product of:
          0.06265242 = sum of:
            0.06265242 = weight(_text_:al in 6299) [ClassicSimilarity], result of:
              0.06265242 = score(doc=6299,freq=12.0), product of:
                0.20205697 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.044089027 = queryNorm
                0.31007302 = fieldWeight in 6299, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=6299)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Content
    Enthält die Beiträge: GREFENSTETTE, G.: The Problem of Cross-Language Information Retrieval; DAVIS, M.W.: On the Effective Use of Large Parallel Corpora in Cross-Language Text Retrieval; BALLESTEROS, L. u. W.B. CROFT: Statistical Methods for Cross-Language Information Retrieval; Distributed Cross-Lingual Information Retrieval; Automatic Cross-Language Information Retrieval Using Latent Semantic Indexing; EVANS, D.A. u.a.: Mapping Vocabularies Using Latent Semantics; PICCHI, E. u. C. PETERS: Cross-Language Information Retrieval: A System for Comparable Corpus Querying; YAMABANA, K. u.a.: A Language Conversion Front-End for Cross-Language Information Retrieval; GACHOT, D.A. u.a.: The Systran NLP Browser: An Application of Machine Translation Technology in Cross-Language Information Retrieval; HULL, D.: A Weighted Boolean Model for Cross-Language Text Retrieval; SHERIDAN, P. u.a. Building a Large Multilingual Test Collection from Comparable News Documents; OARD; D.W. u. B.J. DORR: Evaluating Cross-Language Text Filtering Effectiveness
    Footnote
    Rez. in: Machine translation review: 1999, no.10, S.26-27 (D. Lewis): "Cross Language Information Retrieval (CLIR) addresses the growing need to access large volumes of data across language boundaries. The typical requirement is for the user to input a free form query, usually a brief description of a topic, into a search or retrieval engine which returns a list, in ranked order, of documents or web pages that are relevant to the topic. The search engine matches the terms in the query to indexed terms, usually keywords previously derived from the target documents. Unlike monolingual information retrieval, CLIR requires query terms in one language to be matched to indexed terms in another. Matching can be done by bilingual dictionary lookup, full machine translation, or by applying statistical methods. A query's success is measured in terms of recall (how many potentially relevant target documents are found) and precision (what proportion of documents found are relevant). Issues in CLIR are how to translate query terms into index terms, how to eliminate alternative translations (e.g. to decide that French 'traitement' in a query means 'treatment' and not 'salary'), and how to rank or weight translation alternatives that are retained (e.g. how to order the French terms 'aventure', 'business', 'affaire', and 'liaison' as relevant translations of English 'affair'). Grefenstette provides a lucid and useful overview of the field and the problems. The volume brings together a number of experiments and projects in CLIR. Mark Davies (New Mexico State University) describes Recuerdo, a Spanish retrieval engine which reduces translation ambiguities by scanning indexes for parallel texts; it also uses either a bilingual dictionary or direct equivalents from a parallel corpus in order to compare results for queries on parallel texts. Lisa Ballesteros and Bruce Croft (University of Massachusetts) use a 'local feedback' technique which automatically enhances a query by adding extra terms to it both before and after translation; such terms can be derived from documents known to be relevant to the query.
    Christian Fluhr at al (DIST/SMTI, France) outline the EMIR (European Multilingual Information Retrieval) and ESPRIT projects. They found that using SYSTRAN to machine translate queries and to access material from various multilingual databases produced less relevant results than a method referred to as 'multilingual reformulation' (the mechanics of which are only hinted at). An interesting technique is Latent Semantic Indexing (LSI), described by Michael Littman et al (Brown University) and, most clearly, by David Evans et al (Carnegie Mellon University). LSI involves creating matrices of documents and the terms they contain and 'fitting' related documents into a reduced matrix space. This effectively allows queries to be mapped onto a common semantic representation of the documents. Eugenio Picchi and Carol Peters (Pisa) report on a procedure to create links between translation equivalents in an Italian-English parallel corpus. The links are used to construct parallel linguistic contexts in real-time for any term or combination of terms that is being searched for in either language. Their interest is primarily lexicographic but they plan to apply the same procedure to comparable corpora, i.e. to texts which are not translations of each other but which share the same domain. Kiyoshi Yamabana et al (NEC, Japan) address the issue of how to disambiguate between alternative translations of query terms. Their DMAX (double maximise) method looks at co-occurrence frequencies between both source language words and target language words in order to arrive at the most probable translation. The statistical data for the decision are derived, not from the translation texts but independently from monolingual corpora in each language. An interactive user interface allows the user to influence the selection of terms during the matching process. Denis Gachot et al (SYSTRAN) describe the SYSTRAN NLP browser, a prototype tool which collects parsing information derived from a text or corpus previously translated with SYSTRAN. The user enters queries into the browser in either a structured or free form and receives grammatical and lexical information about the source text and/or its translation.
    The retrieved output from a query including the phrase 'big rockets' may be, for instance, a sentence containing 'giant rocket' which is semantically ranked above 'military ocket'. David Hull (Xerox Research Centre, Grenoble) describes an implementation of a weighted Boolean model for Spanish-English CLIR. Users construct Boolean-type queries, weighting each term in the query, which is then translated by an on-line dictionary before being applied to the database. Comparisons with the performance of unweighted free-form queries ('vector space' models) proved encouraging. Two contributions consider the evaluation of CLIR systems. In order to by-pass the time-consuming and expensive process of assembling a standard collection of documents and of user queries against which the performance of an CLIR system is manually assessed, Páriac Sheridan et al (ETH Zurich) propose a method based on retrieving 'seed documents'. This involves identifying a unique document in a database (the 'seed document') and, for a number of queries, measuring how fast it is retrieved. The authors have also assembled a large database of multilingual news documents for testing purposes. By storing the (fairly short) documents in a structured form tagged with descriptor codes (e.g. for topic, country and area), the test suite is easily expanded while remaining consistent for the purposes of testing. Douglas Ouard and Bonne Dorr (University of Maryland) describe an evaluation methodology which appears to apply LSI techniques in order to filter and rank incoming documents designed for testing CLIR systems. The volume provides the reader an excellent overview of several projects in CLIR. It is well supported with references and is intended as a secondary text for researchers and practitioners. It highlights the need for a good, general tutorial introduction to the field."
  8. Lancaster, F.W.: Evaluating the performance of a large computerized information system (1985) 0.06
    0.055043153 = product of:
      0.07339087 = sum of:
        0.010033101 = weight(_text_:a in 3649) [ClassicSimilarity], result of:
          0.010033101 = score(doc=3649,freq=30.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.19735932 = fieldWeight in 3649, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=3649)
        0.042895574 = weight(_text_:et in 3649) [ClassicSimilarity], result of:
          0.042895574 = score(doc=3649,freq=2.0), product of:
            0.20686594 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.044089027 = queryNorm
            0.20735928 = fieldWeight in 3649, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.03125 = fieldNorm(doc=3649)
        0.020462193 = product of:
          0.040924385 = sum of:
            0.040924385 = weight(_text_:al in 3649) [ClassicSimilarity], result of:
              0.040924385 = score(doc=3649,freq=2.0), product of:
                0.20205697 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.044089027 = queryNorm
                0.20253885 = fieldWeight in 3649, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3649)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Abstract
    F. W. Lancaster is known for his writing an the state of the art in librarylinformation science. His skill in identifying significant contributions and synthesizing literature in fields as diverse as online systems, vocabulary control, measurement and evaluation, and the paperless society have earned him esteem as a chronicler of information science. Equally deserving of repute is his own contribution to research in the discipline-his evaluation of the MEDLARS operating system. The MEDLARS study is notable for several reasons. It was the first large-scale application of retrieval experiment methodology to the evaluation of an actual operating system. As such, problems had to be faced that do not arise in laboratory-like conditions. One example is the problem of recall: how to determine, for a very large and dynamic database, the number of documents relevant to a given search request. By solving this problem and others attendant upon transferring an experimental methodology to the real world, Lancaster created a constructive procedure that could be used to improve the design and functioning of retrieval systems. The MEDLARS study is notable also for its contribution to our understanding of what constitutes a good index language and good indexing. The ideal retrieval system would be one that retrieves all and only relevant documents. The failures that occur in real operating systems, when a relevant document is not retrieved (a recall failure) or an irrelevant document is retrieved (a precision failure), can be analysed to assess the impact of various factors an the performance of the system. This is exactly what Lancaster did. He found both the MEDLARS indexing and the McSH index language to be significant factors affecting retrieval performance. The indexing, primarily because it was insufficiently exhaustive, explained a large number of recall failures. The index language, largely because of its insufficient specificity, accounted for a large number of precision failures. The purpose of identifying factors responsible for a system's failures is ultimately to improve the system. Unlike many user studies, the MEDLARS evaluation yielded recommendations that were eventually implemented.* Indexing exhaustivity was increased and the McSH index language was enriched with more specific terms and a larger entry vocabulary.
    Source
    Theory of subject analysis: a sourcebook. Ed.: L.M. Chan, et al
    Type
    a
  9. Cleverdon, C.W.; Mills, J.: ¬The testing of index language devices (1985) 0.05
    0.054248735 = product of:
      0.072331645 = sum of:
        0.008973878 = weight(_text_:a in 3643) [ClassicSimilarity], result of:
          0.008973878 = score(doc=3643,freq=24.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.17652355 = fieldWeight in 3643, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=3643)
        0.042895574 = weight(_text_:et in 3643) [ClassicSimilarity], result of:
          0.042895574 = score(doc=3643,freq=2.0), product of:
            0.20686594 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.044089027 = queryNorm
            0.20735928 = fieldWeight in 3643, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.03125 = fieldNorm(doc=3643)
        0.020462193 = product of:
          0.040924385 = sum of:
            0.040924385 = weight(_text_:al in 3643) [ClassicSimilarity], result of:
              0.040924385 = score(doc=3643,freq=2.0), product of:
                0.20205697 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.044089027 = queryNorm
                0.20253885 = fieldWeight in 3643, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3643)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Abstract
    A landmark event in the twentieth-century development of subject analysis theory was a retrieval experiment, begun in 1957, by Cyril Cleverdon, Librarian of the Cranfield Institute of Technology. For this work he received the Professional Award of the Special Libraries Association in 1962 and the Award of Merit of the American Society for Information Science in 1970. The objective of the experiment, called Cranfield I, was to test the ability of four indexing systems-UDC, Facet, Uniterm, and Alphabetic-Subject Headings-to retrieve material responsive to questions addressed to a collection of documents. The experiment was ambitious in scale, consisting of eighteen thousand documents and twelve hundred questions. Prior to Cranfield I, the question of what constitutes good indexing was approached subjectively and reference was made to assumptions in the form of principles that should be observed or user needs that should be met. Cranfield I was the first large-scale effort to use objective criteria for determining the parameters of good indexing. Its creative impetus was the definition of user satisfaction in terms of precision and recall. Out of the experiment emerged the definition of recall as the percentage of relevant documents retrieved and precision as the percentage of retrieved documents that were relevant. Operationalizing the concept of user satisfaction, that is, making it measurable, meant that it could be studied empirically and manipulated as a variable in mathematical equations. Much has been made of the fact that the experimental methodology of Cranfield I was seriously flawed. This is unfortunate as it tends to diminish Cleverdon's contribu tion, which was not methodological-such contributions can be left to benchmark researchers-but rather creative: the introduction of a new paradigm, one that proved to be eminently productive. The criticism leveled at the methodological shortcomings of Cranfield I underscored the need for more precise definitions of the variables involved in information retrieval. Particularly important was the need for a definition of the dependent variable index language. Like the definitions of precision and recall, that of index language provided a new way of looking at the indexing process. It was a re-visioning that stimulated research activity and led not only to a better understanding of indexing but also the design of better retrieval systems." Cranfield I was followed by Cranfield II. While Cranfield I was a wholesale comparison of four indexing "systems," Cranfield II aimed to single out various individual factors in index languages, called "indexing devices," and to measure how variations in these affected retrieval performance. The following selection represents the thinking at Cranfield midway between these two notable retrieval experiments.
    Source
    Theory of subject analysis: a sourcebook. Ed.: L.M. Chan, et al
    Type
    a
  10. Greisdorf, H.; O'Connor, B.: Nodes of topicality modeling user notions of on topic documents (2003) 0.03
    0.030775664 = product of:
      0.06155133 = sum of:
        0.007931862 = weight(_text_:a in 5175) [ClassicSimilarity], result of:
          0.007931862 = score(doc=5175,freq=12.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.15602624 = fieldWeight in 5175, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5175)
        0.053619467 = weight(_text_:et in 5175) [ClassicSimilarity], result of:
          0.053619467 = score(doc=5175,freq=2.0), product of:
            0.20686594 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.044089027 = queryNorm
            0.2591991 = fieldWeight in 5175, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5175)
      0.5 = coord(2/4)
    
    Abstract
    Griesdorf and O'Connor attempt to determine the aspects of a retrieved item that provide a questioner with evidence that the item is in fact on the topic searched independent of its relevance. To this end they collect data from 32 participants, 11 from the business community as well as 21 doctoral students at the University of North Texas each of whom were asked to state if they considered material that approaches a topic in each of 14 specific manners as " on topic" or "off topic." Chi-square indicates that the observed values are significantly different from expected values and the chi-square residuals for on topic judgements exceed plus or minus two in eight cases and plus two in five cases. The positive values which indicate a percentage of response greater than that from chance suggest that documents considered topical are only related to the problem at hand, contain terms that were in the query, and describe, explain or expand the topic of the query. The chi-square residuals for off topic judgements exceed plus or minus two in ten cases and plus two in four cases. The positive values suggest that documents considered not topical exhibit a contrasting, contrary, or confounding point of view, or merely spark curiosity. Such material might well be relevant, but is not judged topical. This suggests that topical appropriateness may best be achieved using the Bruza, et alia, left compositional monotonicity approach.
    Type
    a
  11. Radev, D.R.; Libner, K.; Fan, W.: Getting answers to natural language questions on the Web (2002) 0.03
    0.030430116 = product of:
      0.06086023 = sum of:
        0.0072407667 = weight(_text_:a in 5204) [ClassicSimilarity], result of:
          0.0072407667 = score(doc=5204,freq=10.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.14243183 = fieldWeight in 5204, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
        0.053619467 = weight(_text_:et in 5204) [ClassicSimilarity], result of:
          0.053619467 = score(doc=5204,freq=2.0), product of:
            0.20686594 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.044089027 = queryNorm
            0.2591991 = fieldWeight in 5204, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
      0.5 = coord(2/4)
    
    Abstract
    Seven hundred natural language questions from TREC-8 and TREC-9 were sent by Radev, Libner, and Fan to each of nine web search engines. The top 40 sites returned by each system were stored for evaluation of their productivity of correct answers. Each question per engine was scored as the sum of the reciprocal ranks of identified correct answers. The large number of zero scores gave a positive skew violating the normality assumption for ANOVA, so values were transformed to zero for no hit and one for one or more hits. The non-zero values were then square-root transformed to remove the remaining positive skew. Interactions were observed between search engine and answer type (name, place, date, et cetera), search engine and number of proper nouns in the query, search engine and the need for time limitation, and search engine and total query words. All effects were significant. Shortest queries had the highest mean scores. One or more proper nouns present provides a significant advantage. Non-time dependent queries have an advantage. Place, name, person, and text description had mean scores between .85 and .9 with date at .81 and number at .59. There were significant differences in score by search engine. Search engines found at least one correct answer in between 87.7 and 75.45 of the cases. Google and Northern Light were just short of a 90% hit rate. No evidence indicated that a particular engine was better at answering any particular sort of question.
    Type
    a
  12. Dalrymple, P.W.: Retrieval by reformulation in two library catalogs : toward a cognitive model of searching behavior (1990) 0.03
    0.02731834 = product of:
      0.05463668 = sum of:
        0.012822497 = weight(_text_:a in 5089) [ClassicSimilarity], result of:
          0.012822497 = score(doc=5089,freq=4.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.25222903 = fieldWeight in 5089, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=5089)
        0.041814182 = product of:
          0.083628364 = sum of:
            0.083628364 = weight(_text_:22 in 5089) [ClassicSimilarity], result of:
              0.083628364 = score(doc=5089,freq=2.0), product of:
                0.15439226 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044089027 = queryNorm
                0.5416616 = fieldWeight in 5089, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=5089)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Date
    22. 7.2006 18:43:54
    Type
    a
  13. Fuhr, N.; Niewelt, B.: ¬Ein Retrievaltest mit automatisch indexierten Dokumenten (1984) 0.03
    0.025440529 = product of:
      0.050881058 = sum of:
        0.009066874 = weight(_text_:a in 262) [ClassicSimilarity], result of:
          0.009066874 = score(doc=262,freq=2.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.17835285 = fieldWeight in 262, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=262)
        0.041814182 = product of:
          0.083628364 = sum of:
            0.083628364 = weight(_text_:22 in 262) [ClassicSimilarity], result of:
              0.083628364 = score(doc=262,freq=2.0), product of:
                0.15439226 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044089027 = queryNorm
                0.5416616 = fieldWeight in 262, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=262)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Date
    20.10.2000 12:22:23
    Type
    a
  14. Tomaiuolo, N.G.; Parker, J.: Maximizing relevant retrieval : keyword and natural language searching (1998) 0.03
    0.025440529 = product of:
      0.050881058 = sum of:
        0.009066874 = weight(_text_:a in 6418) [ClassicSimilarity], result of:
          0.009066874 = score(doc=6418,freq=2.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.17835285 = fieldWeight in 6418, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=6418)
        0.041814182 = product of:
          0.083628364 = sum of:
            0.083628364 = weight(_text_:22 in 6418) [ClassicSimilarity], result of:
              0.083628364 = score(doc=6418,freq=2.0), product of:
                0.15439226 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044089027 = queryNorm
                0.5416616 = fieldWeight in 6418, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6418)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Source
    Online. 22(1998) no.6, S.57-58
    Type
    a
  15. Voorhees, E.M.; Harman, D.: Overview of the Sixth Text REtrieval Conference (TREC-6) (2000) 0.03
    0.025440529 = product of:
      0.050881058 = sum of:
        0.009066874 = weight(_text_:a in 6438) [ClassicSimilarity], result of:
          0.009066874 = score(doc=6438,freq=2.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.17835285 = fieldWeight in 6438, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=6438)
        0.041814182 = product of:
          0.083628364 = sum of:
            0.083628364 = weight(_text_:22 in 6438) [ClassicSimilarity], result of:
              0.083628364 = score(doc=6438,freq=2.0), product of:
                0.15439226 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044089027 = queryNorm
                0.5416616 = fieldWeight in 6438, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6438)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Date
    11. 8.2001 16:22:19
    Type
    a
  16. Al-Maskari, A.; Sanderson, M.: ¬A review of factors influencing user satisfaction in information retrieval (2010) 0.02
    0.023456722 = product of:
      0.046913445 = sum of:
        0.011104607 = weight(_text_:a in 3447) [ClassicSimilarity], result of:
          0.011104607 = score(doc=3447,freq=12.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.21843673 = fieldWeight in 3447, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3447)
        0.03580884 = product of:
          0.07161768 = sum of:
            0.07161768 = weight(_text_:al in 3447) [ClassicSimilarity], result of:
              0.07161768 = score(doc=3447,freq=2.0), product of:
                0.20205697 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.044089027 = queryNorm
                0.35444298 = fieldWeight in 3447, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3447)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    The authors investigate factors influencing user satisfaction in information retrieval. It is evident from this study that user satisfaction is a subjective variable, which can be influenced by several factors such as system effectiveness, user effectiveness, user effort, and user characteristics and expectations. Therefore, information retrieval evaluators should consider all these factors in obtaining user satisfaction and in using it as a criterion of system effectiveness. Previous studies have conflicting conclusions on the relationship between user satisfaction and system effectiveness; this study has substantiated these findings and supports using user satisfaction as a criterion of system effectiveness.
    Type
    a
  17. Saracevic, T.: On a method for studying the structure and nature of requests in information retrieval (1983) 0.02
    0.0195131 = product of:
      0.0390262 = sum of:
        0.009158926 = weight(_text_:a in 2417) [ClassicSimilarity], result of:
          0.009158926 = score(doc=2417,freq=4.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.18016359 = fieldWeight in 2417, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=2417)
        0.029867273 = product of:
          0.059734546 = sum of:
            0.059734546 = weight(_text_:22 in 2417) [ClassicSimilarity], result of:
              0.059734546 = score(doc=2417,freq=2.0), product of:
                0.15439226 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044089027 = queryNorm
                0.38690117 = fieldWeight in 2417, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2417)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Pages
    S.22-25
    Type
    a
  18. Rijsbergen, C.J. van: ¬A test for the separation of relevant and non-relevant documents in experimental retrieval collections (1973) 0.02
    0.0182924 = product of:
      0.0365848 = sum of:
        0.01269098 = weight(_text_:a in 5002) [ClassicSimilarity], result of:
          0.01269098 = score(doc=5002,freq=12.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.24964198 = fieldWeight in 5002, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=5002)
        0.023893818 = product of:
          0.047787637 = sum of:
            0.047787637 = weight(_text_:22 in 5002) [ClassicSimilarity], result of:
              0.047787637 = score(doc=5002,freq=2.0), product of:
                0.15439226 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044089027 = queryNorm
                0.30952093 = fieldWeight in 5002, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5002)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Many retrievalexperiments are intended to discover ways of improving performance, taking the results obtained with some particular technique as a baseline. The fact that substantial alterations to a system often have little or no effect on particular collections is puzzling. This may be due to the initially poor seperation of relevant and non-relevant documents. The paper presents a procedure for characterizing this seperation for a collection, which can be used to show whether proposed modifications of the base system are likely to be useful.
    Date
    19. 3.1996 11:22:12
    Type
    a
  19. Allan, J.; Callan, J.P.; Croft, W.B.; Ballesteros, L.; Broglio, J.; Xu, J.; Shu, H.: INQUERY at TREC-5 (1997) 0.02
    0.018171806 = product of:
      0.03634361 = sum of:
        0.006476338 = weight(_text_:a in 3103) [ClassicSimilarity], result of:
          0.006476338 = score(doc=3103,freq=2.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.12739488 = fieldWeight in 3103, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=3103)
        0.029867273 = product of:
          0.059734546 = sum of:
            0.059734546 = weight(_text_:22 in 3103) [ClassicSimilarity], result of:
              0.059734546 = score(doc=3103,freq=2.0), product of:
                0.15439226 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044089027 = queryNorm
                0.38690117 = fieldWeight in 3103, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3103)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Date
    27. 2.1999 20:55:22
    Type
    a
  20. Ng, K.B.; Loewenstern, D.; Basu, C.; Hirsh, H.; Kantor, P.B.: Data fusion of machine-learning methods for the TREC5 routing tak (and other work) (1997) 0.02
    0.018171806 = product of:
      0.03634361 = sum of:
        0.006476338 = weight(_text_:a in 3107) [ClassicSimilarity], result of:
          0.006476338 = score(doc=3107,freq=2.0), product of:
            0.05083672 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.044089027 = queryNorm
            0.12739488 = fieldWeight in 3107, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=3107)
        0.029867273 = product of:
          0.059734546 = sum of:
            0.059734546 = weight(_text_:22 in 3107) [ClassicSimilarity], result of:
              0.059734546 = score(doc=3107,freq=2.0), product of:
                0.15439226 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044089027 = queryNorm
                0.38690117 = fieldWeight in 3107, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3107)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Date
    27. 2.1999 20:59:22
    Type
    a

Years

Languages

Types

  • a 443
  • s 10
  • el 9
  • r 5
  • m 4
  • p 2
  • d 1
  • More… Less…