Search (317 results, page 1 of 16)

  • × theme_ss:"Retrievalalgorithmen"
  1. Losada, D.E.; Barreiro, A.: Emebedding term similarity and inverse document frequency into a logical model of information retrieval (2003) 0.09
    0.0853441 = product of:
      0.12801614 = sum of:
        0.02830994 = weight(_text_:information in 1422) [ClassicSimilarity], result of:
          0.02830994 = score(doc=1422,freq=8.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3103276 = fieldWeight in 1422, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=1422)
        0.09970621 = sum of:
          0.0433803 = weight(_text_:systems in 1422) [ClassicSimilarity], result of:
            0.0433803 = score(doc=1422,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.2716328 = fieldWeight in 1422, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0625 = fieldNorm(doc=1422)
          0.05632591 = weight(_text_:22 in 1422) [ClassicSimilarity], result of:
            0.05632591 = score(doc=1422,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.30952093 = fieldWeight in 1422, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=1422)
      0.6666667 = coord(2/3)
    
    Abstract
    We propose a novel approach to incorporate term similarity and inverse document frequency into a logical model of information retrieval. The ability of the logic to handle expressive representations along with the use of such classical notions are promising characteristics for IR systems. The approach proposed here has been efficiently implemented and experiments against test collections are presented.
    Date
    22. 3.2003 19:27:23
    Footnote
    Beitrag eines Themenheftes: Mathematical, logical, and formal methods in information retrieval
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.4, S.285-301
  2. Chang, C.-H.; Hsu, C.-C.: Integrating query expansion and conceptual relevance feedback for personalized Web information retrieval (1998) 0.07
    0.07467609 = product of:
      0.11201413 = sum of:
        0.024771197 = weight(_text_:information in 1319) [ClassicSimilarity], result of:
          0.024771197 = score(doc=1319,freq=8.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.27153665 = fieldWeight in 1319, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1319)
        0.08724293 = sum of:
          0.03795776 = weight(_text_:systems in 1319) [ClassicSimilarity], result of:
            0.03795776 = score(doc=1319,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.23767869 = fieldWeight in 1319, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0546875 = fieldNorm(doc=1319)
          0.04928517 = weight(_text_:22 in 1319) [ClassicSimilarity], result of:
            0.04928517 = score(doc=1319,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.2708308 = fieldWeight in 1319, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=1319)
      0.6666667 = coord(2/3)
    
    Abstract
    Keyword based querying has been an immediate and efficient way to specify and retrieve related information that the user inquired. However, conventional document ranking based on an automatic assessment of document relevance to the query may not be the best approach when little information is given. Proposes an idea to integrate 2 existing techniques, query expansion and relevance feedback to achieve a concept-based information search for the Web
    Date
    1. 8.1996 22:08:06
    Source
    Computer networks and ISDN systems. 30(1998) nos.1/7, S.621-623
  3. Kanaeva, Z.: Ranking: Google und CiteSeer (2005) 0.07
    0.06983921 = product of:
      0.104758814 = sum of:
        0.017515881 = weight(_text_:information in 3276) [ClassicSimilarity], result of:
          0.017515881 = score(doc=3276,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.1920054 = fieldWeight in 3276, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3276)
        0.08724293 = sum of:
          0.03795776 = weight(_text_:systems in 3276) [ClassicSimilarity], result of:
            0.03795776 = score(doc=3276,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.23767869 = fieldWeight in 3276, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0546875 = fieldNorm(doc=3276)
          0.04928517 = weight(_text_:22 in 3276) [ClassicSimilarity], result of:
            0.04928517 = score(doc=3276,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.2708308 = fieldWeight in 3276, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=3276)
      0.6666667 = coord(2/3)
    
    Abstract
    Im Rahmen des klassischen Information Retrieval wurden verschiedene Verfahren für das Ranking sowie die Suche in einer homogenen strukturlosen Dokumentenmenge entwickelt. Die Erfolge der Suchmaschine Google haben gezeigt dass die Suche in einer zwar inhomogenen aber zusammenhängenden Dokumentenmenge wie dem Internet unter Berücksichtigung der Dokumentenverbindungen (Links) sehr effektiv sein kann. Unter den von der Suchmaschine Google realisierten Konzepten ist ein Verfahren zum Ranking von Suchergebnissen (PageRank), das in diesem Artikel kurz erklärt wird. Darüber hinaus wird auf die Konzepte eines Systems namens CiteSeer eingegangen, welches automatisch bibliographische Angaben indexiert (engl. Autonomous Citation Indexing, ACI). Letzteres erzeugt aus einer Menge von nicht vernetzten wissenschaftlichen Dokumenten eine zusammenhängende Dokumentenmenge und ermöglicht den Einsatz von Banking-Verfahren, die auf den von Google genutzten Verfahren basieren.
    Date
    20. 3.2005 16:23:22
    Source
    Information - Wissenschaft und Praxis. 56(2005) H.2, S.87-92
  4. Ravana, S.D.; Rajagopal, P.; Balakrishnan, V.: Ranking retrieval systems using pseudo relevance judgments (2015) 0.07
    0.06709334 = product of:
      0.100640014 = sum of:
        0.012511344 = weight(_text_:information in 2591) [ClassicSimilarity], result of:
          0.012511344 = score(doc=2591,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.13714671 = fieldWeight in 2591, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2591)
        0.08812867 = sum of:
          0.03834313 = weight(_text_:systems in 2591) [ClassicSimilarity], result of:
            0.03834313 = score(doc=2591,freq=4.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.24009174 = fieldWeight in 2591, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2591)
          0.04978554 = weight(_text_:22 in 2591) [ClassicSimilarity], result of:
            0.04978554 = score(doc=2591,freq=4.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.27358043 = fieldWeight in 2591, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2591)
      0.6666667 = coord(2/3)
    
    Abstract
    Purpose In a system-based approach, replicating the web would require large test collections, and judging the relevancy of all documents per topic in creating relevance judgment through human assessors is infeasible. Due to the large amount of documents that requires judgment, there are possible errors introduced by human assessors because of disagreements. The paper aims to discuss these issues. Design/methodology/approach This study explores exponential variation and document ranking methods that generate a reliable set of relevance judgments (pseudo relevance judgments) to reduce human efforts. These methods overcome problems with large amounts of documents for judgment while avoiding human disagreement errors during the judgment process. This study utilizes two key factors: number of occurrences of each document per topic from all the system runs; and document rankings to generate the alternate methods. Findings The effectiveness of the proposed method is evaluated using the correlation coefficient of ranked systems using mean average precision scores between the original Text REtrieval Conference (TREC) relevance judgments and pseudo relevance judgments. The results suggest that the proposed document ranking method with a pool depth of 100 could be a reliable alternative to reduce human effort and disagreement errors involved in generating TREC-like relevance judgments. Originality/value Simple methods proposed in this study show improvement in the correlation coefficient in generating alternate relevance judgment without human assessors while contributing to information retrieval evaluation.
    Date
    20. 1.2015 18:30:22
    18. 9.2018 18:22:56
    Source
    Aslib journal of information management. 67(2015) no.6, S.700-714
  5. Furner, J.: ¬A unifying model of document relatedness for hybrid search engines (2003) 0.07
    0.065914944 = product of:
      0.098872416 = sum of:
        0.010616227 = weight(_text_:information in 2717) [ClassicSimilarity], result of:
          0.010616227 = score(doc=2717,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.116372846 = fieldWeight in 2717, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2717)
        0.08825619 = sum of:
          0.046011757 = weight(_text_:systems in 2717) [ClassicSimilarity], result of:
            0.046011757 = score(doc=2717,freq=4.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.28811008 = fieldWeight in 2717, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.046875 = fieldNorm(doc=2717)
          0.04224443 = weight(_text_:22 in 2717) [ClassicSimilarity], result of:
            0.04224443 = score(doc=2717,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.23214069 = fieldWeight in 2717, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2717)
      0.6666667 = coord(2/3)
    
    Abstract
    Previous work an search-engine design has indicated that information-seekers may benefit from being given the opportunity to exploit multiple sources of evidence of document relatedness. Few existing systems, however, give users more than minimal control over the selections that may be made among methods of exploitation. By applying the methods of "document network analysis" (DNA), a unifying, graph-theoretic model of content-, collaboration-, and context-based systems (CCC) may be developed in which the nature of the similarities between types of document relatedness and document ranking are clarified. The usefulness of the approach to system design suggested by this model may be tested by constructing and evaluating a prototype system (UCXtra) that allows searchers to maintain control over the multiple ways in which document collections may be ranked and re-ranked.
    Date
    11. 9.2004 17:32:22
  6. Song, D.; Bruza, P.D.: Towards context sensitive information inference (2003) 0.06
    0.061105385 = product of:
      0.09165808 = sum of:
        0.0293417 = weight(_text_:information in 1428) [ClassicSimilarity], result of:
          0.0293417 = score(doc=1428,freq=22.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.32163754 = fieldWeight in 1428, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1428)
        0.06231638 = sum of:
          0.027112689 = weight(_text_:systems in 1428) [ClassicSimilarity], result of:
            0.027112689 = score(doc=1428,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.1697705 = fieldWeight in 1428, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1428)
          0.03520369 = weight(_text_:22 in 1428) [ClassicSimilarity], result of:
            0.03520369 = score(doc=1428,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.19345059 = fieldWeight in 1428, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1428)
      0.6666667 = coord(2/3)
    
    Abstract
    Humans can make hasty, but generally robust judgements about what a text fragment is, or is not, about. Such judgements are termed information inference. This article furnishes an account of information inference from a psychologistic stance. By drawing an theories from nonclassical logic and applied cognition, an information inference mechanism is proposed that makes inferences via computations of information flow through an approximation of a conceptual space. Within a conceptual space information is represented geometrically. In this article, geometric representations of words are realized as vectors in a high dimensional semantic space, which is automatically constructed from a text corpus. Two approaches were presented for priming vector representations according to context. The first approach uses a concept combination heuristic to adjust the vector representation of a concept in the light of the representation of another concept. The second approach computes a prototypical concept an the basis of exemplar trace texts and moves it in the dimensional space according to the context. Information inference is evaluated by measuring the effectiveness of query models derived by information flow computations. Results show that information flow contributes significantly to query model effectiveness, particularly with respect to precision. Moreover, retrieval effectiveness compares favorably with two probabilistic query models, and another based an semantic association. More generally, this article can be seen as a contribution towards realizing operational systems that mimic text-based human reasoning.
    Date
    22. 3.2003 19:35:46
    Footnote
    Beitrag eines Themenheftes: Mathematical, logical, and formal methods in information retrieval
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.4, S.321-334
  7. Kelledy, F.; Smeaton, A.F.: Signature files and beyond (1996) 0.06
    0.05986218 = product of:
      0.08979327 = sum of:
        0.015013612 = weight(_text_:information in 6973) [ClassicSimilarity], result of:
          0.015013612 = score(doc=6973,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.16457605 = fieldWeight in 6973, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=6973)
        0.07477966 = sum of:
          0.032535225 = weight(_text_:systems in 6973) [ClassicSimilarity], result of:
            0.032535225 = score(doc=6973,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.2037246 = fieldWeight in 6973, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.046875 = fieldNorm(doc=6973)
          0.04224443 = weight(_text_:22 in 6973) [ClassicSimilarity], result of:
            0.04224443 = score(doc=6973,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.23214069 = fieldWeight in 6973, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=6973)
      0.6666667 = coord(2/3)
    
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
  8. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.06
    0.0564239 = product of:
      0.08463585 = sum of:
        0.02830994 = weight(_text_:information in 402) [ClassicSimilarity], result of:
          0.02830994 = score(doc=402,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3103276 = fieldWeight in 402, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=402)
        0.05632591 = product of:
          0.11265182 = sum of:
            0.11265182 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.11265182 = score(doc=402,freq=2.0), product of:
                0.1819777 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051966466 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  9. Baloh, P.; Desouza, K.C.; Hackney, R.: Contextualizing organizational interventions of knowledge management systems : a design science perspectiveA domain analysis (2012) 0.05
    0.054929122 = product of:
      0.08239368 = sum of:
        0.008846856 = weight(_text_:information in 241) [ClassicSimilarity], result of:
          0.008846856 = score(doc=241,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.09697737 = fieldWeight in 241, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=241)
        0.07354683 = sum of:
          0.03834313 = weight(_text_:systems in 241) [ClassicSimilarity], result of:
            0.03834313 = score(doc=241,freq=4.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.24009174 = fieldWeight in 241, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0390625 = fieldNorm(doc=241)
          0.03520369 = weight(_text_:22 in 241) [ClassicSimilarity], result of:
            0.03520369 = score(doc=241,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.19345059 = fieldWeight in 241, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=241)
      0.6666667 = coord(2/3)
    
    Abstract
    We address how individuals' (workers) knowledge needs influence the design of knowledge management systems (KMS), enabling knowledge creation and utilization. It is evident that KMS technologies and activities are indiscriminately deployed in most organizations with little regard to the actual context of their adoption. Moreover, it is apparent that the extant literature pertaining to knowledge management projects is frequently deficient in identifying the variety of factors indicative for successful KMS. This presents an obvious business practice and research gap that requires a critical analysis of the necessary intervention that will actually improve how workers can leverage and form organization-wide knowledge. This research involved an extensive review of the literature, a grounded theory methodological approach and rigorous data collection and synthesis through an empirical case analysis (Parsons Brinckerhoff and Samsung). The contribution of this study is the formulation of a model for designing KMS based upon the design science paradigm, which aspires to create artifacts that are interdependent of people and organizations. The essential proposition is that KMS design and implementation must be contextualized in relation to knowledge needs and that these will differ for various organizational settings. The findings present valuable insights and further understanding of the way in which KMS design efforts should be focused.
    Date
    11. 6.2012 14:22:34
    Source
    Journal of the American Society for Information Science and Technology. 63(2012) no.5, S.948-966
  10. Back, J.: ¬An evaluation of relevancy ranking techniques used by Internet search engines (2000) 0.05
    0.04937091 = product of:
      0.074056365 = sum of:
        0.024771197 = weight(_text_:information in 3445) [ClassicSimilarity], result of:
          0.024771197 = score(doc=3445,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.27153665 = fieldWeight in 3445, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=3445)
        0.04928517 = product of:
          0.09857034 = sum of:
            0.09857034 = weight(_text_:22 in 3445) [ClassicSimilarity], result of:
              0.09857034 = score(doc=3445,freq=2.0), product of:
                0.1819777 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051966466 = queryNorm
                0.5416616 = fieldWeight in 3445, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3445)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Date
    25. 8.2005 17:42:22
    Source
    Library and information research news. 24(2000) no.77, S.30-34
  11. Perry, R.; Willett, P.: ¬A revies of the use of inverted files for best match searching in information retrieval systems (1983) 0.05
    0.048659682 = product of:
      0.07298952 = sum of:
        0.035031762 = weight(_text_:information in 2701) [ClassicSimilarity], result of:
          0.035031762 = score(doc=2701,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3840108 = fieldWeight in 2701, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=2701)
        0.03795776 = product of:
          0.07591552 = sum of:
            0.07591552 = weight(_text_:systems in 2701) [ClassicSimilarity], result of:
              0.07591552 = score(doc=2701,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.47535738 = fieldWeight in 2701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2701)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Journal of information science. 6(1983), S.59-66
  12. Willett, P.: Best-match text retrieval (1993) 0.04
    0.043102846 = product of:
      0.06465427 = sum of:
        0.017693711 = weight(_text_:information in 7818) [ClassicSimilarity], result of:
          0.017693711 = score(doc=7818,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.19395474 = fieldWeight in 7818, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=7818)
        0.046960555 = product of:
          0.09392111 = sum of:
            0.09392111 = weight(_text_:systems in 7818) [ClassicSimilarity], result of:
              0.09392111 = score(doc=7818,freq=6.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.5881023 = fieldWeight in 7818, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.078125 = fieldNorm(doc=7818)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Provides an introduction to the computational techniques that underlie best match searching retrieval systems. Discusses: problems of traditional Boolean systems; characteristics of best-match searching; automatic indexing; term conflation; matching of documents and queries (dealing with similarity measures, initial weights, relevance weights, and the matching algorithm); and describes operational best-match systems
    Source
    Library and information briefings. 1993, no.49, S.1-11
  13. Nakkouzi, Z.S.; Eastman, C.M.: Query formulation for handling negation in information retrieval systems (1990) 0.04
    0.04139038 = product of:
      0.06208557 = sum of:
        0.024517128 = weight(_text_:information in 3531) [ClassicSimilarity], result of:
          0.024517128 = score(doc=3531,freq=6.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.2687516 = fieldWeight in 3531, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=3531)
        0.037568443 = product of:
          0.075136885 = sum of:
            0.075136885 = weight(_text_:systems in 3531) [ClassicSimilarity], result of:
              0.075136885 = score(doc=3531,freq=6.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.4704818 = fieldWeight in 3531, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3531)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Queries containing negation are widely recognised as presenting problems for both users and systems. In information retrieval systems such problems usually manifest themselves in the use of the NOT operator. Describes an algorithm to transform Boolean queries with negated terms into queries without negation; the transformation process is based on the use of a hierarchical thesaurus. Examines a set of user requests submitted to the Thomas Cooper Library at the University of South Carolina to determine the pattern and frequency of use of negation.
    Source
    Journal of the American Society for Information Science. 41(1990) no.3, S.171-182
  14. Frants, V.I.; Shapiro, J.: Control and feedback in a documentary information retrieval system (1991) 0.04
    0.039322965 = product of:
      0.058984444 = sum of:
        0.02830994 = weight(_text_:information in 416) [ClassicSimilarity], result of:
          0.02830994 = score(doc=416,freq=8.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3103276 = fieldWeight in 416, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=416)
        0.030674506 = product of:
          0.061349012 = sum of:
            0.061349012 = weight(_text_:systems in 416) [ClassicSimilarity], result of:
              0.061349012 = score(doc=416,freq=4.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.38414678 = fieldWeight in 416, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0625 = fieldNorm(doc=416)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Addresses the problem of control in documentary information retrieval systems is analysed and it is shown why an IR system has to be looked at as an adaptive system. The algorithms of feedback are proposed and it is shown how they depend on the type of the collection of documents: static (no change in the collection between searches) and dynamic (when the change occurs between searches). The proposed algorithms are the basis for the development of the fully automated information retrieval systems
    Source
    Journal of the American Society for Information Science. 42(1991) no.9, S.623-634
  15. Loughran, H.: ¬A review of nearest neighbour information retrieval (1994) 0.04
    0.03850607 = product of:
      0.0577591 = sum of:
        0.03064641 = weight(_text_:information in 616) [ClassicSimilarity], result of:
          0.03064641 = score(doc=616,freq=6.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3359395 = fieldWeight in 616, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=616)
        0.027112689 = product of:
          0.054225378 = sum of:
            0.054225378 = weight(_text_:systems in 616) [ClassicSimilarity], result of:
              0.054225378 = score(doc=616,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.339541 = fieldWeight in 616, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.078125 = fieldNorm(doc=616)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Explains the concept of 'nearest neighbour' searching, also known as best match or ranked output, which it is claimed can overcome many of the inadequacies of traditional Boolean methods. Also points to some of the limitations. Identifies a number of commercial information retrieval systems which feature this search technique
    Source
    Information management report. 1994, August, S.11-14
  16. Aizawa, A.: ¬An information-theoretic perspective of tf-idf measures (2003) 0.04
    0.037575066 = product of:
      0.0563626 = sum of:
        0.03467245 = weight(_text_:information in 4155) [ClassicSimilarity], result of:
          0.03467245 = score(doc=4155,freq=12.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.38007212 = fieldWeight in 4155, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=4155)
        0.02169015 = product of:
          0.0433803 = sum of:
            0.0433803 = weight(_text_:systems in 4155) [ClassicSimilarity], result of:
              0.0433803 = score(doc=4155,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.2716328 = fieldWeight in 4155, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4155)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This paper presents a mathematical definition of the "probability-weighted amount of information" (PWI), a measure of specificity of terms in documents that is based on an information-theoretic view of retrieval events. The proposed PWI is expressed as a product of the occurrence probabilities of terms and their amounts of information, and corresponds well with the conventional term frequency - inverse document frequency measures that are commonly used in today's information retrieval systems. The mathematical definition of the PWI is shown, together with some illustrative examples of the calculation.
    Source
    Information processing and management. 39(2003) no.1, S.45-65
  17. Figuerola, C.G.; Zazo, A.F.; Berrocal, J.L.A.: ¬La interaccion con el usuario en los sistemas de cuperacion de informacion realimentacion por relecvancia (2002) 0.04
    0.035845123 = product of:
      0.05376768 = sum of:
        0.021232454 = weight(_text_:information in 2875) [ClassicSimilarity], result of:
          0.021232454 = score(doc=2875,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.23274569 = fieldWeight in 2875, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=2875)
        0.032535225 = product of:
          0.06507045 = sum of:
            0.06507045 = weight(_text_:systems in 2875) [ClassicSimilarity], result of:
              0.06507045 = score(doc=2875,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.4074492 = fieldWeight in 2875, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2875)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Footnote
    Übers. des Titels: User interaction in retrieval information systems through relevance feedback
  18. Hofferer, M.: Heuristic search in information retrieval (1994) 0.04
    0.035561085 = product of:
      0.053341627 = sum of:
        0.031651475 = weight(_text_:information in 1070) [ClassicSimilarity], result of:
          0.031651475 = score(doc=1070,freq=10.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3469568 = fieldWeight in 1070, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=1070)
        0.02169015 = product of:
          0.0433803 = sum of:
            0.0433803 = weight(_text_:systems in 1070) [ClassicSimilarity], result of:
              0.0433803 = score(doc=1070,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.2716328 = fieldWeight in 1070, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1070)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Describes an adaptive information retrieval system: Information Retrieval Algorithm System (IRAS); that uses heuristic searching to sample a document space and retrieve relevant documents according to users' requests; and also a learning module based on a knowledge representation system and an approximate probabilistic characterization of relevant documents; to reproduce a user classification of relevant documents and to provide a rule controlled ranking
    Source
    Information retrieval: new systems and current research. Proceedings of the 15th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Glasgow 1993. Ed.: Ruben Leon
  19. Losee, R.M.; Church Jr., L.: Are two document clusters better than one? : the cluster performance question for information retrieval (2005) 0.03
    0.034407593 = product of:
      0.05161139 = sum of:
        0.024771197 = weight(_text_:information in 3270) [ClassicSimilarity], result of:
          0.024771197 = score(doc=3270,freq=8.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.27153665 = fieldWeight in 3270, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3270)
        0.026840193 = product of:
          0.053680386 = sum of:
            0.053680386 = weight(_text_:systems in 3270) [ClassicSimilarity], result of:
              0.053680386 = score(doc=3270,freq=4.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.33612844 = fieldWeight in 3270, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3270)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    When do information retrieval systems using two document clusters provide better retrieval performance than systems using no clustering? We answer this question for one set of assumptions and suggest how this may be studied with other assumptions. The "Cluster Hypothesis" asks an empirical question about the relationships between documents and user-supplied relevance judgments, while the "Cluster Performance Question" proposed here focuses an the when and why of information retrieval or digital library performance for clustered and unclustered text databases. This may be generalized to study the relative performance of m versus n clusters.
    Source
    Journal of the American Society for Information Science and Technology. 56(2005) no.1, S.106-108
  20. Pfeifer, U.; Pennekamp, S.: Incremental processing of vague queries in interactive retrieval systems (1997) 0.03
    0.033795103 = product of:
      0.050692655 = sum of:
        0.02001815 = weight(_text_:information in 735) [ClassicSimilarity], result of:
          0.02001815 = score(doc=735,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.21943474 = fieldWeight in 735, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=735)
        0.030674506 = product of:
          0.061349012 = sum of:
            0.061349012 = weight(_text_:systems in 735) [ClassicSimilarity], result of:
              0.061349012 = score(doc=735,freq=4.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.38414678 = fieldWeight in 735, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0625 = fieldNorm(doc=735)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The application of information retrieval techniques in interactive environments requires systems capable of effeciently processing vague queries. To reach reasonable response times, new data structures and algorithms have to be developed. In this paper we describe an approach taking advantage of the conditions of interactive usage and special access paths. To have a reference we investigate text queries and compared our algorithms to the well known 'Buckley/Lewit' algorithm. We achieved significant improvements for the response times
    Source
    Hypertext - Information Retrieval - Multimedia '97: Theorien, Modelle und Implementierungen integrierter elektronischer Informationssysteme. Proceedings HIM '97. Hrsg.: N. Fuhr u.a

Languages

  • e 292
  • d 21
  • chi 1
  • m 1
  • sp 1
  • More… Less…

Types

  • a 294
  • m 12
  • el 6
  • s 5
  • r 3
  • p 2
  • x 2
  • More… Less…