Search (229 results, page 1 of 12)

  • × theme_ss:"Retrievalalgorithmen"
  1. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.08
    0.07858937 = product of:
      0.15717874 = sum of:
        0.12579992 = weight(_text_:processing in 402) [ClassicSimilarity], result of:
          0.12579992 = score(doc=402,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.7156181 = fieldWeight in 402, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.125 = fieldNorm(doc=402)
        0.03137882 = product of:
          0.094136454 = sum of:
            0.094136454 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.094136454 = score(doc=402,freq=2.0), product of:
                0.15206799 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043425296 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  2. Costa Carvalho, A. da; Rossi, C.; Moura, E.S. de; Silva, A.S. da; Fernandes, D.: LePrEF: Learn to precompute evidence fusion for efficient query evaluation (2012) 0.06
    0.06359105 = product of:
      0.1271821 = sum of:
        0.09629551 = weight(_text_:processing in 278) [ClassicSimilarity], result of:
          0.09629551 = score(doc=278,freq=12.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.547781 = fieldWeight in 278, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0390625 = fieldNorm(doc=278)
        0.030886576 = product of:
          0.046329863 = sum of:
            0.016645188 = weight(_text_:science in 278) [ClassicSimilarity], result of:
              0.016645188 = score(doc=278,freq=2.0), product of:
                0.11438741 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.043425296 = queryNorm
                0.1455159 = fieldWeight in 278, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=278)
            0.029684676 = weight(_text_:29 in 278) [ClassicSimilarity], result of:
              0.029684676 = score(doc=278,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.19432661 = fieldWeight in 278, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=278)
          0.6666667 = coord(2/3)
      0.5 = coord(2/4)
    
    Abstract
    State-of-the-art search engine ranking methods combine several distinct sources of relevance evidence to produce a high-quality ranking of results for each query. The fusion of information is currently done at query-processing time, which has a direct effect on the response time of search systems. Previous research also shows that an alternative to improve search efficiency in textual databases is to precompute term impacts at indexing time. In this article, we propose a novel alternative to precompute term impacts, providing a generic framework for combining any distinct set of sources of evidence by using a machine-learning technique. This method retains the advantages of producing high-quality results, but avoids the costs of combining evidence at query-processing time. Our method, called Learn to Precompute Evidence Fusion (LePrEF), uses genetic programming to compute a unified precomputed impact value for each term found in each document prior to query processing, at indexing time. Compared with previous research on precomputing term impacts, our method offers the advantage of providing a generic framework to precompute impact using any set of relevance evidence at any text collection, whereas previous research articles do not. The precomputed impact values are indexed and used later for computing document ranking at query-processing time. By doing so, our method effectively reduces the query processing to simple additions of such impacts. We show that this approach, while leading to results comparable to state-of-the-art ranking methods, also can lead to a significant decrease in computational costs during query processing.
    Date
    24. 6.2012 14:29:10
    Source
    Journal of the American Society for Information Science and Technology. 63(2012) no.7, S.1383-1397
  3. Okada, M.; Ando, K.; Lee, S.S.; Hayashi, Y.; Aoe, J.I.: ¬An efficient substring search method by using delayed keyword extraction (2001) 0.06
    0.059048843 = product of:
      0.118097685 = sum of:
        0.09434994 = weight(_text_:processing in 6415) [ClassicSimilarity], result of:
          0.09434994 = score(doc=6415,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.53671354 = fieldWeight in 6415, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.09375 = fieldNorm(doc=6415)
        0.02374774 = product of:
          0.07124322 = sum of:
            0.07124322 = weight(_text_:29 in 6415) [ClassicSimilarity], result of:
              0.07124322 = score(doc=6415,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.46638384 = fieldWeight in 6415, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6415)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Date
    29. 3.2002 17:24:03
    Source
    Information processing and management. 37(2001) no.5, S.741-761
  4. Cole, C.: Intelligent information retrieval: diagnosing information need : Part II: uncertainty expansion in a prototype of a diagnostic IR tool (1998) 0.06
    0.059048843 = product of:
      0.118097685 = sum of:
        0.09434994 = weight(_text_:processing in 6432) [ClassicSimilarity], result of:
          0.09434994 = score(doc=6432,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.53671354 = fieldWeight in 6432, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.09375 = fieldNorm(doc=6432)
        0.02374774 = product of:
          0.07124322 = sum of:
            0.07124322 = weight(_text_:29 in 6432) [ClassicSimilarity], result of:
              0.07124322 = score(doc=6432,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.46638384 = fieldWeight in 6432, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6432)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Date
    11. 8.2001 14:48:29
    Source
    Information processing and management. 34(1998) no.6, S.721-731
  5. MacFarlane, A.; Robertson, S.E.; McCann, J.A.: Parallel computing for passage retrieval (2004) 0.05
    0.0523217 = product of:
      0.1046434 = sum of:
        0.08895399 = weight(_text_:processing in 5108) [ClassicSimilarity], result of:
          0.08895399 = score(doc=5108,freq=4.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.5060184 = fieldWeight in 5108, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0625 = fieldNorm(doc=5108)
        0.01568941 = product of:
          0.047068227 = sum of:
            0.047068227 = weight(_text_:22 in 5108) [ClassicSimilarity], result of:
              0.047068227 = score(doc=5108,freq=2.0), product of:
                0.15206799 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043425296 = queryNorm
                0.30952093 = fieldWeight in 5108, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5108)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    In this paper methods for both speeding up passage processing and examining more passages using parallel computers are explored. The number of passages processed are varied in order to examine the effect on retrieval effectiveness and efficiency. The particular algorithm applied has previously been used to good effect in Okapi experiments at TREC. This algorithm and the mechanism for applying parallel computing to speed up processing are described.
    Date
    20. 1.2007 18:30:22
  6. Salton, G.: ¬A simple blueprint for automatic Boolean query processing (1988) 0.04
    0.044476993 = product of:
      0.17790797 = sum of:
        0.17790797 = weight(_text_:processing in 6774) [ClassicSimilarity], result of:
          0.17790797 = score(doc=6774,freq=4.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            1.0120368 = fieldWeight in 6774, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.125 = fieldNorm(doc=6774)
      0.25 = coord(1/4)
    
    Source
    Information processing and management. 24(1988) no.3, S.269-280
  7. Uratani, N.; Takeda, M.: ¬A fast string-searching algorithm for multiple patterns (1993) 0.04
    0.039365895 = product of:
      0.07873179 = sum of:
        0.06289996 = weight(_text_:processing in 6275) [ClassicSimilarity], result of:
          0.06289996 = score(doc=6275,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.35780904 = fieldWeight in 6275, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0625 = fieldNorm(doc=6275)
        0.015831826 = product of:
          0.047495477 = sum of:
            0.047495477 = weight(_text_:29 in 6275) [ClassicSimilarity], result of:
              0.047495477 = score(doc=6275,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.31092256 = fieldWeight in 6275, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6275)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Source
    Information processing and management. 29(1993) no.6, S.775-791
  8. Thompson, P.: Looking back: on relevance, probabilistic indexing and information retrieval (2008) 0.04
    0.039365895 = product of:
      0.07873179 = sum of:
        0.06289996 = weight(_text_:processing in 2074) [ClassicSimilarity], result of:
          0.06289996 = score(doc=2074,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.35780904 = fieldWeight in 2074, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0625 = fieldNorm(doc=2074)
        0.015831826 = product of:
          0.047495477 = sum of:
            0.047495477 = weight(_text_:29 in 2074) [ClassicSimilarity], result of:
              0.047495477 = score(doc=2074,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.31092256 = fieldWeight in 2074, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2074)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Date
    31. 7.2008 19:58:29
    Source
    Information processing and management. 44(2008) no.2, S.963-970
  9. Vechtomova, O.; Karamuftuoglu, M.: Lexical cohesion and term proximity in document ranking (2008) 0.04
    0.039365895 = product of:
      0.07873179 = sum of:
        0.06289996 = weight(_text_:processing in 2101) [ClassicSimilarity], result of:
          0.06289996 = score(doc=2101,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.35780904 = fieldWeight in 2101, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0625 = fieldNorm(doc=2101)
        0.015831826 = product of:
          0.047495477 = sum of:
            0.047495477 = weight(_text_:29 in 2101) [ClassicSimilarity], result of:
              0.047495477 = score(doc=2101,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.31092256 = fieldWeight in 2101, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2101)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Date
    1. 8.2008 12:29:05
    Source
    Information processing and management. 44(2008) no.4, S.1485-1502
  10. Berry, M.W.; Browne, M.: Understanding search engines : mathematical modeling and text retrieval (1999) 0.04
    0.038065713 = product of:
      0.076131426 = sum of:
        0.06671549 = weight(_text_:processing in 5777) [ClassicSimilarity], result of:
          0.06671549 = score(doc=5777,freq=4.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.3795138 = fieldWeight in 5777, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.046875 = fieldNorm(doc=5777)
        0.00941594 = product of:
          0.02824782 = sum of:
            0.02824782 = weight(_text_:science in 5777) [ClassicSimilarity], result of:
              0.02824782 = score(doc=5777,freq=4.0), product of:
                0.11438741 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.043425296 = queryNorm
                0.24694869 = fieldWeight in 5777, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5777)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    LCSH
    Text processing (Computer science)
    Subject
    Text processing (Computer science)
  11. Zhu, B.; Chen, H.: Validating a geographical image retrieval system (2000) 0.04
    0.03668678 = product of:
      0.07337356 = sum of:
        0.06671549 = weight(_text_:processing in 4769) [ClassicSimilarity], result of:
          0.06671549 = score(doc=4769,freq=4.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.3795138 = fieldWeight in 4769, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.046875 = fieldNorm(doc=4769)
        0.006658075 = product of:
          0.019974224 = sum of:
            0.019974224 = weight(_text_:science in 4769) [ClassicSimilarity], result of:
              0.019974224 = score(doc=4769,freq=2.0), product of:
                0.11438741 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.043425296 = queryNorm
                0.17461908 = fieldWeight in 4769, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4769)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    This paper summarizes a prototype geographical image retrieval system that demonstrates how to integrate image processing and information analysis techniques to support large-scale content-based image retrieval. By using an image as its interface, the prototype system addresses a troublesome aspect of traditional retrieval models, which require users to have complete knowledge of the low-level features of an image. In addition we describe an experiment to validate against that of human subjects in an effort to address the scarcity of research evaluating performance of an algorithm against that of human beings. The results of the experiment indicate that the system could do as well as human subjects in accomplishing the tasks of similarity analysis and image categorization. We also found that under some circumstances texture features of an image are insufficient to represent an geographic image. We believe, however, that our image retrieval system provides a promising approach to integrating image processing techniques and information retrieval algorithms
    Source
    Journal of the American Society for Information Science. 51(2000) no.7, S.625-634
  12. Paris, L.A.H.; Tibbo, H.R.: Freestyle vs. Boolean : a comparison of partial and exact match retrieval systems (1998) 0.03
    0.03444516 = product of:
      0.06889032 = sum of:
        0.05503747 = weight(_text_:processing in 3329) [ClassicSimilarity], result of:
          0.05503747 = score(doc=3329,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.3130829 = fieldWeight in 3329, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3329)
        0.013852848 = product of:
          0.04155854 = sum of:
            0.04155854 = weight(_text_:29 in 3329) [ClassicSimilarity], result of:
              0.04155854 = score(doc=3329,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.27205724 = fieldWeight in 3329, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3329)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Date
    12. 3.1999 10:29:27
    Source
    Information processing and management. 34(1998) nos.2/3, S.175-190
  13. Otterbacher, J.; Erkan, G.; Radev, D.R.: Biased LexRank : passage retrieval using random walks with question-based priors (2009) 0.03
    0.03444516 = product of:
      0.06889032 = sum of:
        0.05503747 = weight(_text_:processing in 2450) [ClassicSimilarity], result of:
          0.05503747 = score(doc=2450,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.3130829 = fieldWeight in 2450, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2450)
        0.013852848 = product of:
          0.04155854 = sum of:
            0.04155854 = weight(_text_:29 in 2450) [ClassicSimilarity], result of:
              0.04155854 = score(doc=2450,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.27205724 = fieldWeight in 2450, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2450)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Date
    22.11.2008 17:11:29
    Source
    Information processing and management. 45(2009) no.1, S.42-54
  14. Reddaway, S.: High speed text retrieval from large databases on a massively parallel processor (1991) 0.03
    0.03144998 = product of:
      0.12579992 = sum of:
        0.12579992 = weight(_text_:processing in 7745) [ClassicSimilarity], result of:
          0.12579992 = score(doc=7745,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.7156181 = fieldWeight in 7745, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.125 = fieldNorm(doc=7745)
      0.25 = coord(1/4)
    
    Source
    Information processing and management. 27(1991), S.311-316
  15. Wollf, J.G.: ¬A scalable technique for best-match retrieval of sequential information using metrics-guided search (1994) 0.03
    0.03140261 = product of:
      0.06280522 = sum of:
        0.05503747 = weight(_text_:processing in 5334) [ClassicSimilarity], result of:
          0.05503747 = score(doc=5334,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.3130829 = fieldWeight in 5334, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5334)
        0.0077677546 = product of:
          0.023303263 = sum of:
            0.023303263 = weight(_text_:science in 5334) [ClassicSimilarity], result of:
              0.023303263 = score(doc=5334,freq=2.0), product of:
                0.11438741 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.043425296 = queryNorm
                0.20372227 = fieldWeight in 5334, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5334)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    Describes a new technique for retrieving information by finding the best match or matches between a textual query and a textual database. The technique uses principles of beam search with a measure of probability to guide the search and prune the search tree. Unlike many methods for comparing strings, the method gives a set of alternative matches, graded by the quality of the matching. The new technique is embodies in a software simulation SP21 which runs on a conventional computer. Presnts examples showing best-match retrieval of information from a textual database. Presents analytic and emprirical evidence on the performance of the technique. It lends itself well to parallel processing. Discusses planned developments
    Source
    Journal of information science. 20(1994) no.1, S.16-28
  16. Hoenkamp, E.; Bruza, P.: How everyday language can and will boost effective information retrieval (2015) 0.03
    0.030572318 = product of:
      0.061144635 = sum of:
        0.05559624 = weight(_text_:processing in 2123) [ClassicSimilarity], result of:
          0.05559624 = score(doc=2123,freq=4.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.3162615 = fieldWeight in 2123, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2123)
        0.005548396 = product of:
          0.016645188 = sum of:
            0.016645188 = weight(_text_:science in 2123) [ClassicSimilarity], result of:
              0.016645188 = score(doc=2123,freq=2.0), product of:
                0.11438741 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.043425296 = queryNorm
                0.1455159 = fieldWeight in 2123, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2123)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    Typing 2 or 3 keywords into a browser has become an easy and efficient way to find information. Yet, typing even short queries becomes tedious on ever shrinking (virtual) keyboards. Meanwhile, speech processing is maturing rapidly, facilitating everyday language input. Also, wearable technology can inform users proactively by listening in on their conversations or processing their social media interactions. Given these developments, everyday language may soon become the new input of choice. We present an information retrieval (IR) algorithm specifically designed to accept everyday language. It integrates two paradigms of information retrieval, previously studied in isolation; one directed mainly at the surface structure of language, the other primarily at the underlying meaning. The integration was achieved by a Markov machine that encodes meaning by its transition graph, and surface structure by the language it generates. A rigorous evaluation of the approach showed, first, that it can compete with the quality of existing language models, second, that it is more effective the more verbose the input, and third, as a consequence, that it is promising for an imminent transition from keyword input, where the onus is on the user to formulate concise queries, to a modality where users can express more freely, more informal, and more natural their need for information in everyday language.
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.8, S.1546-1558
  17. Drucker, H.; Shahrary, B.; Gibbon, D.C.: Support vector machines : relevance feedback and information retrieval (2002) 0.03
    0.029524421 = product of:
      0.059048843 = sum of:
        0.04717497 = weight(_text_:processing in 2581) [ClassicSimilarity], result of:
          0.04717497 = score(doc=2581,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.26835677 = fieldWeight in 2581, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.046875 = fieldNorm(doc=2581)
        0.01187387 = product of:
          0.03562161 = sum of:
            0.03562161 = weight(_text_:29 in 2581) [ClassicSimilarity], result of:
              0.03562161 = score(doc=2581,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.23319192 = fieldWeight in 2581, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2581)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Date
    15. 8.2004 18:55:29
    Source
    Information processing and management. 38(2002) no.3, S.305-323
  18. Kekäläinen, J.: Binary and graded relevance in IR evaluations : comparison of the effects on ranking of IR systems (2005) 0.03
    0.029524421 = product of:
      0.059048843 = sum of:
        0.04717497 = weight(_text_:processing in 1036) [ClassicSimilarity], result of:
          0.04717497 = score(doc=1036,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.26835677 = fieldWeight in 1036, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.046875 = fieldNorm(doc=1036)
        0.01187387 = product of:
          0.03562161 = sum of:
            0.03562161 = weight(_text_:29 in 1036) [ClassicSimilarity], result of:
              0.03562161 = score(doc=1036,freq=2.0), product of:
                0.15275662 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.043425296 = queryNorm
                0.23319192 = fieldWeight in 1036, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1036)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Date
    26.12.2007 20:29:18
    Source
    Information processing and management. 41(2005) no.5, S.1019-1034
  19. Witschel, H.F.: Global term weights in distributed environments (2008) 0.03
    0.029471014 = product of:
      0.058942027 = sum of:
        0.04717497 = weight(_text_:processing in 2096) [ClassicSimilarity], result of:
          0.04717497 = score(doc=2096,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.26835677 = fieldWeight in 2096, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.046875 = fieldNorm(doc=2096)
        0.011767056 = product of:
          0.035301168 = sum of:
            0.035301168 = weight(_text_:22 in 2096) [ClassicSimilarity], result of:
              0.035301168 = score(doc=2096,freq=2.0), product of:
                0.15206799 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043425296 = queryNorm
                0.23214069 = fieldWeight in 2096, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2096)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Date
    1. 8.2008 9:44:22
    Source
    Information processing and management. 44(2008) no.3, S.1049-1061
  20. Chen, H.; Zhang, Y.; Houston, A.L.: Semantic indexing and searching using a Hopfield net (1998) 0.03
    0.028295456 = product of:
      0.05659091 = sum of:
        0.04717497 = weight(_text_:processing in 5704) [ClassicSimilarity], result of:
          0.04717497 = score(doc=5704,freq=2.0), product of:
            0.175792 = queryWeight, product of:
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.043425296 = queryNorm
            0.26835677 = fieldWeight in 5704, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.048147 = idf(docFreq=2097, maxDocs=44218)
              0.046875 = fieldNorm(doc=5704)
        0.00941594 = product of:
          0.02824782 = sum of:
            0.02824782 = weight(_text_:science in 5704) [ClassicSimilarity], result of:
              0.02824782 = score(doc=5704,freq=4.0), product of:
                0.11438741 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.043425296 = queryNorm
                0.24694869 = fieldWeight in 5704, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5704)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    Presents a neural network approach to document semantic indexing. Reports results of a study to apply a Hopfield net algorithm to simulate human associative memory for concept exploration in the domain of computer science and engineering. The INSPEC database, consisting of 320.000 abstracts from leading periodical articles was used as the document test bed. Benchmark tests conformed that 3 parameters: maximum number of activated nodes; maximum allowable error; and maximum number of iterations; were useful in positively influencing network convergence behaviour without negatively impacting central processing unit performance. Another series of benchmark tests was performed to determine the effectiveness of various filtering techniques in reducing the negative impact of noisy input terms. Preliminary user tests conformed expectations that the Hopfield net is potentially useful as an associative memory technique to improve document recall and precision by solving discrepancies between indexer vocabularies and end user vocabularies
    Source
    Journal of information science. 24(1998) no.1, S.3-18

Years

Languages

  • e 216
  • d 12
  • pt 1
  • More… Less…

Types

  • a 219
  • m 6
  • el 2
  • s 2
  • x 2
  • r 1
  • More… Less…