Search (66 results, page 1 of 4)

  • × theme_ss:"Retrievalalgorithmen"
  1. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.11
    0.10595073 = product of:
      0.21190146 = sum of:
        0.21190146 = sum of:
          0.12767978 = weight(_text_:n in 58) [ClassicSimilarity], result of:
            0.12767978 = score(doc=58,freq=2.0), product of:
              0.22335295 = queryWeight, product of:
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.05180212 = queryNorm
              0.57165027 = fieldWeight in 58, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.09375 = fieldNorm(doc=58)
          0.08422167 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
            0.08422167 = score(doc=58,freq=2.0), product of:
              0.1814022 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05180212 = queryNorm
              0.46428138 = fieldWeight in 58, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.09375 = fieldNorm(doc=58)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:44
  2. Fuhr, N.: Rankingexperimente mit gewichteter Indexierung (1986) 0.11
    0.10595073 = product of:
      0.21190146 = sum of:
        0.21190146 = sum of:
          0.12767978 = weight(_text_:n in 2051) [ClassicSimilarity], result of:
            0.12767978 = score(doc=2051,freq=2.0), product of:
              0.22335295 = queryWeight, product of:
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.05180212 = queryNorm
              0.57165027 = fieldWeight in 2051, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.09375 = fieldNorm(doc=2051)
          0.08422167 = weight(_text_:22 in 2051) [ClassicSimilarity], result of:
            0.08422167 = score(doc=2051,freq=2.0), product of:
              0.1814022 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05180212 = queryNorm
              0.46428138 = fieldWeight in 2051, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.09375 = fieldNorm(doc=2051)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:56
  3. Chen, Z.; Fu, B.: On the complexity of Rocchio's similarity-based relevance feedback algorithm (2007) 0.09
    0.08618352 = sum of:
      0.021027196 = product of:
        0.084108785 = sum of:
          0.084108785 = weight(_text_:authors in 578) [ClassicSimilarity], result of:
            0.084108785 = score(doc=578,freq=4.0), product of:
              0.23615624 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.05180212 = queryNorm
              0.35615736 = fieldWeight in 578, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0390625 = fieldNorm(doc=578)
        0.25 = coord(1/4)
      0.06515632 = product of:
        0.13031264 = sum of:
          0.13031264 = weight(_text_:n in 578) [ClassicSimilarity], result of:
            0.13031264 = score(doc=578,freq=12.0), product of:
              0.22335295 = queryWeight, product of:
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.05180212 = queryNorm
              0.58343816 = fieldWeight in 578, product of:
                3.4641016 = tf(freq=12.0), with freq of:
                  12.0 = termFreq=12.0
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.0390625 = fieldNorm(doc=578)
        0.5 = coord(1/2)
    
    Abstract
    Rocchio's similarity-based relevance feedback algorithm, one of the most important query reformation methods in information retrieval, is essentially an adaptive learning algorithm from examples in searching for documents represented by a linear classifier. Despite its popularity in various applications, there is little rigorous analysis of its learning complexity in literature. In this article, the authors prove for the first time that the learning complexity of Rocchio's algorithm is O(d + d**2(log d + log n)) over the discretized vector space {0, ... , n - 1 }**d when the inner product similarity measure is used. The upper bound on the learning complexity for searching for documents represented by a monotone linear classifier (q, 0) over {0, ... , n - 1 }d can be improved to, at most, 1 + 2k (n - 1) (log d + log(n - 1)), where k is the number of nonzero components in q. Several lower bounds on the learning complexity are also obtained for Rocchio's algorithm. For example, the authors prove that Rocchio's algorithm has a lower bound Omega((d über 2)log n) on its learning complexity over the Boolean vector space {0,1}**d.
  4. Dannenberg, R.B.; Birmingham, W.P.; Pardo, B.; Hu, N.; Meek, C.; Tzanetakis, G.: ¬A comparative evaluation of search techniques for query-by-humming using the MUSART testbed (2007) 0.06
    0.060940944 = sum of:
      0.014868473 = product of:
        0.05947389 = sum of:
          0.05947389 = weight(_text_:authors in 269) [ClassicSimilarity], result of:
            0.05947389 = score(doc=269,freq=2.0), product of:
              0.23615624 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.05180212 = queryNorm
              0.25184128 = fieldWeight in 269, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0390625 = fieldNorm(doc=269)
        0.25 = coord(1/4)
      0.046072472 = product of:
        0.092144944 = sum of:
          0.092144944 = weight(_text_:n in 269) [ClassicSimilarity], result of:
            0.092144944 = score(doc=269,freq=6.0), product of:
              0.22335295 = queryWeight, product of:
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.05180212 = queryNorm
              0.41255307 = fieldWeight in 269, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.0390625 = fieldNorm(doc=269)
        0.5 = coord(1/2)
    
    Abstract
    Query-by-humming systems offer content-based searching for melodies and require no special musical training or knowledge. Many such systems have been built, but there has not been much useful evaluation and comparison in the literature due to the lack of shared databases and queries. The MUSART project testbed allows various search algorithms to be compared using a shared framework that automatically runs experiments and summarizes results. Using this testbed, the authors compared algorithms based on string alignment, melodic contour matching, a hidden Markov model, n-grams, and CubyHum. Retrieval performance is very sensitive to distance functions and the representation of pitch and rhythm, which raises questions about some previously published conclusions. Some algorithms are particularly sensitive to the quality of queries. Our queries, which are taken from human subjects in a realistic setting, are quite difficult, especially for n-gram models. Finally, simulations on query-by-humming performance as a function of database size indicate that retrieval performance falls only slowly as the database size increases.
  5. Klas, C.-P.; Fuhr, N.; Schaefer, A.: Evaluating strategic support for information access in the DAFFODIL system (2004) 0.05
    0.052975364 = product of:
      0.10595073 = sum of:
        0.10595073 = sum of:
          0.06383989 = weight(_text_:n in 2419) [ClassicSimilarity], result of:
            0.06383989 = score(doc=2419,freq=2.0), product of:
              0.22335295 = queryWeight, product of:
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.05180212 = queryNorm
              0.28582513 = fieldWeight in 2419, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.046875 = fieldNorm(doc=2419)
          0.042110834 = weight(_text_:22 in 2419) [ClassicSimilarity], result of:
            0.042110834 = score(doc=2419,freq=2.0), product of:
              0.1814022 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05180212 = queryNorm
              0.23214069 = fieldWeight in 2419, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2419)
      0.5 = coord(1/2)
    
    Date
    16.11.2008 16:22:48
  6. Soulier, L.; Jabeur, L.B.; Tamine, L.; Bahsoun, W.: On ranking relevant entities in heterogeneous networks using a language-based model (2013) 0.04
    0.038573377 = sum of:
      0.021027196 = product of:
        0.084108785 = sum of:
          0.084108785 = weight(_text_:authors in 664) [ClassicSimilarity], result of:
            0.084108785 = score(doc=664,freq=4.0), product of:
              0.23615624 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.05180212 = queryNorm
              0.35615736 = fieldWeight in 664, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0390625 = fieldNorm(doc=664)
        0.25 = coord(1/4)
      0.01754618 = product of:
        0.03509236 = sum of:
          0.03509236 = weight(_text_:22 in 664) [ClassicSimilarity], result of:
            0.03509236 = score(doc=664,freq=2.0), product of:
              0.1814022 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05180212 = queryNorm
              0.19345059 = fieldWeight in 664, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=664)
        0.5 = coord(1/2)
    
    Abstract
    A new challenge, accessing multiple relevant entities, arises from the availability of linked heterogeneous data. In this article, we address more specifically the problem of accessing relevant entities, such as publications and authors within a bibliographic network, given an information need. We propose a novel algorithm, called BibRank, that estimates a joint relevance of documents and authors within a bibliographic network. This model ranks each type of entity using a score propagation algorithm with respect to the query topic and the structure of the underlying bi-type information entity network. Evidence sources, namely content-based and network-based scores, are both used to estimate the topical similarity between connected entities. For this purpose, authorship relationships are analyzed through a language model-based score on the one hand and on the other hand, non topically related entities of the same type are detected through marginal citations. The article reports the results of experiments using the Bibrank algorithm for an information retrieval task. The CiteSeerX bibliographic data set forms the basis for the topical query automatic generation and evaluation. We show that a statistically significant improvement over closely related ranking models is achieved.
    Date
    22. 3.2013 19:34:49
  7. Cheng, C.-S.; Chung, C.-P.; Shann, J.J.-J.: Fast query evaluation through document identifier assignment for inverted file-based information retrieval systems (2006) 0.03
    0.029739656 = product of:
      0.05947931 = sum of:
        0.05947931 = product of:
          0.11895862 = sum of:
            0.11895862 = weight(_text_:n in 979) [ClassicSimilarity], result of:
              0.11895862 = score(doc=979,freq=10.0), product of:
                0.22335295 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.05180212 = queryNorm
                0.53260374 = fieldWeight in 979, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=979)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Compressing an inverted file can greatly improve query performance of an information retrieval system (IRS) by reducing disk I/Os. We observe that a good document identifier assignment (DIA) can make the document identifiers in the posting lists more clustered, and result in better compression as well as shorter query processing time. In this paper, we tackle the NP-complete problem of finding an optimal DIA to minimize the average query processing time in an IRS when the probability distribution of query terms is given. We indicate that the greedy nearest neighbor (Greedy-NN) algorithm can provide excellent performance for this problem. However, the Greedy-NN algorithm is inappropriate if used in large-scale IRSs, due to its high complexity O(N2 × n), where N denotes the number of documents and n denotes the number of distinct terms. In real-world IRSs, the distribution of query terms is skewed. Based on this fact, we propose a fast O(N × n) heuristic, called partition-based document identifier assignment (PBDIA) algorithm, which can efficiently assign consecutive document identifiers to those documents containing frequently used query terms, and improve compression efficiency of the posting lists for those terms. This can result in reduced query processing time. The experimental results show that the PBDIA algorithm can yield a competitive performance versus the Greedy-NN for the DIA problem, and that this optimization problem has significant advantages for both long queries and parallel information retrieval (IR).
  8. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.03
    0.02807389 = product of:
      0.05614778 = sum of:
        0.05614778 = product of:
          0.11229556 = sum of:
            0.11229556 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.11229556 = score(doc=402,freq=2.0), product of:
                0.1814022 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05180212 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  9. Fuhr, N.: Modelle im Information Retrieval (2013) 0.03
    0.026599953 = product of:
      0.053199906 = sum of:
        0.053199906 = product of:
          0.10639981 = sum of:
            0.10639981 = weight(_text_:n in 724) [ClassicSimilarity], result of:
              0.10639981 = score(doc=724,freq=2.0), product of:
                0.22335295 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.05180212 = queryNorm
                0.47637522 = fieldWeight in 724, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.078125 = fieldNorm(doc=724)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  10. Karlsson, A.; Hammarfelt, B.; Steinhauer, H.J.; Falkman, G.; Olson, N.; Nelhans, G.; Nolin, J.: Modeling uncertainty in bibliometrics and information retrieval : an information fusion approach (2015) 0.03
    0.026599953 = product of:
      0.053199906 = sum of:
        0.053199906 = product of:
          0.10639981 = sum of:
            0.10639981 = weight(_text_:n in 1696) [ClassicSimilarity], result of:
              0.10639981 = score(doc=1696,freq=2.0), product of:
                0.22335295 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.05180212 = queryNorm
                0.47637522 = fieldWeight in 1696, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1696)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  11. Smeaton, A.F.; Rijsbergen, C.J. van: ¬The retrieval effects of query expansion on a feedback document retrieval system (1983) 0.02
    0.024564654 = product of:
      0.049129307 = sum of:
        0.049129307 = product of:
          0.098258615 = sum of:
            0.098258615 = weight(_text_:22 in 2134) [ClassicSimilarity], result of:
              0.098258615 = score(doc=2134,freq=2.0), product of:
                0.1814022 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05180212 = queryNorm
                0.5416616 = fieldWeight in 2134, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2134)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    30. 3.2001 13:32:22
  12. Back, J.: ¬An evaluation of relevancy ranking techniques used by Internet search engines (2000) 0.02
    0.024564654 = product of:
      0.049129307 = sum of:
        0.049129307 = product of:
          0.098258615 = sum of:
            0.098258615 = weight(_text_:22 in 3445) [ClassicSimilarity], result of:
              0.098258615 = score(doc=3445,freq=2.0), product of:
                0.1814022 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05180212 = queryNorm
                0.5416616 = fieldWeight in 3445, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3445)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    25. 8.2005 17:42:22
  13. Dang, E.K.F.; Luk, R.W.P.; Allan, J.: Beyond bag-of-words : bigram-enhanced context-dependent term weights (2014) 0.02
    0.023036236 = product of:
      0.046072472 = sum of:
        0.046072472 = product of:
          0.092144944 = sum of:
            0.092144944 = weight(_text_:n in 1283) [ClassicSimilarity], result of:
              0.092144944 = score(doc=1283,freq=6.0), product of:
                0.22335295 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.05180212 = queryNorm
                0.41255307 = fieldWeight in 1283, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1283)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    While term independence is a widely held assumption in most of the established information retrieval approaches, it is clearly not true and various works in the past have investigated a relaxation of the assumption. One approach is to use n-grams in document representation instead of unigrams. However, the majority of early works on n-grams obtained only modest performance improvement. On the other hand, the use of information based on supporting terms or "contexts" of queries has been found to be promising. In particular, recent studies showed that using new context-dependent term weights improved the performance of relevance feedback (RF) retrieval compared with using traditional bag-of-words BM25 term weights. Calculation of the new term weights requires an estimation of the local probability of relevance of each query term occurrence. In previous studies, the estimation of this probability was based on unigrams that occur in the neighborhood of a query term. We explore an integration of the n-gram and context approaches by computing context-dependent term weights based on a mixture of unigrams and bigrams. Extensive experiments are performed using the title queries of the Text Retrieval Conference (TREC)-6, TREC-7, TREC-8, and TREC-2005 collections, for RF with relevance judgment of either the top 10 or top 20 documents of an initial retrieval. We identify some crucial elements needed in the use of bigrams in our methods, such as proper inverse document frequency (IDF) weighting of the bigrams and noise reduction by pruning bigrams with large document frequency values. We show that enhancing context-dependent term weights with bigrams is effective in further improving retrieval performance.
  14. Khoo, C.S.G.; Wan, K.-W.: ¬A simple relevancy-ranking strategy for an interface to Boolean OPACs (2004) 0.02
    0.022690257 = sum of:
      0.01040793 = product of:
        0.04163172 = sum of:
          0.04163172 = weight(_text_:authors in 2509) [ClassicSimilarity], result of:
            0.04163172 = score(doc=2509,freq=2.0), product of:
              0.23615624 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.05180212 = queryNorm
              0.17628889 = fieldWeight in 2509, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.02734375 = fieldNorm(doc=2509)
        0.25 = coord(1/4)
      0.012282327 = product of:
        0.024564654 = sum of:
          0.024564654 = weight(_text_:22 in 2509) [ClassicSimilarity], result of:
            0.024564654 = score(doc=2509,freq=2.0), product of:
              0.1814022 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05180212 = queryNorm
              0.1354154 = fieldWeight in 2509, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.02734375 = fieldNorm(doc=2509)
        0.5 = coord(1/2)
    
    Abstract
    A relevancy-ranking algorithm for a natural language interface to Boolean online public access catalogs (OPACs) was formulated and compared with that currently used in a knowledge-based search interface called the E-Referencer, being developed by the authors. The algorithm makes use of seven weIl-known ranking criteria: breadth of match, section weighting, proximity of query words, variant word forms (stemming), document frequency, term frequency and document length. The algorithm converts a natural language query into a series of increasingly broader Boolean search statements. In a small experiment with ten subjects in which the algorithm was simulated by hand, the algorithm obtained good results with a mean overall precision of 0.42 and mean average precision of 0.62, representing a 27 percent improvement in precision and 41 percent improvement in average precision compared to the E-Referencer. The usefulness of each step in the algorithm was analyzed and suggestions are made for improving the algorithm.
    Source
    Electronic library. 22(2004) no.2, S.112-120
  15. Ding, Y.; Yan, E.; Frazho, A.; Caverlee, J.: PageRank for ranking authors in co-citation networks (2009) 0.02
    0.0218521 = product of:
      0.0437042 = sum of:
        0.0437042 = product of:
          0.1748168 = sum of:
            0.1748168 = weight(_text_:authors in 3161) [ClassicSimilarity], result of:
              0.1748168 = score(doc=3161,freq=12.0), product of:
                0.23615624 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.05180212 = queryNorm
                0.7402591 = fieldWeight in 3161, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3161)
          0.25 = coord(1/4)
      0.5 = coord(1/2)
    
    Abstract
    This paper studies how varied damping factors in the PageRank algorithm influence the ranking of authors and proposes weighted PageRank algorithms. We selected the 108 most highly cited authors in the information retrieval (IR) area from the 1970s to 2008 to form the author co-citation network. We calculated the ranks of these 108 authors based on PageRank with the damping factor ranging from 0.05 to 0.95. In order to test the relationship between different measures, we compared PageRank and weighted PageRank results with the citation ranking, h-index, and centrality measures. We found that in our author co-citation network, citation rank is highly correlated with PageRank with different damping factors and also with different weighted PageRank algorithms; citation rank and PageRank are not significantly correlated with centrality measures; and h-index rank does not significantly correlate with centrality measures but does significantly correlate with other measures. The key factors that have impact on the PageRank of authors in the author co-citation network are being co-cited with important authors.
  16. Uratani, N.; Takeda, M.: ¬A fast string-searching algorithm for multiple patterns (1993) 0.02
    0.021279963 = product of:
      0.042559925 = sum of:
        0.042559925 = product of:
          0.08511985 = sum of:
            0.08511985 = weight(_text_:n in 6275) [ClassicSimilarity], result of:
              0.08511985 = score(doc=6275,freq=2.0), product of:
                0.22335295 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.05180212 = queryNorm
                0.38110018 = fieldWeight in 6275, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6275)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  17. Fuhr, N.: Zur Überwindung der Diskrepanz zwischen Retrievalforschung und -praxis (1990) 0.02
    0.021279963 = product of:
      0.042559925 = sum of:
        0.042559925 = product of:
          0.08511985 = sum of:
            0.08511985 = weight(_text_:n in 6625) [ClassicSimilarity], result of:
              0.08511985 = score(doc=6625,freq=2.0), product of:
                0.22335295 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.05180212 = queryNorm
                0.38110018 = fieldWeight in 6625, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6625)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  18. Pfeifer, U.; Pennekamp, S.: Incremental processing of vague queries in interactive retrieval systems (1997) 0.02
    0.021279963 = product of:
      0.042559925 = sum of:
        0.042559925 = product of:
          0.08511985 = sum of:
            0.08511985 = weight(_text_:n in 735) [ClassicSimilarity], result of:
              0.08511985 = score(doc=735,freq=2.0), product of:
                0.22335295 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.05180212 = queryNorm
                0.38110018 = fieldWeight in 735, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0625 = fieldNorm(doc=735)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Hypertext - Information Retrieval - Multimedia '97: Theorien, Modelle und Implementierungen integrierter elektronischer Informationssysteme. Proceedings HIM '97. Hrsg.: N. Fuhr u.a
  19. Frakes, W.B.: Stemming algorithms (1992) 0.02
    0.021279963 = product of:
      0.042559925 = sum of:
        0.042559925 = product of:
          0.08511985 = sum of:
            0.08511985 = weight(_text_:n in 3503) [ClassicSimilarity], result of:
              0.08511985 = score(doc=3503,freq=2.0), product of:
                0.22335295 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.05180212 = queryNorm
                0.38110018 = fieldWeight in 3503, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3503)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Desribes stemming algorithms - programs that relate morphologically similar indexing and search terms. Stemming is used to improve retrieval effectiveness and to reduce the size of indexing files. Several approaches to stemming are describes - table lookup, affix removal, successor variety, and n-gram. empirical studies of stemming are summarized. The Porter stemmer is described in detail, and a full implementation in C is presented
  20. Wartik, S.; Fox, E.; Heath, L.; Chen, Q.-F.: Hashing algorithms (1992) 0.02
    0.021279963 = product of:
      0.042559925 = sum of:
        0.042559925 = product of:
          0.08511985 = sum of:
            0.08511985 = weight(_text_:n in 3510) [ClassicSimilarity], result of:
              0.08511985 = score(doc=3510,freq=2.0), product of:
                0.22335295 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.05180212 = queryNorm
                0.38110018 = fieldWeight in 3510, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3510)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Discusses hashing, an information storage and retrieval technique useful for implementing many of the other structures in this book. The concepts underlying hashing are presented, along with 2 implementation strategies. The chapter also contains an extensive discussion of perfect hashing, an important optimization in information retrieval, and an O(n) algorithm to find minimal perfect hash functions for a set of keys

Years

Languages

  • e 58
  • d 8

Types

  • a 59
  • m 5
  • s 2
  • el 1
  • r 1
  • More… Less…