Search (60 results, page 1 of 3)

  • × theme_ss:"Retrievalalgorithmen"
  1. Wei, F.; Li, W.; Liu, S.: iRANK: a rank-learn-combine framework for unsupervised ensemble ranking (2010) 0.05
    0.04866091 = sum of:
      0.025846891 = product of:
        0.103387564 = sum of:
          0.103387564 = weight(_text_:authors in 3472) [ClassicSimilarity], result of:
            0.103387564 = score(doc=3472,freq=6.0), product of:
              0.23701768 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.051991083 = queryNorm
              0.43620193 = fieldWeight in 3472, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3472)
        0.25 = coord(1/4)
      0.02281402 = product of:
        0.04562804 = sum of:
          0.04562804 = weight(_text_:f in 3472) [ClassicSimilarity], result of:
            0.04562804 = score(doc=3472,freq=2.0), product of:
              0.20722532 = queryWeight, product of:
                3.985786 = idf(docFreq=2232, maxDocs=44218)
                0.051991083 = queryNorm
              0.22018565 = fieldWeight in 3472, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.985786 = idf(docFreq=2232, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3472)
        0.5 = coord(1/2)
    
    Abstract
    The authors address the problem of unsupervised ensemble ranking. Traditional approaches either combine multiple ranking criteria into a unified representation to obtain an overall ranking score or to utilize certain rank fusion or aggregation techniques to combine the ranking results. Beyond the aforementioned combine-then-rank and rank-then-combine approaches, the authors propose a novel rank-learn-combine ranking framework, called Interactive Ranking (iRANK), which allows two base rankers to teach each other before combination during the ranking process by providing their own ranking results as feedback to the others to boost the ranking performance. This mutual ranking refinement process continues until the two base rankers cannot learn from each other any more. The overall performance is improved by the enhancement of the base rankers through the mutual learning mechanism. The authors further design two ranking refinement strategies to efficiently and effectively use the feedback based on reasonable assumptions and rational analysis. Although iRANK is applicable to many applications, as a case study, they apply this framework to the sentence ranking problem in query-focused summarization and evaluate its effectiveness on the DUC 2005 and 2006 data sets. The results are encouraging with consistent and promising improvements.
  2. Kelledy, F.; Smeaton, A.F.: Signature files and beyond (1996) 0.05
    0.048509046 = product of:
      0.09701809 = sum of:
        0.09701809 = sum of:
          0.05475365 = weight(_text_:f in 6973) [ClassicSimilarity], result of:
            0.05475365 = score(doc=6973,freq=2.0), product of:
              0.20722532 = queryWeight, product of:
                3.985786 = idf(docFreq=2232, maxDocs=44218)
                0.051991083 = queryNorm
              0.26422277 = fieldWeight in 6973, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.985786 = idf(docFreq=2232, maxDocs=44218)
                0.046875 = fieldNorm(doc=6973)
          0.042264443 = weight(_text_:22 in 6973) [ClassicSimilarity], result of:
            0.042264443 = score(doc=6973,freq=2.0), product of:
              0.1820639 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051991083 = queryNorm
              0.23214069 = fieldWeight in 6973, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=6973)
      0.5 = coord(1/2)
    
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
  3. Crestani, F.; Dominich, S.; Lalmas, M.; Rijsbergen, C.J.K. van: Mathematical, logical, and formal methods in information retrieval : an introduction to the special issue (2003) 0.05
    0.048509046 = product of:
      0.09701809 = sum of:
        0.09701809 = sum of:
          0.05475365 = weight(_text_:f in 1451) [ClassicSimilarity], result of:
            0.05475365 = score(doc=1451,freq=2.0), product of:
              0.20722532 = queryWeight, product of:
                3.985786 = idf(docFreq=2232, maxDocs=44218)
                0.051991083 = queryNorm
              0.26422277 = fieldWeight in 1451, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.985786 = idf(docFreq=2232, maxDocs=44218)
                0.046875 = fieldNorm(doc=1451)
          0.042264443 = weight(_text_:22 in 1451) [ClassicSimilarity], result of:
            0.042264443 = score(doc=1451,freq=2.0), product of:
              0.1820639 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051991083 = queryNorm
              0.23214069 = fieldWeight in 1451, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=1451)
      0.5 = coord(1/2)
    
    Date
    22. 3.2003 19:27:36
  4. Soulier, L.; Jabeur, L.B.; Tamine, L.; Bahsoun, W.: On ranking relevant entities in heterogeneous networks using a language-based model (2013) 0.04
    0.03871408 = sum of:
      0.021103898 = product of:
        0.08441559 = sum of:
          0.08441559 = weight(_text_:authors in 664) [ClassicSimilarity], result of:
            0.08441559 = score(doc=664,freq=4.0), product of:
              0.23701768 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.051991083 = queryNorm
              0.35615736 = fieldWeight in 664, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0390625 = fieldNorm(doc=664)
        0.25 = coord(1/4)
      0.017610185 = product of:
        0.03522037 = sum of:
          0.03522037 = weight(_text_:22 in 664) [ClassicSimilarity], result of:
            0.03522037 = score(doc=664,freq=2.0), product of:
              0.1820639 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051991083 = queryNorm
              0.19345059 = fieldWeight in 664, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=664)
        0.5 = coord(1/2)
    
    Abstract
    A new challenge, accessing multiple relevant entities, arises from the availability of linked heterogeneous data. In this article, we address more specifically the problem of accessing relevant entities, such as publications and authors within a bibliographic network, given an information need. We propose a novel algorithm, called BibRank, that estimates a joint relevance of documents and authors within a bibliographic network. This model ranks each type of entity using a score propagation algorithm with respect to the query topic and the structure of the underlying bi-type information entity network. Evidence sources, namely content-based and network-based scores, are both used to estimate the topical similarity between connected entities. For this purpose, authorship relationships are analyzed through a language model-based score on the one hand and on the other hand, non topically related entities of the same type are detected through marginal citations. The article reports the results of experiments using the Bibrank algorithm for an information retrieval task. The CiteSeerX bibliographic data set forms the basis for the topical query automatic generation and evaluation. We show that a statistically significant improvement over closely related ranking models is achieved.
    Date
    22. 3.2013 19:34:49
  5. Crestani, F.: Combination of similarity measures for effective spoken document retrieval (2003) 0.03
    0.03193963 = product of:
      0.06387926 = sum of:
        0.06387926 = product of:
          0.12775852 = sum of:
            0.12775852 = weight(_text_:f in 4690) [ClassicSimilarity], result of:
              0.12775852 = score(doc=4690,freq=2.0), product of:
                0.20722532 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.051991083 = queryNorm
                0.6165198 = fieldWeight in 4690, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4690)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  6. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.03
    0.028176295 = product of:
      0.05635259 = sum of:
        0.05635259 = product of:
          0.11270518 = sum of:
            0.11270518 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.11270518 = score(doc=402,freq=2.0), product of:
                0.1820639 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051991083 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  7. Jacso, P.: Testing the calculation of a realistic h-index in Google Scholar, Scopus, and Web of Science for F. W. Lancaster (2008) 0.03
    0.025506852 = product of:
      0.051013704 = sum of:
        0.051013704 = product of:
          0.10202741 = sum of:
            0.10202741 = weight(_text_:f in 5586) [ClassicSimilarity], result of:
              0.10202741 = score(doc=5586,freq=10.0), product of:
                0.20722532 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.051991083 = queryNorm
                0.4923501 = fieldWeight in 5586, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5586)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper focuses on the practical limitations in the content and software of the databases that are used to calculate the h-index for assessing the publishing productivity and impact of researchers. To celebrate F. W. Lancaster's biological age of seventy-five, and "scientific age" of forty-five, this paper discusses the related features of Google Scholar, Scopus, and Web of Science (WoS), and demonstrates in the latter how a much more realistic and fair h-index can be computed for F. W. Lancaster than the one produced automatically. Browsing and searching the cited reference index of the 1945-2007 edition of WoS, which in my estimate has over a hundred million "orphan references" that have no counterpart master records to be attached to, and "stray references" that cite papers which do have master records but cannot be identified by the matching algorithm because of errors of omission and commission in the references of the citing works, can bring up hundreds of additional cited references given to works of an accomplished author but are ignored in the automatic process of calculating the h-index. The partially manual process doubled the h-index value for F. W. Lancaster from 13 to 26, which is a much more realistic value for an information scientist and professor of his stature.
    Content
    Beitrag in einem Themenheft 'The Influence of F. W. Lancaster on Information Science and on Libraries', das als Festschrift für F.W. Lancaster deklariert ist.
  8. Smeaton, A.F.; Rijsbergen, C.J. van: ¬The retrieval effects of query expansion on a feedback document retrieval system (1983) 0.02
    0.024654258 = product of:
      0.049308516 = sum of:
        0.049308516 = product of:
          0.09861703 = sum of:
            0.09861703 = weight(_text_:22 in 2134) [ClassicSimilarity], result of:
              0.09861703 = score(doc=2134,freq=2.0), product of:
                0.1820639 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051991083 = queryNorm
                0.5416616 = fieldWeight in 2134, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2134)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    30. 3.2001 13:32:22
  9. Back, J.: ¬An evaluation of relevancy ranking techniques used by Internet search engines (2000) 0.02
    0.024654258 = product of:
      0.049308516 = sum of:
        0.049308516 = product of:
          0.09861703 = sum of:
            0.09861703 = weight(_text_:22 in 3445) [ClassicSimilarity], result of:
              0.09861703 = score(doc=3445,freq=2.0), product of:
                0.1820639 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051991083 = queryNorm
                0.5416616 = fieldWeight in 3445, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3445)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    25. 8.2005 17:42:22
  10. Khoo, C.S.G.; Wan, K.-W.: ¬A simple relevancy-ranking strategy for an interface to Boolean OPACs (2004) 0.02
    0.022773024 = sum of:
      0.010445896 = product of:
        0.041783582 = sum of:
          0.041783582 = weight(_text_:authors in 2509) [ClassicSimilarity], result of:
            0.041783582 = score(doc=2509,freq=2.0), product of:
              0.23701768 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.051991083 = queryNorm
              0.17628889 = fieldWeight in 2509, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.02734375 = fieldNorm(doc=2509)
        0.25 = coord(1/4)
      0.012327129 = product of:
        0.024654258 = sum of:
          0.024654258 = weight(_text_:22 in 2509) [ClassicSimilarity], result of:
            0.024654258 = score(doc=2509,freq=2.0), product of:
              0.1820639 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051991083 = queryNorm
              0.1354154 = fieldWeight in 2509, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.02734375 = fieldNorm(doc=2509)
        0.5 = coord(1/2)
    
    Abstract
    A relevancy-ranking algorithm for a natural language interface to Boolean online public access catalogs (OPACs) was formulated and compared with that currently used in a knowledge-based search interface called the E-Referencer, being developed by the authors. The algorithm makes use of seven weIl-known ranking criteria: breadth of match, section weighting, proximity of query words, variant word forms (stemming), document frequency, term frequency and document length. The algorithm converts a natural language query into a series of increasingly broader Boolean search statements. In a small experiment with ten subjects in which the algorithm was simulated by hand, the algorithm obtained good results with a mean overall precision of 0.42 and mean average precision of 0.62, representing a 27 percent improvement in precision and 41 percent improvement in average precision compared to the E-Referencer. The usefulness of each step in the algorithm was analyzed and suggestions are made for improving the algorithm.
    Source
    Electronic library. 22(2004) no.2, S.112-120
  11. Ding, Y.; Yan, E.; Frazho, A.; Caverlee, J.: PageRank for ranking authors in co-citation networks (2009) 0.02
    0.021931812 = product of:
      0.043863624 = sum of:
        0.043863624 = product of:
          0.1754545 = sum of:
            0.1754545 = weight(_text_:authors in 3161) [ClassicSimilarity], result of:
              0.1754545 = score(doc=3161,freq=12.0), product of:
                0.23701768 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.051991083 = queryNorm
                0.7402591 = fieldWeight in 3161, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3161)
          0.25 = coord(1/4)
      0.5 = coord(1/2)
    
    Abstract
    This paper studies how varied damping factors in the PageRank algorithm influence the ranking of authors and proposes weighted PageRank algorithms. We selected the 108 most highly cited authors in the information retrieval (IR) area from the 1970s to 2008 to form the author co-citation network. We calculated the ranks of these 108 authors based on PageRank with the damping factor ranging from 0.05 to 0.95. In order to test the relationship between different measures, we compared PageRank and weighted PageRank results with the citation ranking, h-index, and centrality measures. We found that in our author co-citation network, citation rank is highly correlated with PageRank with different damping factors and also with different weighted PageRank algorithms; citation rank and PageRank are not significantly correlated with centrality measures; and h-index rank does not significantly correlate with centrality measures but does significantly correlate with other measures. The key factors that have impact on the PageRank of authors in the author co-citation network are being co-cited with important authors.
  12. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.02
    0.021132221 = product of:
      0.042264443 = sum of:
        0.042264443 = product of:
          0.084528886 = sum of:
            0.084528886 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
              0.084528886 = score(doc=58,freq=2.0), product of:
                0.1820639 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051991083 = queryNorm
                0.46428138 = fieldWeight in 58, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=58)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:44
  13. Fuhr, N.: Rankingexperimente mit gewichteter Indexierung (1986) 0.02
    0.021132221 = product of:
      0.042264443 = sum of:
        0.042264443 = product of:
          0.084528886 = sum of:
            0.084528886 = weight(_text_:22 in 2051) [ClassicSimilarity], result of:
              0.084528886 = score(doc=2051,freq=2.0), product of:
                0.1820639 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051991083 = queryNorm
                0.46428138 = fieldWeight in 2051, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2051)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:56
  14. Can, F.: Incremental clustering for dynamic information processing (1993) 0.02
    0.018251216 = product of:
      0.036502432 = sum of:
        0.036502432 = product of:
          0.073004864 = sum of:
            0.073004864 = weight(_text_:f in 6627) [ClassicSimilarity], result of:
              0.073004864 = score(doc=6627,freq=2.0), product of:
                0.20722532 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.051991083 = queryNorm
                0.35229704 = fieldWeight in 6627, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6627)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  15. Wartik, S.; Fox, E.; Heath, L.; Chen, Q.-F.: Hashing algorithms (1992) 0.02
    0.018251216 = product of:
      0.036502432 = sum of:
        0.036502432 = product of:
          0.073004864 = sum of:
            0.073004864 = weight(_text_:f in 3510) [ClassicSimilarity], result of:
              0.073004864 = score(doc=3510,freq=2.0), product of:
                0.20722532 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.051991083 = queryNorm
                0.35229704 = fieldWeight in 3510, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3510)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  16. Guerrero-Bote, V.P.; Moya Anegón, F. de; Herrero Solana, V.: Document organization using Kohonen's algorithm (2002) 0.02
    0.018251216 = product of:
      0.036502432 = sum of:
        0.036502432 = product of:
          0.073004864 = sum of:
            0.073004864 = weight(_text_:f in 2564) [ClassicSimilarity], result of:
              0.073004864 = score(doc=2564,freq=2.0), product of:
                0.20722532 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.051991083 = queryNorm
                0.35229704 = fieldWeight in 2564, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2564)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  17. López-Pujalte, C.; Guerrero-Bote, V.P.; Moya-Anegón, F. de: Genetic algorithms in relevance feedback : a second test and new contributions (2003) 0.02
    0.015969815 = product of:
      0.03193963 = sum of:
        0.03193963 = product of:
          0.06387926 = sum of:
            0.06387926 = weight(_text_:f in 1076) [ClassicSimilarity], result of:
              0.06387926 = score(doc=1076,freq=2.0), product of:
                0.20722532 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.051991083 = queryNorm
                0.3082599 = fieldWeight in 1076, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1076)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  18. López-Herrera, A.G.; Herrera-Viedma, E.; Herrera, F.: ¬A study of the use of multi-objective evolutionary algorithms to learn Boolean queries : a comparative study (2009) 0.02
    0.015969815 = product of:
      0.03193963 = sum of:
        0.03193963 = product of:
          0.06387926 = sum of:
            0.06387926 = weight(_text_:f in 1751) [ClassicSimilarity], result of:
              0.06387926 = score(doc=1751,freq=2.0), product of:
                0.20722532 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.051991083 = queryNorm
                0.3082599 = fieldWeight in 1751, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1751)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  19. Ding, Y.: Topic-based PageRank on author cocitation networks (2011) 0.02
    0.015508134 = product of:
      0.031016268 = sum of:
        0.031016268 = product of:
          0.12406507 = sum of:
            0.12406507 = weight(_text_:authors in 4348) [ClassicSimilarity], result of:
              0.12406507 = score(doc=4348,freq=6.0), product of:
                0.23701768 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.051991083 = queryNorm
                0.52344227 = fieldWeight in 4348, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4348)
          0.25 = coord(1/4)
      0.5 = coord(1/2)
    
    Abstract
    Ranking authors is vital for identifying a researcher's impact and standing within a scientific field. There are many different ranking methods (e.g., citations, publications, h-index, PageRank, and weighted PageRank), but most of them are topic-independent. This paper proposes topic-dependent ranks based on the combination of a topic model and a weighted PageRank algorithm. The author-conference-topic (ACT) model was used to extract topic distribution of individual authors. Two ways for combining the ACT model with the PageRank algorithm are proposed: simple combination (I_PR) or using a topic distribution as a weighted vector for PageRank (PR_t). Information retrieval was chosen as the test field and representative authors for different topics at different time phases were identified. Principal component analysis (PCA) was applied to analyze the ranking difference between I_PR and PR_t.
  20. MacFarlane, A.; Robertson, S.E.; McCann, J.A.: Parallel computing for passage retrieval (2004) 0.01
    0.014088147 = product of:
      0.028176295 = sum of:
        0.028176295 = product of:
          0.05635259 = sum of:
            0.05635259 = weight(_text_:22 in 5108) [ClassicSimilarity], result of:
              0.05635259 = score(doc=5108,freq=2.0), product of:
                0.1820639 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051991083 = queryNorm
                0.30952093 = fieldWeight in 5108, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5108)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    20. 1.2007 18:30:22

Years

Languages

  • e 55
  • d 5

Types

  • a 52
  • m 5
  • s 2
  • r 1
  • x 1
  • More… Less…