Search (355 results, page 1 of 18)

  • × theme_ss:"Retrievalalgorithmen"
  1. Soulier, L.; Jabeur, L.B.; Tamine, L.; Bahsoun, W.: On ranking relevant entities in heterogeneous networks using a language-based model (2013) 0.07
    0.06978634 = sum of:
      0.021531772 = product of:
        0.08612709 = sum of:
          0.08612709 = weight(_text_:authors in 664) [ClassicSimilarity], result of:
            0.08612709 = score(doc=664,freq=4.0), product of:
              0.24182312 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.053045183 = queryNorm
              0.35615736 = fieldWeight in 664, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0390625 = fieldNorm(doc=664)
        0.25 = coord(1/4)
      0.04825457 = sum of:
        0.012320121 = weight(_text_:a in 664) [ClassicSimilarity], result of:
          0.012320121 = score(doc=664,freq=20.0), product of:
            0.06116359 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.053045183 = queryNorm
            0.20142901 = fieldWeight in 664, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=664)
        0.035934452 = weight(_text_:22 in 664) [ClassicSimilarity], result of:
          0.035934452 = score(doc=664,freq=2.0), product of:
            0.1857552 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.053045183 = queryNorm
            0.19345059 = fieldWeight in 664, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=664)
    
    Abstract
    A new challenge, accessing multiple relevant entities, arises from the availability of linked heterogeneous data. In this article, we address more specifically the problem of accessing relevant entities, such as publications and authors within a bibliographic network, given an information need. We propose a novel algorithm, called BibRank, that estimates a joint relevance of documents and authors within a bibliographic network. This model ranks each type of entity using a score propagation algorithm with respect to the query topic and the structure of the underlying bi-type information entity network. Evidence sources, namely content-based and network-based scores, are both used to estimate the topical similarity between connected entities. For this purpose, authorship relationships are analyzed through a language model-based score on the one hand and on the other hand, non topically related entities of the same type are detected through marginal citations. The article reports the results of experiments using the Bibrank algorithm for an information retrieval task. The CiteSeerX bibliographic data set forms the basis for the topical query automatic generation and evaluation. We show that a statistically significant improvement over closely related ranking models is achieved.
    Date
    22. 3.2013 19:34:49
    Type
    a
  2. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.06
    0.06372866 = product of:
      0.12745732 = sum of:
        0.12745732 = sum of:
          0.012467085 = weight(_text_:a in 402) [ClassicSimilarity], result of:
            0.012467085 = score(doc=402,freq=2.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.20383182 = fieldWeight in 402, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.125 = fieldNorm(doc=402)
          0.11499024 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
            0.11499024 = score(doc=402,freq=2.0), product of:
              0.1857552 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.053045183 = queryNorm
              0.61904186 = fieldWeight in 402, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.125 = fieldNorm(doc=402)
      0.5 = coord(1/2)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
    Type
    a
  3. Smeaton, A.F.; Rijsbergen, C.J. van: ¬The retrieval effects of query expansion on a feedback document retrieval system (1983) 0.06
    0.058021847 = product of:
      0.116043694 = sum of:
        0.116043694 = sum of:
          0.015427233 = weight(_text_:a in 2134) [ClassicSimilarity], result of:
            0.015427233 = score(doc=2134,freq=4.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.25222903 = fieldWeight in 2134, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.109375 = fieldNorm(doc=2134)
          0.10061646 = weight(_text_:22 in 2134) [ClassicSimilarity], result of:
            0.10061646 = score(doc=2134,freq=2.0), product of:
              0.1857552 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.053045183 = queryNorm
              0.5416616 = fieldWeight in 2134, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.109375 = fieldNorm(doc=2134)
      0.5 = coord(1/2)
    
    Date
    30. 3.2001 13:32:22
    Type
    a
  4. Back, J.: ¬An evaluation of relevancy ranking techniques used by Internet search engines (2000) 0.06
    0.05576258 = product of:
      0.11152516 = sum of:
        0.11152516 = sum of:
          0.0109087005 = weight(_text_:a in 3445) [ClassicSimilarity], result of:
            0.0109087005 = score(doc=3445,freq=2.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.17835285 = fieldWeight in 3445, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.109375 = fieldNorm(doc=3445)
          0.10061646 = weight(_text_:22 in 3445) [ClassicSimilarity], result of:
            0.10061646 = score(doc=3445,freq=2.0), product of:
              0.1857552 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.053045183 = queryNorm
              0.5416616 = fieldWeight in 3445, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.109375 = fieldNorm(doc=3445)
      0.5 = coord(1/2)
    
    Date
    25. 8.2005 17:42:22
    Type
    a
  5. Khoo, C.S.G.; Wan, K.-W.: ¬A simple relevancy-ranking strategy for an interface to Boolean OPACs (2004) 0.05
    0.048603386 = sum of:
      0.010657682 = product of:
        0.04263073 = sum of:
          0.04263073 = weight(_text_:authors in 2509) [ClassicSimilarity], result of:
            0.04263073 = score(doc=2509,freq=2.0), product of:
              0.24182312 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.053045183 = queryNorm
              0.17628889 = fieldWeight in 2509, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.02734375 = fieldNorm(doc=2509)
        0.25 = coord(1/4)
      0.037945703 = sum of:
        0.012791585 = weight(_text_:a in 2509) [ClassicSimilarity], result of:
          0.012791585 = score(doc=2509,freq=44.0), product of:
            0.06116359 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.053045183 = queryNorm
            0.20913726 = fieldWeight in 2509, product of:
              6.6332498 = tf(freq=44.0), with freq of:
                44.0 = termFreq=44.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2509)
        0.025154116 = weight(_text_:22 in 2509) [ClassicSimilarity], result of:
          0.025154116 = score(doc=2509,freq=2.0), product of:
            0.1857552 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.053045183 = queryNorm
            0.1354154 = fieldWeight in 2509, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2509)
    
    Abstract
    A relevancy-ranking algorithm for a natural language interface to Boolean online public access catalogs (OPACs) was formulated and compared with that currently used in a knowledge-based search interface called the E-Referencer, being developed by the authors. The algorithm makes use of seven weIl-known ranking criteria: breadth of match, section weighting, proximity of query words, variant word forms (stemming), document frequency, term frequency and document length. The algorithm converts a natural language query into a series of increasingly broader Boolean search statements. In a small experiment with ten subjects in which the algorithm was simulated by hand, the algorithm obtained good results with a mean overall precision of 0.42 and mean average precision of 0.62, representing a 27 percent improvement in precision and 41 percent improvement in average precision compared to the E-Referencer. The usefulness of each step in the algorithm was analyzed and suggestions are made for improving the algorithm.
    Content
    "Most Web search engines accept natural language queries, perform some kind of fuzzy matching and produce ranked output, displaying first the documents that are most likely to be relevant. On the other hand, most library online public access catalogs (OPACs) an the Web are still Boolean retrieval systems that perform exact matching, and require users to express their search requests precisely in a Boolean search language and to refine their search statements to improve the search results. It is well-documented that users have difficulty searching Boolean OPACs effectively (e.g. Borgman, 1996; Ensor, 1992; Wallace, 1993). One approach to making OPACs easier to use is to develop a natural language search interface that acts as a middleware between the user's Web browser and the OPAC system. The search interface can accept a natural language query from the user and reformulate it as a series of Boolean search statements that are then submitted to the OPAC. The records retrieved by the OPAC are ranked by the search interface before forwarding them to the user's Web browser. The user, then, does not need to interact directly with the Boolean OPAC but with the natural language search interface or search intermediary. The search interface interacts with the OPAC system an the user's behalf. The advantage of this approach is that no modification to the OPAC or library system is required. Furthermore, the search interface can access multiple OPACs, acting as a meta search engine, and integrate search results from various OPACs before sending them to the user. The search interface needs to incorporate a method for converting the user's natural language query into a series of Boolean search statements, and for ranking the OPAC records retrieved. The purpose of this study was to develop a relevancyranking algorithm for a search interface to Boolean OPAC systems. This is part of an on-going effort to develop a knowledge-based search interface to OPACs called the E-Referencer (Khoo et al., 1998, 1999; Poo et al., 2000). E-Referencer v. 2 that has been implemented applies a repertoire of initial search strategies and reformulation strategies to retrieve records from OPACs using the Z39.50 protocol, and also assists users in mapping query keywords to the Library of Congress subject headings."
    Source
    Electronic library. 22(2004) no.2, S.112-120
    Type
    a
  6. Ding, Y.; Yan, E.; Frazho, A.; Caverlee, J.: PageRank for ranking authors in co-citation networks (2009) 0.05
    0.048058774 = sum of:
      0.04475294 = product of:
        0.17901176 = sum of:
          0.17901176 = weight(_text_:authors in 3161) [ClassicSimilarity], result of:
            0.17901176 = score(doc=3161,freq=12.0), product of:
              0.24182312 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.053045183 = queryNorm
              0.7402591 = fieldWeight in 3161, product of:
                3.4641016 = tf(freq=12.0), with freq of:
                  12.0 = termFreq=12.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.046875 = fieldNorm(doc=3161)
        0.25 = coord(1/4)
      0.0033058354 = product of:
        0.006611671 = sum of:
          0.006611671 = weight(_text_:a in 3161) [ClassicSimilarity], result of:
            0.006611671 = score(doc=3161,freq=4.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.10809815 = fieldWeight in 3161, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=3161)
        0.5 = coord(1/2)
    
    Abstract
    This paper studies how varied damping factors in the PageRank algorithm influence the ranking of authors and proposes weighted PageRank algorithms. We selected the 108 most highly cited authors in the information retrieval (IR) area from the 1970s to 2008 to form the author co-citation network. We calculated the ranks of these 108 authors based on PageRank with the damping factor ranging from 0.05 to 0.95. In order to test the relationship between different measures, we compared PageRank and weighted PageRank results with the citation ranking, h-index, and centrality measures. We found that in our author co-citation network, citation rank is highly correlated with PageRank with different damping factors and also with different weighted PageRank algorithms; citation rank and PageRank are not significantly correlated with centrality measures; and h-index rank does not significantly correlate with centrality measures but does significantly correlate with other measures. The key factors that have impact on the PageRank of authors in the author co-citation network are being co-cited with important authors.
    Type
    a
  7. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.05
    0.047796495 = product of:
      0.09559299 = sum of:
        0.09559299 = sum of:
          0.009350315 = weight(_text_:a in 58) [ClassicSimilarity], result of:
            0.009350315 = score(doc=58,freq=2.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.15287387 = fieldWeight in 58, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.09375 = fieldNorm(doc=58)
          0.086242676 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
            0.086242676 = score(doc=58,freq=2.0), product of:
              0.1857552 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.053045183 = queryNorm
              0.46428138 = fieldWeight in 58, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.09375 = fieldNorm(doc=58)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:44
    Type
    a
  8. Fuhr, N.: Rankingexperimente mit gewichteter Indexierung (1986) 0.05
    0.047796495 = product of:
      0.09559299 = sum of:
        0.09559299 = sum of:
          0.009350315 = weight(_text_:a in 2051) [ClassicSimilarity], result of:
            0.009350315 = score(doc=2051,freq=2.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.15287387 = fieldWeight in 2051, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.09375 = fieldNorm(doc=2051)
          0.086242676 = weight(_text_:22 in 2051) [ClassicSimilarity], result of:
            0.086242676 = score(doc=2051,freq=2.0), product of:
              0.1857552 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.053045183 = queryNorm
              0.46428138 = fieldWeight in 2051, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.09375 = fieldNorm(doc=2051)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:56
    Type
    a
  9. Ding, Y.: Topic-based PageRank on author cocitation networks (2011) 0.04
    0.037829764 = sum of:
      0.03164511 = product of:
        0.12658045 = sum of:
          0.12658045 = weight(_text_:authors in 4348) [ClassicSimilarity], result of:
            0.12658045 = score(doc=4348,freq=6.0), product of:
              0.24182312 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.053045183 = queryNorm
              0.52344227 = fieldWeight in 4348, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.046875 = fieldNorm(doc=4348)
        0.25 = coord(1/4)
      0.0061846524 = product of:
        0.012369305 = sum of:
          0.012369305 = weight(_text_:a in 4348) [ClassicSimilarity], result of:
            0.012369305 = score(doc=4348,freq=14.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.20223314 = fieldWeight in 4348, product of:
                3.7416575 = tf(freq=14.0), with freq of:
                  14.0 = termFreq=14.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=4348)
        0.5 = coord(1/2)
    
    Abstract
    Ranking authors is vital for identifying a researcher's impact and standing within a scientific field. There are many different ranking methods (e.g., citations, publications, h-index, PageRank, and weighted PageRank), but most of them are topic-independent. This paper proposes topic-dependent ranks based on the combination of a topic model and a weighted PageRank algorithm. The author-conference-topic (ACT) model was used to extract topic distribution of individual authors. Two ways for combining the ACT model with the PageRank algorithm are proposed: simple combination (I_PR) or using a topic distribution as a weighted vector for PageRank (PR_t). Information retrieval was chosen as the test field and representative authors for different topics at different time phases were identified. Principal component analysis (PCA) was applied to analyze the ranking difference between I_PR and PR_t.
    Type
    a
  10. Losada, D.E.; Barreiro, A.: Emebedding term similarity and inverse document frequency into a logical model of information retrieval (2003) 0.04
    0.035716873 = product of:
      0.071433745 = sum of:
        0.071433745 = sum of:
          0.013938627 = weight(_text_:a in 1422) [ClassicSimilarity], result of:
            0.013938627 = score(doc=1422,freq=10.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.22789092 = fieldWeight in 1422, product of:
                3.1622777 = tf(freq=10.0), with freq of:
                  10.0 = termFreq=10.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0625 = fieldNorm(doc=1422)
          0.05749512 = weight(_text_:22 in 1422) [ClassicSimilarity], result of:
            0.05749512 = score(doc=1422,freq=2.0), product of:
              0.1857552 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.053045183 = queryNorm
              0.30952093 = fieldWeight in 1422, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=1422)
      0.5 = coord(1/2)
    
    Abstract
    We propose a novel approach to incorporate term similarity and inverse document frequency into a logical model of information retrieval. The ability of the logic to handle expressive representations along with the use of such classical notions are promising characteristics for IR systems. The approach proposed here has been efficiently implemented and experiments against test collections are presented.
    Date
    22. 3.2003 19:27:23
    Type
    a
  11. Faloutsos, C.: Signature files (1992) 0.03
    0.0349811 = product of:
      0.0699622 = sum of:
        0.0699622 = sum of:
          0.012467085 = weight(_text_:a in 3499) [ClassicSimilarity], result of:
            0.012467085 = score(doc=3499,freq=8.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.20383182 = fieldWeight in 3499, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0625 = fieldNorm(doc=3499)
          0.05749512 = weight(_text_:22 in 3499) [ClassicSimilarity], result of:
            0.05749512 = score(doc=3499,freq=2.0), product of:
              0.1857552 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.053045183 = queryNorm
              0.30952093 = fieldWeight in 3499, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=3499)
      0.5 = coord(1/2)
    
    Abstract
    Presents a survey and discussion on signature-based text retrieval methods. It describes the main idea behind the signature approach and its advantages over other text retrieval methods, it provides a classification of the signature methods that have appeared in the literature, it describes the main representatives of each class, together with the relative advantages and drawbacks, and it gives a list of applications as well as commercial or university prototypes that use the signature approach
    Date
    7. 5.1999 15:22:48
    Type
    a
  12. Bornmann, L.; Mutz, R.: From P100 to P100' : a new citation-rank approach (2014) 0.03
    0.0349811 = product of:
      0.0699622 = sum of:
        0.0699622 = sum of:
          0.012467085 = weight(_text_:a in 1431) [ClassicSimilarity], result of:
            0.012467085 = score(doc=1431,freq=8.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.20383182 = fieldWeight in 1431, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0625 = fieldNorm(doc=1431)
          0.05749512 = weight(_text_:22 in 1431) [ClassicSimilarity], result of:
            0.05749512 = score(doc=1431,freq=2.0), product of:
              0.1857552 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.053045183 = queryNorm
              0.30952093 = fieldWeight in 1431, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=1431)
      0.5 = coord(1/2)
    
    Abstract
    Properties of a percentile-based rating scale needed in bibliometrics are formulated. Based on these properties, P100 was recently introduced as a new citation-rank approach (Bornmann, Leydesdorff, & Wang, 2013). In this paper, we conceptualize P100 and propose an improvement which we call P100'. Advantages and disadvantages of citation-rank indicators are noted.
    Date
    22. 8.2014 17:05:18
    Type
    a
  13. MacFarlane, A.; Robertson, S.E.; McCann, J.A.: Parallel computing for passage retrieval (2004) 0.03
    0.03315534 = product of:
      0.06631068 = sum of:
        0.06631068 = sum of:
          0.008815561 = weight(_text_:a in 5108) [ClassicSimilarity], result of:
            0.008815561 = score(doc=5108,freq=4.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.14413087 = fieldWeight in 5108, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0625 = fieldNorm(doc=5108)
          0.05749512 = weight(_text_:22 in 5108) [ClassicSimilarity], result of:
            0.05749512 = score(doc=5108,freq=2.0), product of:
              0.1857552 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.053045183 = queryNorm
              0.30952093 = fieldWeight in 5108, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=5108)
      0.5 = coord(1/2)
    
    Date
    20. 1.2007 18:30:22
    Type
    a
  14. Wei, F.; Li, W.; Liu, S.: iRANK: a rank-learn-combine framework for unsupervised ensemble ranking (2010) 0.03
    0.03072675 = sum of:
      0.026370928 = product of:
        0.10548371 = sum of:
          0.10548371 = weight(_text_:authors in 3472) [ClassicSimilarity], result of:
            0.10548371 = score(doc=3472,freq=6.0), product of:
              0.24182312 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.053045183 = queryNorm
              0.43620193 = fieldWeight in 3472, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3472)
        0.25 = coord(1/4)
      0.004355821 = product of:
        0.008711642 = sum of:
          0.008711642 = weight(_text_:a in 3472) [ClassicSimilarity], result of:
            0.008711642 = score(doc=3472,freq=10.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.14243183 = fieldWeight in 3472, product of:
                3.1622777 = tf(freq=10.0), with freq of:
                  10.0 = termFreq=10.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3472)
        0.5 = coord(1/2)
    
    Abstract
    The authors address the problem of unsupervised ensemble ranking. Traditional approaches either combine multiple ranking criteria into a unified representation to obtain an overall ranking score or to utilize certain rank fusion or aggregation techniques to combine the ranking results. Beyond the aforementioned combine-then-rank and rank-then-combine approaches, the authors propose a novel rank-learn-combine ranking framework, called Interactive Ranking (iRANK), which allows two base rankers to teach each other before combination during the ranking process by providing their own ranking results as feedback to the others to boost the ranking performance. This mutual ranking refinement process continues until the two base rankers cannot learn from each other any more. The overall performance is improved by the enhancement of the base rankers through the mutual learning mechanism. The authors further design two ranking refinement strategies to efficiently and effectively use the feedback based on reasonable assumptions and rational analysis. Although iRANK is applicable to many applications, as a case study, they apply this framework to the sentence ranking problem in query-focused summarization and evaluate its effectiveness on the DUC 2005 and 2006 data sets. The results are encouraging with consistent and promising improvements.
    Type
    a
  15. Ojala, M.: Commands that RANKle (1997) 0.03
    0.03059396 = sum of:
      0.024360416 = product of:
        0.097441666 = sum of:
          0.097441666 = weight(_text_:authors in 428) [ClassicSimilarity], result of:
            0.097441666 = score(doc=428,freq=2.0), product of:
              0.24182312 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.053045183 = queryNorm
              0.40294603 = fieldWeight in 428, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0625 = fieldNorm(doc=428)
        0.25 = coord(1/4)
      0.0062335427 = product of:
        0.012467085 = sum of:
          0.012467085 = weight(_text_:a in 428) [ClassicSimilarity], result of:
            0.012467085 = score(doc=428,freq=8.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.20383182 = fieldWeight in 428, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0625 = fieldNorm(doc=428)
        0.5 = coord(1/2)
    
    Abstract
    Examines the RANK command on DIALOG using a statistical analysis of articles in DATABASE as an example. The RANK command was used to find authors, company names, and length of articles. Use of the command revealed a number of complexities and revealed some problematic indexing on the part of the database producers. The LEXIS-NEXIS RANK command was also used, but this fulfils a different function to the command of the same name in DIALOG
    Type
    a
  16. Yan, E.; Ding, Y.; Sugimoto, C.R.: P-Rank: an indicator measuring prestige in heterogeneous scholarly networks (2011) 0.03
    0.030513281 = sum of:
      0.025838124 = product of:
        0.103352495 = sum of:
          0.103352495 = weight(_text_:authors in 4349) [ClassicSimilarity], result of:
            0.103352495 = score(doc=4349,freq=4.0), product of:
              0.24182312 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.053045183 = queryNorm
              0.42738882 = fieldWeight in 4349, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.046875 = fieldNorm(doc=4349)
        0.25 = coord(1/4)
      0.0046751574 = product of:
        0.009350315 = sum of:
          0.009350315 = weight(_text_:a in 4349) [ClassicSimilarity], result of:
            0.009350315 = score(doc=4349,freq=8.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.15287387 = fieldWeight in 4349, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=4349)
        0.5 = coord(1/2)
    
    Abstract
    Ranking scientific productivity and prestige are often limited to homogeneous networks. These networks are unable to account for the multiple factors that constitute the scholarly communication and reward system. This study proposes a new informetric indicator, P-Rank, for measuring prestige in heterogeneous scholarly networks containing articles, authors, and journals. P-Rank differentiates the weight of each citation based on its citing papers, citing journals, and citing authors. Articles from 16 representative library and information science journals are selected as the dataset. Principle Component Analysis is conducted to examine the relationship between P-Rank and other bibliometric indicators. We also compare the correlation and rank variances between citation counts and P-Rank scores. This work provides a new approach to examining prestige in scholarly communication networks in a more comprehensive and nuanced way.
    Type
    a
  17. Biskri, I.; Rompré, L.: Using association rules for query reformulation (2012) 0.03
    0.030513281 = sum of:
      0.025838124 = product of:
        0.103352495 = sum of:
          0.103352495 = weight(_text_:authors in 92) [ClassicSimilarity], result of:
            0.103352495 = score(doc=92,freq=4.0), product of:
              0.24182312 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.053045183 = queryNorm
              0.42738882 = fieldWeight in 92, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.046875 = fieldNorm(doc=92)
        0.25 = coord(1/4)
      0.0046751574 = product of:
        0.009350315 = sum of:
          0.009350315 = weight(_text_:a in 92) [ClassicSimilarity], result of:
            0.009350315 = score(doc=92,freq=8.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.15287387 = fieldWeight in 92, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=92)
        0.5 = coord(1/2)
    
    Abstract
    In this paper the authors will present research on the combination of two methods of data mining: text classification and maximal association rules. Text classification has been the focus of interest of many researchers for a long time. However, the results take the form of lists of words (classes) that people often do not know what to do with. The use of maximal association rules induced a number of advantages: (1) the detection of dependencies and correlations between the relevant units of information (words) of different classes, (2) the extraction of hidden knowledge, often relevant, from a large volume of data. The authors will show how this combination can improve the process of information retrieval.
    Type
    a
  18. Chang, C.-H.; Hsu, C.-C.: Integrating query expansion and conceptual relevance feedback for personalized Web information retrieval (1998) 0.03
    0.029877722 = product of:
      0.059755445 = sum of:
        0.059755445 = sum of:
          0.009447212 = weight(_text_:a in 1319) [ClassicSimilarity], result of:
            0.009447212 = score(doc=1319,freq=6.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.1544581 = fieldWeight in 1319, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0546875 = fieldNorm(doc=1319)
          0.05030823 = weight(_text_:22 in 1319) [ClassicSimilarity], result of:
            0.05030823 = score(doc=1319,freq=2.0), product of:
              0.1857552 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.053045183 = queryNorm
              0.2708308 = fieldWeight in 1319, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=1319)
      0.5 = coord(1/2)
    
    Abstract
    Keyword based querying has been an immediate and efficient way to specify and retrieve related information that the user inquired. However, conventional document ranking based on an automatic assessment of document relevance to the query may not be the best approach when little information is given. Proposes an idea to integrate 2 existing techniques, query expansion and relevance feedback to achieve a concept-based information search for the Web
    Date
    1. 8.1996 22:08:06
    Footnote
    Contribution to a special issue devoted to the Proceedings of the 7th International World Wide Web Conference, held 14-18 April 1998, Brisbane, Australia
    Type
    a
  19. Ravana, S.D.; Rajagopal, P.; Balakrishnan, V.: Ranking retrieval systems using pseudo relevance judgments (2015) 0.03
    0.029765315 = product of:
      0.05953063 = sum of:
        0.05953063 = sum of:
          0.008711642 = weight(_text_:a in 2591) [ClassicSimilarity], result of:
            0.008711642 = score(doc=2591,freq=10.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.14243183 = fieldWeight in 2591, product of:
                3.1622777 = tf(freq=10.0), with freq of:
                  10.0 = termFreq=10.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2591)
          0.050818987 = weight(_text_:22 in 2591) [ClassicSimilarity], result of:
            0.050818987 = score(doc=2591,freq=4.0), product of:
              0.1857552 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.053045183 = queryNorm
              0.27358043 = fieldWeight in 2591, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2591)
      0.5 = coord(1/2)
    
    Abstract
    Purpose In a system-based approach, replicating the web would require large test collections, and judging the relevancy of all documents per topic in creating relevance judgment through human assessors is infeasible. Due to the large amount of documents that requires judgment, there are possible errors introduced by human assessors because of disagreements. The paper aims to discuss these issues. Design/methodology/approach This study explores exponential variation and document ranking methods that generate a reliable set of relevance judgments (pseudo relevance judgments) to reduce human efforts. These methods overcome problems with large amounts of documents for judgment while avoiding human disagreement errors during the judgment process. This study utilizes two key factors: number of occurrences of each document per topic from all the system runs; and document rankings to generate the alternate methods. Findings The effectiveness of the proposed method is evaluated using the correlation coefficient of ranked systems using mean average precision scores between the original Text REtrieval Conference (TREC) relevance judgments and pseudo relevance judgments. The results suggest that the proposed document ranking method with a pool depth of 100 could be a reliable alternative to reduce human effort and disagreement errors involved in generating TREC-like relevance judgments. Originality/value Simple methods proposed in this study show improvement in the correlation coefficient in generating alternate relevance judgment without human assessors while contributing to information retrieval evaluation.
    Date
    20. 1.2015 18:30:22
    18. 9.2018 18:22:56
    Type
    a
  20. Shah, B.; Raghavan, V.; Dhatric, P.; Zhao, X.: ¬A cluster-based approach for efficient content-based image retrieval using a similarity-preserving space transformation method (2006) 0.03
    0.027992498 = sum of:
      0.021531772 = product of:
        0.08612709 = sum of:
          0.08612709 = weight(_text_:authors in 6118) [ClassicSimilarity], result of:
            0.08612709 = score(doc=6118,freq=4.0), product of:
              0.24182312 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.053045183 = queryNorm
              0.35615736 = fieldWeight in 6118, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0390625 = fieldNorm(doc=6118)
        0.25 = coord(1/4)
      0.0064607267 = product of:
        0.012921453 = sum of:
          0.012921453 = weight(_text_:a in 6118) [ClassicSimilarity], result of:
            0.012921453 = score(doc=6118,freq=22.0), product of:
              0.06116359 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.053045183 = queryNorm
              0.21126054 = fieldWeight in 6118, product of:
                4.690416 = tf(freq=22.0), with freq of:
                  22.0 = termFreq=22.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=6118)
        0.5 = coord(1/2)
    
    Abstract
    The techniques of clustering and space transformation have been successfully used in the past to solve a number of pattern recognition problems. In this article, the authors propose a new approach to content-based image retrieval (CBIR) that uses (a) a newly proposed similarity-preserving space transformation method to transform the original low-level image space into a highlevel vector space that enables efficient query processing, and (b) a clustering scheme that further improves the efficiency of our retrieval system. This combination is unique and the resulting system provides synergistic advantages of using both clustering and space transformation. The proposed space transformation method is shown to preserve the order of the distances in the transformed feature space. This strategy makes this approach to retrieval generic as it can be applied to object types, other than images, and feature spaces more general than metric spaces. The CBIR approach uses the inexpensive "estimated" distance in the transformed space, as opposed to the computationally inefficient "real" distance in the original space, to retrieve the desired results for a given query image. The authors also provide a theoretical analysis of the complexity of their CBIR approach when used for color-based retrieval, which shows that it is computationally more efficient than other comparable approaches. An extensive set of experiments to test the efficiency and effectiveness of the proposed approach has been performed. The results show that the approach offers superior response time (improvement of 1-2 orders of magnitude compared to retrieval approaches that either use pruning techniques like indexing, clustering, etc., or space transformation, but not both) with sufficiently high retrieval accuracy.
    Type
    a

Years

Languages

Types

  • a 337
  • el 8
  • m 8
  • s 3
  • p 2
  • r 2
  • More… Less…