Search (61 results, page 1 of 4)

  • × theme_ss:"Retrievalalgorithmen"
  1. Biskri, I.; Rompré, L.: Using association rules for query reformulation (2012) 0.05
    0.049952157 = sum of:
      0.025381705 = product of:
        0.10152682 = sum of:
          0.10152682 = weight(_text_:authors in 92) [ClassicSimilarity], result of:
            0.10152682 = score(doc=92,freq=4.0), product of:
              0.23755142 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.05210816 = queryNorm
              0.42738882 = fieldWeight in 92, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.046875 = fieldNorm(doc=92)
        0.25 = coord(1/4)
      0.024570452 = product of:
        0.049140904 = sum of:
          0.049140904 = weight(_text_:i in 92) [ClassicSimilarity], result of:
            0.049140904 = score(doc=92,freq=2.0), product of:
              0.1965379 = queryWeight, product of:
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.05210816 = queryNorm
              0.25003272 = fieldWeight in 92, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.046875 = fieldNorm(doc=92)
        0.5 = coord(1/2)
    
    Abstract
    In this paper the authors will present research on the combination of two methods of data mining: text classification and maximal association rules. Text classification has been the focus of interest of many researchers for a long time. However, the results take the form of lists of words (classes) that people often do not know what to do with. The use of maximal association rules induced a number of advantages: (1) the detection of dependencies and correlations between the relevant units of information (words) of different classes, (2) the extraction of hidden knowledge, often relevant, from a large volume of data. The authors will show how this combination can improve the process of information retrieval.
  2. Habernal, I.; Konopík, M.; Rohlík, O.: Question answering (2012) 0.04
    0.042518027 = sum of:
      0.017947575 = product of:
        0.0717903 = sum of:
          0.0717903 = weight(_text_:authors in 101) [ClassicSimilarity], result of:
            0.0717903 = score(doc=101,freq=2.0), product of:
              0.23755142 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.05210816 = queryNorm
              0.30220953 = fieldWeight in 101, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.046875 = fieldNorm(doc=101)
        0.25 = coord(1/4)
      0.024570452 = product of:
        0.049140904 = sum of:
          0.049140904 = weight(_text_:i in 101) [ClassicSimilarity], result of:
            0.049140904 = score(doc=101,freq=2.0), product of:
              0.1965379 = queryWeight, product of:
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.05210816 = queryNorm
              0.25003272 = fieldWeight in 101, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.046875 = fieldNorm(doc=101)
        0.5 = coord(1/2)
    
    Abstract
    Question Answering is an area of information retrieval with the added challenge of applying sophisticated techniques to identify the complex syntactic and semantic relationships present in text in order to provide a more sophisticated and satisfactory response to the user's information needs. For this reason, the authors see question answering as the next step beyond standard information retrieval. In this chapter state of the art question answering is covered focusing on providing an overview of systems, techniques and approaches that are likely to be employed in the next generations of search engines. Special attention is paid to question answering using the World Wide Web as the data source and to question answering exploiting the possibilities of Semantic Web. Considerations about the current issues and prospects for promising future research are also provided.
  3. Soulier, L.; Jabeur, L.B.; Tamine, L.; Bahsoun, W.: On ranking relevant entities in heterogeneous networks using a language-based model (2013) 0.04
    0.03880126 = sum of:
      0.021151422 = product of:
        0.08460569 = sum of:
          0.08460569 = weight(_text_:authors in 664) [ClassicSimilarity], result of:
            0.08460569 = score(doc=664,freq=4.0), product of:
              0.23755142 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.05210816 = queryNorm
              0.35615736 = fieldWeight in 664, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0390625 = fieldNorm(doc=664)
        0.25 = coord(1/4)
      0.01764984 = product of:
        0.03529968 = sum of:
          0.03529968 = weight(_text_:22 in 664) [ClassicSimilarity], result of:
            0.03529968 = score(doc=664,freq=2.0), product of:
              0.1824739 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05210816 = queryNorm
              0.19345059 = fieldWeight in 664, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=664)
        0.5 = coord(1/2)
    
    Abstract
    A new challenge, accessing multiple relevant entities, arises from the availability of linked heterogeneous data. In this article, we address more specifically the problem of accessing relevant entities, such as publications and authors within a bibliographic network, given an information need. We propose a novel algorithm, called BibRank, that estimates a joint relevance of documents and authors within a bibliographic network. This model ranks each type of entity using a score propagation algorithm with respect to the query topic and the structure of the underlying bi-type information entity network. Evidence sources, namely content-based and network-based scores, are both used to estimate the topical similarity between connected entities. For this purpose, authorship relationships are analyzed through a language model-based score on the one hand and on the other hand, non topically related entities of the same type are detected through marginal citations. The article reports the results of experiments using the Bibrank algorithm for an information retrieval task. The CiteSeerX bibliographic data set forms the basis for the topical query automatic generation and evaluation. We show that a statistically significant improvement over closely related ranking models is achieved.
    Date
    22. 3.2013 19:34:49
  4. Henzinger, M.R.: Hyperlink analysis for the Web (2001) 0.03
    0.033301435 = sum of:
      0.016921137 = product of:
        0.067684546 = sum of:
          0.067684546 = weight(_text_:authors in 8) [ClassicSimilarity], result of:
            0.067684546 = score(doc=8,freq=4.0), product of:
              0.23755142 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.05210816 = queryNorm
              0.28492588 = fieldWeight in 8, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.03125 = fieldNorm(doc=8)
        0.25 = coord(1/4)
      0.0163803 = product of:
        0.0327606 = sum of:
          0.0327606 = weight(_text_:i in 8) [ClassicSimilarity], result of:
            0.0327606 = score(doc=8,freq=2.0), product of:
              0.1965379 = queryWeight, product of:
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.05210816 = queryNorm
              0.16668847 = fieldWeight in 8, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.03125 = fieldNorm(doc=8)
        0.5 = coord(1/2)
    
    Content
    Information retrieval is a computer science subfield whose goal is to find all documents relevant to a user query in a given collection of documents. As such, information retrieval should really be called document retrieval. Before the advent of the Web, IR systems were typically installed in libraries for use mostly by reference librarians. The retrieval algorithm for these systems was usually based exclusively on analysis of the words in the document. The Web changed all this. Now each Web user has access to various search engines whose retrieval algorithms often use not only the words in the documents but also information like the hyperlink structure of the Web or markup language tags. How are hyperlinks useful? The hyperlink functionality alone-that is, the hyperlink to Web page B that is contained in Web page A-is not directly useful in information retrieval. However, the way Web page authors use hyperlinks can give them valuable information content. Authors usually create hyperlinks they think will be useful to readers. Some may be navigational aids that, for example, take the reader back to the site's home page; others provide access to documents that augment the content of the current page. The latter tend to point to highquality pages that might be on the same topic as the page containing the hyperlink. Web information retrieval systems can exploit this information to refine searches for relevant documents. Hyperlink analysis significantly improves the relevance of the search results, so much so that all major Web search engines claim to use some type of hyperlink analysis. However, the search engines do not disclose details about the type of hyperlink analysis they perform- mostly to avoid manipulation of search results by Web-positioning companies. In this article, I discuss how hyperlink analysis can be applied to ranking algorithms, and survey other ways Web search engines can use this analysis.
  5. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.03
    0.028239746 = product of:
      0.05647949 = sum of:
        0.05647949 = product of:
          0.11295898 = sum of:
            0.11295898 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.11295898 = score(doc=402,freq=2.0), product of:
                0.1824739 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05210816 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  6. Langville, A.N.; Meyer, C.D.: Google's PageRank and beyond : the science of search engine rankings (2006) 0.03
    0.026347723 = sum of:
      0.008973788 = product of:
        0.03589515 = sum of:
          0.03589515 = weight(_text_:authors in 6) [ClassicSimilarity], result of:
            0.03589515 = score(doc=6,freq=2.0), product of:
              0.23755142 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.05210816 = queryNorm
              0.15110476 = fieldWeight in 6, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.0234375 = fieldNorm(doc=6)
        0.25 = coord(1/4)
      0.017373934 = product of:
        0.03474787 = sum of:
          0.03474787 = weight(_text_:i in 6) [ClassicSimilarity], result of:
            0.03474787 = score(doc=6,freq=4.0), product of:
              0.1965379 = queryWeight, product of:
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.05210816 = queryNorm
              0.17679983 = fieldWeight in 6, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.7717297 = idf(docFreq=2765, maxDocs=44218)
                0.0234375 = fieldNorm(doc=6)
        0.5 = coord(1/2)
    
    Abstract
    Why doesn't your home page appear on the first page of search results, even when you query your own name? How do other Web pages always appear at the top? What creates these powerful rankings? And how? The first book ever about the science of Web page rankings, "Google's PageRank and Beyond" supplies the answers to these and other questions and more. The book serves two very different audiences: the curious science reader and the technical computational reader. The chapters build in mathematical sophistication, so that the first five are accessible to the general academic reader. While other chapters are much more mathematical in nature, each one contains something for both audiences. For example, the authors include entertaining asides such as how search engines make money and how the Great Firewall of China influences research. The book includes an extensive background chapter designed to help readers learn more about the mathematics of search engines, and it contains several MATLAB codes and links to sample Web data sets. The philosophy throughout is to encourage readers to experiment with the ideas and algorithms in the text. Any business seriously interested in improving its rankings in the major search engines can benefit from the clear examples, sample code, and list of resources provided. It includes: many illustrative examples and entertaining asides; MATLAB code; accessible and informal style; and complete and self-contained section for mathematics review.
    Content
    Inhalt: Chapter 1. Introduction to Web Search Engines: 1.1 A Short History of Information Retrieval - 1.2 An Overview of Traditional Information Retrieval - 1.3 Web Information Retrieval Chapter 2. Crawling, Indexing, and Query Processing: 2.1 Crawling - 2.2 The Content Index - 2.3 Query Processing Chapter 3. Ranking Webpages by Popularity: 3.1 The Scene in 1998 - 3.2 Two Theses - 3.3 Query-Independence Chapter 4. The Mathematics of Google's PageRank: 4.1 The Original Summation Formula for PageRank - 4.2 Matrix Representation of the Summation Equations - 4.3 Problems with the Iterative Process - 4.4 A Little Markov Chain Theory - 4.5 Early Adjustments to the Basic Model - 4.6 Computation of the PageRank Vector - 4.7 Theorem and Proof for Spectrum of the Google Matrix Chapter 5. Parameters in the PageRank Model: 5.1 The a Factor - 5.2 The Hyperlink Matrix H - 5.3 The Teleportation Matrix E Chapter 6. The Sensitivity of PageRank; 6.1 Sensitivity with respect to alpha - 6.2 Sensitivity with respect to H - 6.3 Sensitivity with respect to vT - 6.4 Other Analyses of Sensitivity - 6.5 Sensitivity Theorems and Proofs Chapter 7. The PageRank Problem as a Linear System: 7.1 Properties of (I - alphaS) - 7.2 Properties of (I - alphaH) - 7.3 Proof of the PageRank Sparse Linear System Chapter 8. Issues in Large-Scale Implementation of PageRank: 8.1 Storage Issues - 8.2 Convergence Criterion - 8.3 Accuracy - 8.4 Dangling Nodes - 8.5 Back Button Modeling
  7. Smeaton, A.F.; Rijsbergen, C.J. van: ¬The retrieval effects of query expansion on a feedback document retrieval system (1983) 0.02
    0.024709776 = product of:
      0.049419552 = sum of:
        0.049419552 = product of:
          0.098839104 = sum of:
            0.098839104 = weight(_text_:22 in 2134) [ClassicSimilarity], result of:
              0.098839104 = score(doc=2134,freq=2.0), product of:
                0.1824739 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05210816 = queryNorm
                0.5416616 = fieldWeight in 2134, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2134)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    30. 3.2001 13:32:22
  8. Back, J.: ¬An evaluation of relevancy ranking techniques used by Internet search engines (2000) 0.02
    0.024709776 = product of:
      0.049419552 = sum of:
        0.049419552 = product of:
          0.098839104 = sum of:
            0.098839104 = weight(_text_:22 in 3445) [ClassicSimilarity], result of:
              0.098839104 = score(doc=3445,freq=2.0), product of:
                0.1824739 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05210816 = queryNorm
                0.5416616 = fieldWeight in 3445, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3445)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    25. 8.2005 17:42:22
  9. Cole, C.: Intelligent information retrieval: diagnosing information need : Part I: the theoretical framework for developing an intelligent IR tool (1998) 0.02
    0.024570452 = product of:
      0.049140904 = sum of:
        0.049140904 = product of:
          0.09828181 = sum of:
            0.09828181 = weight(_text_:i in 6431) [ClassicSimilarity], result of:
              0.09828181 = score(doc=6431,freq=2.0), product of:
                0.1965379 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.05210816 = queryNorm
                0.50006545 = fieldWeight in 6431, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6431)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  10. Wills, R.S.: Google's PageRank : the math behind the search engine (2006) 0.02
    0.023165245 = product of:
      0.04633049 = sum of:
        0.04633049 = product of:
          0.09266098 = sum of:
            0.09266098 = weight(_text_:i in 5954) [ClassicSimilarity], result of:
              0.09266098 = score(doc=5954,freq=16.0), product of:
                0.1965379 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.05210816 = queryNorm
                0.4714662 = fieldWeight in 5954, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5954)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Approximately 91 million American adults use the Internet on a typical day The number-one Internet activity is reading and writing e-mail. Search engine use is next in line and continues to increase in popularity. In fact, survey findings indicate that nearly 60 million American adults use search engines on a given day. Even though there are many Internet search engines, Google, Yahoo!, and MSN receive over 81% of all search requests. Despite claims that the quality of search provided by Yahoo! and MSN now equals that of Google, Google continues to thrive as the search engine of choice, receiving over 46% of all search requests, nearly double the volume of Yahoo! and over four times that of MSN. I use Google's search engine on a daily basis and rarely request information from other search engines. One day, I decided to visit the homepages of Google. Yahoo!, and MSN to compare the quality of search results. Coffee was on my mind that day, so I entered the simple query "coffee" in the search box at each homepage. Table 1 shows the top ten (unsponsored) results returned by each search engine. Although ordered differently, two webpages, www.peets.com and www.coffeegeek.com, appear in all three top ten lists. In addition, each pairing of top ten lists has two additional results in common. Depending on the information I hoped to obtain about coffee by using the search engines, I could argue that any one of the three returned better results: however, I was not looking for a particular webpage, so all three listings of search results seemed of equal quality. Thus, I plan to continue using Google. My decision is indicative of the problem Yahoo!, MSN, and other search engine companies face in the quest to obtain a larger percentage of Internet search volume. Search engine users are loyal to one or a few search engines and are generally happy with search results. Thus, as long as Google continues to provide results deemed high in quality, Google likely will remain the top search engine. But what set Google apart from its competitors in the first place? The answer is PageRank. In this article I explain this simple mathematical algorithm that revolutionized Web search.
  11. Khoo, C.S.G.; Wan, K.-W.: ¬A simple relevancy-ranking strategy for an interface to Boolean OPACs (2004) 0.02
    0.022824306 = sum of:
      0.010469419 = product of:
        0.041877676 = sum of:
          0.041877676 = weight(_text_:authors in 2509) [ClassicSimilarity], result of:
            0.041877676 = score(doc=2509,freq=2.0), product of:
              0.23755142 = queryWeight, product of:
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.05210816 = queryNorm
              0.17628889 = fieldWeight in 2509, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.558814 = idf(docFreq=1258, maxDocs=44218)
                0.02734375 = fieldNorm(doc=2509)
        0.25 = coord(1/4)
      0.012354888 = product of:
        0.024709776 = sum of:
          0.024709776 = weight(_text_:22 in 2509) [ClassicSimilarity], result of:
            0.024709776 = score(doc=2509,freq=2.0), product of:
              0.1824739 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05210816 = queryNorm
              0.1354154 = fieldWeight in 2509, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.02734375 = fieldNorm(doc=2509)
        0.5 = coord(1/2)
    
    Abstract
    A relevancy-ranking algorithm for a natural language interface to Boolean online public access catalogs (OPACs) was formulated and compared with that currently used in a knowledge-based search interface called the E-Referencer, being developed by the authors. The algorithm makes use of seven weIl-known ranking criteria: breadth of match, section weighting, proximity of query words, variant word forms (stemming), document frequency, term frequency and document length. The algorithm converts a natural language query into a series of increasingly broader Boolean search statements. In a small experiment with ten subjects in which the algorithm was simulated by hand, the algorithm obtained good results with a mean overall precision of 0.42 and mean average precision of 0.62, representing a 27 percent improvement in precision and 41 percent improvement in average precision compared to the E-Referencer. The usefulness of each step in the algorithm was analyzed and suggestions are made for improving the algorithm.
    Source
    Electronic library. 22(2004) no.2, S.112-120
  12. Ding, Y.; Yan, E.; Frazho, A.; Caverlee, J.: PageRank for ranking authors in co-citation networks (2009) 0.02
    0.0219812 = product of:
      0.0439624 = sum of:
        0.0439624 = product of:
          0.1758496 = sum of:
            0.1758496 = weight(_text_:authors in 3161) [ClassicSimilarity], result of:
              0.1758496 = score(doc=3161,freq=12.0), product of:
                0.23755142 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.05210816 = queryNorm
                0.7402591 = fieldWeight in 3161, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3161)
          0.25 = coord(1/4)
      0.5 = coord(1/2)
    
    Abstract
    This paper studies how varied damping factors in the PageRank algorithm influence the ranking of authors and proposes weighted PageRank algorithms. We selected the 108 most highly cited authors in the information retrieval (IR) area from the 1970s to 2008 to form the author co-citation network. We calculated the ranks of these 108 authors based on PageRank with the damping factor ranging from 0.05 to 0.95. In order to test the relationship between different measures, we compared PageRank and weighted PageRank results with the citation ranking, h-index, and centrality measures. We found that in our author co-citation network, citation rank is highly correlated with PageRank with different damping factors and also with different weighted PageRank algorithms; citation rank and PageRank are not significantly correlated with centrality measures; and h-index rank does not significantly correlate with centrality measures but does significantly correlate with other measures. The key factors that have impact on the PageRank of authors in the author co-citation network are being co-cited with important authors.
  13. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.02
    0.021179808 = product of:
      0.042359617 = sum of:
        0.042359617 = product of:
          0.08471923 = sum of:
            0.08471923 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
              0.08471923 = score(doc=58,freq=2.0), product of:
                0.1824739 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05210816 = queryNorm
                0.46428138 = fieldWeight in 58, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=58)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:44
  14. Fuhr, N.: Rankingexperimente mit gewichteter Indexierung (1986) 0.02
    0.021179808 = product of:
      0.042359617 = sum of:
        0.042359617 = product of:
          0.08471923 = sum of:
            0.08471923 = weight(_text_:22 in 2051) [ClassicSimilarity], result of:
              0.08471923 = score(doc=2051,freq=2.0), product of:
                0.1824739 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05210816 = queryNorm
                0.46428138 = fieldWeight in 2051, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2051)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:56
  15. Ruthven, I.; Lalmas, M.: Selective relevance feedback using term characteristics (1999) 0.02
    0.020475375 = product of:
      0.04095075 = sum of:
        0.04095075 = product of:
          0.0819015 = sum of:
            0.0819015 = weight(_text_:i in 3824) [ClassicSimilarity], result of:
              0.0819015 = score(doc=3824,freq=2.0), product of:
                0.1965379 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.05210816 = queryNorm
                0.41672117 = fieldWeight in 3824, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3824)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  16. Ding, Y.: Topic-based PageRank on author cocitation networks (2011) 0.02
    0.015543057 = product of:
      0.031086113 = sum of:
        0.031086113 = product of:
          0.12434445 = sum of:
            0.12434445 = weight(_text_:authors in 4348) [ClassicSimilarity], result of:
              0.12434445 = score(doc=4348,freq=6.0), product of:
                0.23755142 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.05210816 = queryNorm
                0.52344227 = fieldWeight in 4348, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4348)
          0.25 = coord(1/4)
      0.5 = coord(1/2)
    
    Abstract
    Ranking authors is vital for identifying a researcher's impact and standing within a scientific field. There are many different ranking methods (e.g., citations, publications, h-index, PageRank, and weighted PageRank), but most of them are topic-independent. This paper proposes topic-dependent ranks based on the combination of a topic model and a weighted PageRank algorithm. The author-conference-topic (ACT) model was used to extract topic distribution of individual authors. Two ways for combining the ACT model with the PageRank algorithm are proposed: simple combination (I_PR) or using a topic distribution as a weighted vector for PageRank (PR_t). Information retrieval was chosen as the test field and representative authors for different topics at different time phases were identified. Principal component analysis (PCA) was applied to analyze the ranking difference between I_PR and PR_t.
  17. Lalmas, M.; Ruthven, I.: Representing and retrieving structured documents using the Dempster-Shafer theory of evidence : modelling and evaluation (1998) 0.01
    0.014332764 = product of:
      0.028665528 = sum of:
        0.028665528 = product of:
          0.057331055 = sum of:
            0.057331055 = weight(_text_:i in 1076) [ClassicSimilarity], result of:
              0.057331055 = score(doc=1076,freq=2.0), product of:
                0.1965379 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.05210816 = queryNorm
                0.29170483 = fieldWeight in 1076, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1076)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  18. Na, S.-H.; Kang, I.-S.; Roh, J.-E.; Lee, J.-H.: ¬An empirical study of query expansion and cluster-based retrieval in language modeling approach (2007) 0.01
    0.014332764 = product of:
      0.028665528 = sum of:
        0.028665528 = product of:
          0.057331055 = sum of:
            0.057331055 = weight(_text_:i in 906) [ClassicSimilarity], result of:
              0.057331055 = score(doc=906,freq=2.0), product of:
                0.1965379 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.05210816 = queryNorm
                0.29170483 = fieldWeight in 906, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=906)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  19. Abdelkareem, M.A.A.: In terms of publication index, what indicator is the best for researchers indexing, Google Scholar, Scopus, Clarivate or others? (2018) 0.01
    0.014332764 = product of:
      0.028665528 = sum of:
        0.028665528 = product of:
          0.057331055 = sum of:
            0.057331055 = weight(_text_:i in 4548) [ClassicSimilarity], result of:
              0.057331055 = score(doc=4548,freq=2.0), product of:
                0.1965379 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.05210816 = queryNorm
                0.29170483 = fieldWeight in 4548, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4548)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    I believe that Google Scholar is the most popular academic indexing way for researchers and citations. However, some other indexing institutions may be more professional than Google Scholar but not as popular as Google Scholar. Other indexing websites like Scopus and Clarivate are providing more statistical figures for scholars, institutions or even journals. On account of publication citations, always Google Scholar shows higher citations for a paper than other indexing websites since Google Scholar consider most of the publication platforms so he can easily count the citations. While other databases just consider the citations come from those journals that are already indexed in their database
  20. MacFarlane, A.; Robertson, S.E.; McCann, J.A.: Parallel computing for passage retrieval (2004) 0.01
    0.014119873 = product of:
      0.028239746 = sum of:
        0.028239746 = product of:
          0.05647949 = sum of:
            0.05647949 = weight(_text_:22 in 5108) [ClassicSimilarity], result of:
              0.05647949 = score(doc=5108,freq=2.0), product of:
                0.1824739 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05210816 = queryNorm
                0.30952093 = fieldWeight in 5108, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5108)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    20. 1.2007 18:30:22

Years

Languages

  • e 55
  • d 6

Types

  • a 52
  • m 5
  • s 2
  • el 1
  • r 1
  • x 1
  • More… Less…