Search (72 results, page 1 of 4)

  • × theme_ss:"Retrievalalgorithmen"
  1. Fuhr, N.: Modelle im Information Retrieval (2023) 0.01
    0.0062544122 = product of:
      0.05628971 = sum of:
        0.007527269 = product of:
          0.015054538 = sum of:
            0.015054538 = weight(_text_:29 in 800) [ClassicSimilarity], result of:
              0.015054538 = score(doc=800,freq=2.0), product of:
                0.07747029 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.022023074 = queryNorm
                0.19432661 = fieldWeight in 800, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=800)
          0.5 = coord(1/2)
        0.04876244 = weight(_text_:formale in 800) [ClassicSimilarity], result of:
          0.04876244 = score(doc=800,freq=2.0), product of:
            0.13942613 = queryWeight, product of:
              6.330911 = idf(docFreq=213, maxDocs=44218)
              0.022023074 = queryNorm
            0.34973675 = fieldWeight in 800, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.330911 = idf(docFreq=213, maxDocs=44218)
              0.0390625 = fieldNorm(doc=800)
      0.11111111 = coord(2/18)
    
    Abstract
    Information-Retrieval-Modelle -(IR-Modelle) spezifizieren, wie zu einer gegebenen Anfrage die Antwortdokumente aus einer Dokumentenkollektion bestimmt werden. Ausgangsbasis jedes Modells sind dabei zunächst bestimmte Annahmen über die Wissensrepräsentation (s. Teil B Methoden und Systeme der Inhaltserschließung) von Fragen und Dokumenten. Hier bezeichnen wir die Elemente dieser Repräsentationen als Terme, wobei es aus der Sicht des Modells egal ist, wie diese Terme aus dem Dokument (und analog aus der von Benutzenden eingegebenen Anfrage) abgeleitet werden: Bei Texten werden hierzu häufig computerlinguistische Methoden eingesetzt, aber auch komplexere automatische oder manuelle Erschließungsverfahren können zur Anwendung kommen. Repräsentationen besitzen ferner eine bestimmte Struktur. Ein Dokument wird meist als Menge oder Multimenge von Termen aufgefasst, wobei im zweiten Fall das Mehrfachvorkommen berücksichtigt wird. Diese Dokumentrepräsentation wird wiederum auf eine sogenannte Dokumentbeschreibung abgebildet, in der die einzelnen Terme gewichtet sein können. Im Folgenden unterscheiden wir nur zwischen ungewichteter (Gewicht eines Terms ist entweder 0 oder 1) und gewichteter Indexierung (das Gewicht ist eine nichtnegative reelle Zahl). Analog dazu gibt es eine Fragerepräsentation; legt man eine natürlichsprachige Anfrage zugrunde, so kann man die o. g. Verfahren für Dokumenttexte anwenden. Alternativ werden auch grafische oder formale Anfragesprachen verwendet, wobei aus Sicht der Modelle insbesondere deren logische Struktur (etwa beim Booleschen Retrieval) relevant ist. Die Fragerepräsentation wird dann in eine Fragebeschreibung überführt.
    Date
    24.11.2022 17:20:29
  2. Mandl, T.: Tolerantes Information Retrieval : Neuronale Netze zur Erhöhung der Adaptivität und Flexibilität bei der Informationssuche (2001) 0.01
    0.0054112542 = product of:
      0.048701286 = sum of:
        0.019504976 = weight(_text_:formale in 5965) [ClassicSimilarity], result of:
          0.019504976 = score(doc=5965,freq=2.0), product of:
            0.13942613 = queryWeight, product of:
              6.330911 = idf(docFreq=213, maxDocs=44218)
              0.022023074 = queryNorm
            0.1398947 = fieldWeight in 5965, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.330911 = idf(docFreq=213, maxDocs=44218)
              0.015625 = fieldNorm(doc=5965)
        0.02919631 = weight(_text_:komponente in 5965) [ClassicSimilarity], result of:
          0.02919631 = score(doc=5965,freq=2.0), product of:
            0.17058289 = queryWeight, product of:
              7.7456436 = idf(docFreq=51, maxDocs=44218)
              0.022023074 = queryNorm
            0.17115615 = fieldWeight in 5965, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7456436 = idf(docFreq=51, maxDocs=44218)
              0.015625 = fieldNorm(doc=5965)
      0.11111111 = coord(2/18)
    
    Footnote
    Rez. in: nfd - Information 54(2003) H.6, S.379-380 (U. Thiel): "Kannte G. Salton bei der Entwicklung des Vektorraummodells die kybernetisch orientierten Versuche mit assoziativen Speicherstrukturen? An diese und ähnliche Vermutungen, die ich vor einigen Jahren mit Reginald Ferber und anderen Kollegen diskutierte, erinnerte mich die Thematik des vorliegenden Buches. Immerhin lässt sich feststellen, dass die Vektorrepräsentation eine genial einfache Darstellung sowohl der im Information Retrieval (IR) als grundlegende Datenstruktur benutzten "inverted files" als auch der assoziativen Speichermatrizen darstellt, die sich im Laufe der Zeit Über Perzeptrons zu Neuronalen Netzen (NN) weiterentwickelten. Dieser formale Zusammenhang stimulierte in der Folge eine Reihe von Ansätzen, die Netzwerke im Retrieval zu verwenden, wobei sich, wie auch im vorliegenden Band, hybride Ansätze, die Methoden aus beiden Disziplinen kombinieren, als sehr geeignet erweisen. Aber der Reihe nach... Das Buch wurde vom Autor als Dissertation beim Fachbereich IV "Sprachen und Technik" der Universität Hildesheim eingereicht und resultiert aus einer Folge von Forschungsbeiträgen zu mehreren Projekten, an denen der Autor in der Zeit von 1995 bis 2000 an verschiedenen Standorten beteiligt war. Dies erklärt die ungewohnte Breite der Anwendungen, Szenarien und Domänen, in denen die Ergebnisse gewonnen wurden. So wird das in der Arbeit entwickelte COSIMIR Modell (COgnitive SIMilarity learning in Information Retrieval) nicht nur anhand der klassischen Cranfield-Kollektion evaluiert, sondern auch im WING-Projekt der Universität Regensburg im Faktenretrieval aus einer Werkstoffdatenbank eingesetzt. Weitere Versuche mit der als "Transformations-Netzwerk" bezeichneten Komponente, deren Aufgabe die Abbildung von Gewichtungsfunktionen zwischen zwei Termräumen ist, runden das Spektrum der Experimente ab. Aber nicht nur die vorgestellten Resultate sind vielfältig, auch der dem Leser angebotene "State-of-the-Art"-Überblick fasst in hoch informativer Breite Wesentliches aus den Gebieten IR und NN zusammen und beleuchtet die Schnittpunkte der beiden Bereiche. So werden neben den Grundlagen des Text- und Faktenretrieval die Ansätze zur Verbesserung der Adaptivität und zur Beherrschung von Heterogenität vorgestellt, während als Grundlagen Neuronaler Netze neben einer allgemeinen Einführung in die Grundbegriffe u.a. das Backpropagation-Modell, KohonenNetze und die Adaptive Resonance Theory (ART) geschildert werden. Einweiteres Kapitel stellt die bisherigen NN-orientierten Ansätze im IR vor und rundet den Abriss der relevanten Forschungslandschaft ab. Als Vorbereitung der Präsentation des COSIMIR-Modells schiebt der Autor an dieser Stelle ein diskursives Kapitel zum Thema Heterogenität im IR ein, wodurch die Ziele und Grundannahmen der Arbeit noch einmal reflektiert werden. Als Dimensionen der Heterogenität werden der Objekttyp, die Qualität der Objekte und ihrer Erschließung und die Mehrsprachigkeit genannt. Wenn auch diese Systematik im Wesentlichen die Akzente auf Probleme aus den hier tangierten Projekten legt, und weniger eine umfassende Aufbereitung z.B. der Literatur zum Problem der Relevanz anstrebt, ist sie dennoch hilfreich zum Verständnis der in den nachfolgenden Kapitel oft nur implizit angesprochenen Designentscheidungen bei der Konzeption der entwickelten Prototypen. Der Ansatz, Heterogenität durch Transformationen zu behandeln, wird im speziellen Kontext der NN konkretisiert, wobei andere Möglichkeiten, die z.B. Instrumente der Logik und Probabilistik einzusetzen, nur kurz diskutiert werden. Eine weitergehende Analyse hätte wohl auch den Rahmen der Arbeit zu weit gespannt,
  3. Chang, C.-H.; Hsu, C.-C.: Integrating query expansion and conceptual relevance feedback for personalized Web information retrieval (1998) 0.00
    0.0028905603 = product of:
      0.026015043 = sum of:
        0.01905279 = product of:
          0.03810558 = sum of:
            0.03810558 = weight(_text_:networks in 1319) [ClassicSimilarity], result of:
              0.03810558 = score(doc=1319,freq=2.0), product of:
                0.10416738 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.022023074 = queryNorm
                0.36581108 = fieldWeight in 1319, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1319)
          0.5 = coord(1/2)
        0.006962253 = product of:
          0.020886758 = sum of:
            0.020886758 = weight(_text_:22 in 1319) [ClassicSimilarity], result of:
              0.020886758 = score(doc=1319,freq=2.0), product of:
                0.07712106 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.022023074 = queryNorm
                0.2708308 = fieldWeight in 1319, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1319)
          0.33333334 = coord(1/3)
      0.11111111 = coord(2/18)
    
    Date
    1. 8.1996 22:08:06
    Source
    Computer networks and ISDN systems. 30(1998) nos.1/7, S.621-623
  4. Mutschke, P.: Autorennetzwerke : Verfahren zur Netzwerkanalyse als Mehrwertdienste für Informationssysteme (2004) 0.00
    0.0028329256 = product of:
      0.05099266 = sum of:
        0.05099266 = weight(_text_:virtuelle in 4050) [ClassicSimilarity], result of:
          0.05099266 = score(doc=4050,freq=2.0), product of:
            0.13015614 = queryWeight, product of:
              5.90999 = idf(docFreq=325, maxDocs=44218)
              0.022023074 = queryNorm
            0.39178067 = fieldWeight in 4050, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.90999 = idf(docFreq=325, maxDocs=44218)
              0.046875 = fieldNorm(doc=4050)
      0.055555556 = coord(1/18)
    
    Abstract
    Virtuelle Bibliotheken enthalten eine Fülle an Informationen, die in ihrer Vielfalt und Tiefe von Standardsuchmaschinen nicht erschöpfend erfasst wird. Der Arbeitsbericht informiert über Entwicklungen am IZ, die darauf abzielen, Wissen über das Interaktionsgeschehen in wissenschaftlichen Communities und den sozialen Status ihrer Akteure für das Retrieval auszunutzen. Grundlage hierfür sind soziale Netzwerke, die sich durch Kooperation der wissenschaftlichen Akteure konstituieren und in den Dokumenten der Datenbasis z.B. als Koautorbeziehungen repräsentiert sind (Autorennetzwerke). Die in dem Bericht beschriebenen Studien zur Small-World-Topologie von Autorennetzwerken zeigen, dass diese Netzwerke ein erhebliches Potential für Informationssysteme haben. Der Bericht diskutiert Szenarios, die beschreiben, wie Autorennetzwerke und hier insbesondere das Konzept der Akteurszentralität für die Informationssuche in Datenbanken sinnvoll genutzt werden können. Kernansatz dieser Retrievalmodelle ist die Suche nach Experten und das Ranking von Dokumenten auf der Basis der Zentralität von Autoren in Autorennetzwerken.
  5. Carpineto, C.; Romano, G.: Order-theoretical ranking (2000) 0.00
    0.0027090246 = product of:
      0.04876244 = sum of:
        0.04876244 = weight(_text_:formale in 4766) [ClassicSimilarity], result of:
          0.04876244 = score(doc=4766,freq=2.0), product of:
            0.13942613 = queryWeight, product of:
              6.330911 = idf(docFreq=213, maxDocs=44218)
              0.022023074 = queryNorm
            0.34973675 = fieldWeight in 4766, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.330911 = idf(docFreq=213, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4766)
      0.055555556 = coord(1/18)
    
    Theme
    Formale Begriffsanalyse
  6. Kulyukin, V.A.; Settle, A.: Ranked retrieval with semantic networks and vector spaces (2001) 0.00
    0.0024194021 = product of:
      0.043549236 = sum of:
        0.043549236 = product of:
          0.08709847 = sum of:
            0.08709847 = weight(_text_:networks in 6934) [ClassicSimilarity], result of:
              0.08709847 = score(doc=6934,freq=8.0), product of:
                0.10416738 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.022023074 = queryNorm
                0.8361396 = fieldWeight in 6934, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6934)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
    Abstract
    The equivalence of semantic networks with spreading activation and vector spaces with dot product is investigated under ranked retrieval. Semantic networks are viewed as networks of concepts organized in terms of abstraction and packaging relations. It is shown that the two models can be effectively constructed from each other. A formal method is suggested to analyze the models in terms of their relative performance in the same universe of objects
  7. Zhang, D.; Dong, Y.: ¬An effective algorithm to rank Web resources (2000) 0.00
    0.0021169768 = product of:
      0.03810558 = sum of:
        0.03810558 = product of:
          0.07621116 = sum of:
            0.07621116 = weight(_text_:networks in 3662) [ClassicSimilarity], result of:
              0.07621116 = score(doc=3662,freq=2.0), product of:
                0.10416738 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.022023074 = queryNorm
                0.73162216 = fieldWeight in 3662, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3662)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
    Source
    Computer networks. 33(2000) nos.1/6, S.449-455
  8. Finding anything in the billion page Web : are algorithms the key? (1999) 0.00
    0.0021169768 = product of:
      0.03810558 = sum of:
        0.03810558 = product of:
          0.07621116 = sum of:
            0.07621116 = weight(_text_:networks in 6248) [ClassicSimilarity], result of:
              0.07621116 = score(doc=6248,freq=2.0), product of:
                0.10416738 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.022023074 = queryNorm
                0.73162216 = fieldWeight in 6248, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6248)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
    Source
    Computer networks. 31(1999) no.11-16, S.1760-1761
  9. Soulier, L.; Jabeur, L.B.; Tamine, L.; Bahsoun, W.: On ranking relevant entities in heterogeneous networks using a language-based model (2013) 0.00
    0.0020646858 = product of:
      0.018582173 = sum of:
        0.0136091355 = product of:
          0.027218271 = sum of:
            0.027218271 = weight(_text_:networks in 664) [ClassicSimilarity], result of:
              0.027218271 = score(doc=664,freq=2.0), product of:
                0.10416738 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.022023074 = queryNorm
                0.26129362 = fieldWeight in 664, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=664)
          0.5 = coord(1/2)
        0.004973038 = product of:
          0.014919113 = sum of:
            0.014919113 = weight(_text_:22 in 664) [ClassicSimilarity], result of:
              0.014919113 = score(doc=664,freq=2.0), product of:
                0.07712106 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.022023074 = queryNorm
                0.19345059 = fieldWeight in 664, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=664)
          0.33333334 = coord(1/3)
      0.11111111 = coord(2/18)
    
    Date
    22. 3.2013 19:34:49
  10. Yan, E.; Ding, Y.; Sugimoto, C.R.: P-Rank: an indicator measuring prestige in heterogeneous scholarly networks (2011) 0.00
    0.0020287302 = product of:
      0.036517143 = sum of:
        0.036517143 = product of:
          0.07303429 = sum of:
            0.07303429 = weight(_text_:networks in 4349) [ClassicSimilarity], result of:
              0.07303429 = score(doc=4349,freq=10.0), product of:
                0.10416738 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.022023074 = queryNorm
                0.70112437 = fieldWeight in 4349, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4349)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
    Abstract
    Ranking scientific productivity and prestige are often limited to homogeneous networks. These networks are unable to account for the multiple factors that constitute the scholarly communication and reward system. This study proposes a new informetric indicator, P-Rank, for measuring prestige in heterogeneous scholarly networks containing articles, authors, and journals. P-Rank differentiates the weight of each citation based on its citing papers, citing journals, and citing authors. Articles from 16 representative library and information science journals are selected as the dataset. Principle Component Analysis is conducted to examine the relationship between P-Rank and other bibliometric indicators. We also compare the correlation and rank variances between citation counts and P-Rank scores. This work provides a new approach to examining prestige in scholarly communication networks in a more comprehensive and nuanced way.
  11. Archuby, C.G.: Interfaces se recuperacion para catalogos en linea con salidas ordenadas por probable relevancia (2000) 0.00
    0.0011827962 = product of:
      0.021290332 = sum of:
        0.021290332 = product of:
          0.042580664 = sum of:
            0.042580664 = weight(_text_:29 in 5727) [ClassicSimilarity], result of:
              0.042580664 = score(doc=5727,freq=4.0), product of:
                0.07747029 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.022023074 = queryNorm
                0.5496386 = fieldWeight in 5727, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=5727)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
    Date
    29. 1.1996 18:23:13
    Source
    Ciencia da informacao. 29(2000) no.3, S.5-13
  12. Crestani, F.: Combination of similarity measures for effective spoken document retrieval (2003) 0.00
    0.0011709085 = product of:
      0.021076351 = sum of:
        0.021076351 = product of:
          0.042152703 = sum of:
            0.042152703 = weight(_text_:29 in 4690) [ClassicSimilarity], result of:
              0.042152703 = score(doc=4690,freq=2.0), product of:
                0.07747029 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.022023074 = queryNorm
                0.5441145 = fieldWeight in 4690, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4690)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
    Source
    Journal of information science. 29(2003) no.2, S.87-96
  13. Li, J.; Willett, P.: ArticleRank : a PageRank-based alternative to numbers of citations for analysing citation networks (2009) 0.00
    0.0010692347 = product of:
      0.019246224 = sum of:
        0.019246224 = product of:
          0.03849245 = sum of:
            0.03849245 = weight(_text_:networks in 751) [ClassicSimilarity], result of:
              0.03849245 = score(doc=751,freq=4.0), product of:
                0.10416738 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.022023074 = queryNorm
                0.369525 = fieldWeight in 751, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=751)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
    Abstract
    Purpose - The purpose of this paper is to suggest an alternative to the widely used Times Cited criterion for analysing citation networks. The approach involves taking account of the natures of the papers that cite a given paper, so as to differentiate between papers that attract the same number of citations. Design/methodology/approach - ArticleRank is an algorithm that has been derived from Google's PageRank algorithm to measure the influence of journal articles. ArticleRank is applied to two datasets - a citation network based on an early paper on webometrics, and a self-citation network based on the 19 most cited papers in the Journal of Documentation - using citation data taken from the Web of Knowledge database. Findings - ArticleRank values provide a different ranking of a set of papers from that provided by the corresponding Times Cited values, and overcomes the inability of the latter to differentiate between papers with the same numbers of citations. The difference in rankings between Times Cited and ArticleRank is greatest for the most heavily cited articles in a dataset. Originality/value - This is a novel application of the PageRank algorithm.
  14. Jiang, X.; Sun, X.; Yang, Z.; Zhuge, H.; Lapshinova-Koltunski, E.; Yao, J.: Exploiting heterogeneous scientific literature networks to combat ranking bias : evidence from the computational linguistics area (2016) 0.00
    0.0010692347 = product of:
      0.019246224 = sum of:
        0.019246224 = product of:
          0.03849245 = sum of:
            0.03849245 = weight(_text_:networks in 3017) [ClassicSimilarity], result of:
              0.03849245 = score(doc=3017,freq=4.0), product of:
                0.10416738 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.022023074 = queryNorm
                0.369525 = fieldWeight in 3017, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3017)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
    Abstract
    It is important to help researchers find valuable papers from a large literature collection. To this end, many graph-based ranking algorithms have been proposed. However, most of these algorithms suffer from the problem of ranking bias. Ranking bias hurts the usefulness of a ranking algorithm because it returns a ranking list with an undesirable time distribution. This paper is a focused study on how to alleviate ranking bias by leveraging the heterogeneous network structure of the literature collection. We propose a new graph-based ranking algorithm, MutualRank, that integrates mutual reinforcement relationships among networks of papers, researchers, and venues to achieve a more synthetic, accurate, and less-biased ranking than previous methods. MutualRank provides a unified model that involves both intra- and inter-network information for ranking papers, researchers, and venues simultaneously. We use the ACL Anthology Network as the benchmark data set and construct the gold standard from computer linguistics course websites of well-known universities and two well-known textbooks. The experimental results show that MutualRank greatly outperforms the state-of-the-art competitors, including PageRank, HITS, CoRank, Future Rank, and P-Rank, in ranking papers in both improving ranking effectiveness and alleviating ranking bias. Rankings of researchers and venues by MutualRank are also quite reasonable.
  15. Okada, M.; Ando, K.; Lee, S.S.; Hayashi, Y.; Aoe, J.I.: ¬An efficient substring search method by using delayed keyword extraction (2001) 0.00
    0.0010036358 = product of:
      0.018065445 = sum of:
        0.018065445 = product of:
          0.03613089 = sum of:
            0.03613089 = weight(_text_:29 in 6415) [ClassicSimilarity], result of:
              0.03613089 = score(doc=6415,freq=2.0), product of:
                0.07747029 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.022023074 = queryNorm
                0.46638384 = fieldWeight in 6415, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6415)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
    Date
    29. 3.2002 17:24:03
  16. Cole, C.: Intelligent information retrieval: diagnosing information need : Part II: uncertainty expansion in a prototype of a diagnostic IR tool (1998) 0.00
    0.0010036358 = product of:
      0.018065445 = sum of:
        0.018065445 = product of:
          0.03613089 = sum of:
            0.03613089 = weight(_text_:29 in 6432) [ClassicSimilarity], result of:
              0.03613089 = score(doc=6432,freq=2.0), product of:
                0.07747029 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.022023074 = queryNorm
                0.46638384 = fieldWeight in 6432, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6432)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
    Date
    11. 8.2001 14:48:29
  17. Ding, Y.; Yan, E.; Frazho, A.; Caverlee, J.: PageRank for ranking authors in co-citation networks (2009) 0.00
    9.072757E-4 = product of:
      0.016330963 = sum of:
        0.016330963 = product of:
          0.032661926 = sum of:
            0.032661926 = weight(_text_:networks in 3161) [ClassicSimilarity], result of:
              0.032661926 = score(doc=3161,freq=2.0), product of:
                0.10416738 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.022023074 = queryNorm
                0.31355235 = fieldWeight in 3161, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3161)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
  18. Ding, Y.: Topic-based PageRank on author cocitation networks (2011) 0.00
    9.072757E-4 = product of:
      0.016330963 = sum of:
        0.016330963 = product of:
          0.032661926 = sum of:
            0.032661926 = weight(_text_:networks in 4348) [ClassicSimilarity], result of:
              0.032661926 = score(doc=4348,freq=2.0), product of:
                0.10416738 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.022023074 = queryNorm
                0.31355235 = fieldWeight in 4348, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4348)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
  19. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.00
    8.840956E-4 = product of:
      0.015913721 = sum of:
        0.015913721 = product of:
          0.047741164 = sum of:
            0.047741164 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.047741164 = score(doc=402,freq=2.0), product of:
                0.07712106 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.022023074 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.33333334 = coord(1/3)
      0.055555556 = coord(1/18)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  20. Zhang, W.; Korf, R.E.: Performance of linear-space search algorithms (1995) 0.00
    8.3636324E-4 = product of:
      0.015054538 = sum of:
        0.015054538 = product of:
          0.030109076 = sum of:
            0.030109076 = weight(_text_:29 in 4744) [ClassicSimilarity], result of:
              0.030109076 = score(doc=4744,freq=2.0), product of:
                0.07747029 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.022023074 = queryNorm
                0.38865322 = fieldWeight in 4744, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4744)
          0.5 = coord(1/2)
      0.055555556 = coord(1/18)
    
    Date
    2. 8.1996 10:29:15

Years

Languages

  • e 58
  • d 13
  • pt 1
  • More… Less…

Types

  • a 67
  • m 2
  • r 2
  • el 1
  • x 1
  • More… Less…