Search (78 results, page 4 of 4)

  • × theme_ss:"Retrievalalgorithmen"
  1. Schaefer, A.; Jordan, M.; Klas, C.-P.; Fuhr, N.: Active support for query formulation in virtual digital libraries : a case study with DAFFODIL (2005) 0.01
    0.008123662 = product of:
      0.016247325 = sum of:
        0.016247325 = product of:
          0.03249465 = sum of:
            0.03249465 = weight(_text_:p in 4296) [ClassicSimilarity], result of:
              0.03249465 = score(doc=4296,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19862589 = fieldWeight in 4296, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4296)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  2. Symonds, M.; Bruza, P.; Zuccon, G.; Koopman, B.; Sitbon, L.; Turner, I.: Automatic query expansion : a structural linguistic perspective (2014) 0.01
    0.008123662 = product of:
      0.016247325 = sum of:
        0.016247325 = product of:
          0.03249465 = sum of:
            0.03249465 = weight(_text_:p in 1338) [ClassicSimilarity], result of:
              0.03249465 = score(doc=1338,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19862589 = fieldWeight in 1338, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1338)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  3. Hoenkamp, E.; Bruza, P.: How everyday language can and will boost effective information retrieval (2015) 0.01
    0.008123662 = product of:
      0.016247325 = sum of:
        0.016247325 = product of:
          0.03249465 = sum of:
            0.03249465 = weight(_text_:p in 2123) [ClassicSimilarity], result of:
              0.03249465 = score(doc=2123,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19862589 = fieldWeight in 2123, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2123)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  4. Karisani, P.; Rahgozar, M.; Oroumchian, F.: Transforming LSA space dimensions into a rubric for an automatic assessment and feedback system (2016) 0.01
    0.008123662 = product of:
      0.016247325 = sum of:
        0.016247325 = product of:
          0.03249465 = sum of:
            0.03249465 = weight(_text_:p in 2970) [ClassicSimilarity], result of:
              0.03249465 = score(doc=2970,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19862589 = fieldWeight in 2970, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2970)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  5. Jiang, X.; Sun, X.; Yang, Z.; Zhuge, H.; Lapshinova-Koltunski, E.; Yao, J.: Exploiting heterogeneous scientific literature networks to combat ranking bias : evidence from the computational linguistics area (2016) 0.01
    0.008123662 = product of:
      0.016247325 = sum of:
        0.016247325 = product of:
          0.03249465 = sum of:
            0.03249465 = weight(_text_:p in 3017) [ClassicSimilarity], result of:
              0.03249465 = score(doc=3017,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19862589 = fieldWeight in 3017, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3017)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    It is important to help researchers find valuable papers from a large literature collection. To this end, many graph-based ranking algorithms have been proposed. However, most of these algorithms suffer from the problem of ranking bias. Ranking bias hurts the usefulness of a ranking algorithm because it returns a ranking list with an undesirable time distribution. This paper is a focused study on how to alleviate ranking bias by leveraging the heterogeneous network structure of the literature collection. We propose a new graph-based ranking algorithm, MutualRank, that integrates mutual reinforcement relationships among networks of papers, researchers, and venues to achieve a more synthetic, accurate, and less-biased ranking than previous methods. MutualRank provides a unified model that involves both intra- and inter-network information for ranking papers, researchers, and venues simultaneously. We use the ACL Anthology Network as the benchmark data set and construct the gold standard from computer linguistics course websites of well-known universities and two well-known textbooks. The experimental results show that MutualRank greatly outperforms the state-of-the-art competitors, including PageRank, HITS, CoRank, Future Rank, and P-Rank, in ranking papers in both improving ranking effectiveness and alleviating ranking bias. Rankings of researchers and venues by MutualRank are also quite reasonable.
  6. Jiang, J.-D.; Jiang, J.-Y.; Cheng, P.-J.: Cocluster hypothesis and ranking consistency for relevance ranking in web search (2019) 0.01
    0.008123662 = product of:
      0.016247325 = sum of:
        0.016247325 = product of:
          0.03249465 = sum of:
            0.03249465 = weight(_text_:p in 5247) [ClassicSimilarity], result of:
              0.03249465 = score(doc=5247,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19862589 = fieldWeight in 5247, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5247)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  7. Jacso, P.: Testing the calculation of a realistic h-index in Google Scholar, Scopus, and Web of Science for F. W. Lancaster (2008) 0.01
    0.008123662 = product of:
      0.016247325 = sum of:
        0.016247325 = product of:
          0.03249465 = sum of:
            0.03249465 = weight(_text_:p in 5586) [ClassicSimilarity], result of:
              0.03249465 = score(doc=5586,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19862589 = fieldWeight in 5586, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5586)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  8. Pan, M.; Huang, J.X.; He, T.; Mao, Z.; Ying, Z.; Tu, X.: ¬A simple kernel co-occurrence-based enhancement for pseudo-relevance feedback (2020) 0.01
    0.008123662 = product of:
      0.016247325 = sum of:
        0.016247325 = product of:
          0.03249465 = sum of:
            0.03249465 = weight(_text_:p in 5678) [ClassicSimilarity], result of:
              0.03249465 = score(doc=5678,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19862589 = fieldWeight in 5678, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5678)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Pseudo-relevance feedback is a well-studied query expansion technique in which it is assumed that the top-ranked documents in an initial set of retrieval results are relevant and expansion terms are then extracted from those documents. When selecting expansion terms, most traditional models do not simultaneously consider term frequency and the co-occurrence relationships between candidate terms and query terms. Intuitively, however, a term that has a higher co-occurrence with a query term is more likely to be related to the query topic. In this article, we propose a kernel co-occurrence-based framework to enhance retrieval performance by integrating term co-occurrence information into the Rocchio model and a relevance language model (RM3). Specifically, a kernel co-occurrence-based Rocchio method (KRoc) and a kernel co-occurrence-based RM3 method (KRM3) are proposed. In our framework, co-occurrence information is incorporated into both the factor of the term discrimination power and the factor of the within-document term weight to boost retrieval performance. The results of a series of experiments show that our proposed methods significantly outperform the corresponding strong baselines over all data sets in terms of the mean average precision and over most data sets in terms of P@10. A direct comparison of standard Text Retrieval Conference data sets indicates that our proposed methods are at least comparable to state-of-the-art approaches.
  9. Burgin, R.: ¬The retrieval effectiveness of 5 clustering algorithms as a function of indexing exhaustivity (1995) 0.01
    0.007705845 = product of:
      0.01541169 = sum of:
        0.01541169 = product of:
          0.03082338 = sum of:
            0.03082338 = weight(_text_:22 in 3365) [ClassicSimilarity], result of:
              0.03082338 = score(doc=3365,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19345059 = fieldWeight in 3365, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3365)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 2.1996 11:20:06
  10. Efthimiadis, E.N.: User choices : a new yardstick for the evaluation of ranking algorithms for interactive query expansion (1995) 0.01
    0.007705845 = product of:
      0.01541169 = sum of:
        0.01541169 = product of:
          0.03082338 = sum of:
            0.03082338 = weight(_text_:22 in 5697) [ClassicSimilarity], result of:
              0.03082338 = score(doc=5697,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19345059 = fieldWeight in 5697, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5697)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 2.1996 13:14:10
  11. Song, D.; Bruza, P.D.: Towards context sensitive information inference (2003) 0.01
    0.007705845 = product of:
      0.01541169 = sum of:
        0.01541169 = product of:
          0.03082338 = sum of:
            0.03082338 = weight(_text_:22 in 1428) [ClassicSimilarity], result of:
              0.03082338 = score(doc=1428,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19345059 = fieldWeight in 1428, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1428)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 3.2003 19:35:46
  12. Shiri, A.A.; Revie, C.: Query expansion behavior within a thesaurus-enhanced search environment : a user-centered evaluation (2006) 0.01
    0.007705845 = product of:
      0.01541169 = sum of:
        0.01541169 = product of:
          0.03082338 = sum of:
            0.03082338 = weight(_text_:22 in 56) [ClassicSimilarity], result of:
              0.03082338 = score(doc=56,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19345059 = fieldWeight in 56, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=56)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 7.2006 16:32:43
  13. Dominich, S.: Mathematical foundations of information retrieval (2001) 0.01
    0.007705845 = product of:
      0.01541169 = sum of:
        0.01541169 = product of:
          0.03082338 = sum of:
            0.03082338 = weight(_text_:22 in 1753) [ClassicSimilarity], result of:
              0.03082338 = score(doc=1753,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19345059 = fieldWeight in 1753, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1753)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 3.2008 12:26:32
  14. Soulier, L.; Jabeur, L.B.; Tamine, L.; Bahsoun, W.: On ranking relevant entities in heterogeneous networks using a language-based model (2013) 0.01
    0.007705845 = product of:
      0.01541169 = sum of:
        0.01541169 = product of:
          0.03082338 = sum of:
            0.03082338 = weight(_text_:22 in 664) [ClassicSimilarity], result of:
              0.03082338 = score(doc=664,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.19345059 = fieldWeight in 664, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=664)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 3.2013 19:34:49
  15. Mayr, P.: Re-Ranking auf Basis von Bradfordizing für die verteilte Suche in Digitalen Bibliotheken (2009) 0.01
    0.00649893 = product of:
      0.01299786 = sum of:
        0.01299786 = product of:
          0.02599572 = sum of:
            0.02599572 = weight(_text_:p in 4302) [ClassicSimilarity], result of:
              0.02599572 = score(doc=4302,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.15890071 = fieldWeight in 4302, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4302)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  16. Khoo, C.S.G.; Wan, K.-W.: ¬A simple relevancy-ranking strategy for an interface to Boolean OPACs (2004) 0.01
    0.0053940914 = product of:
      0.010788183 = sum of:
        0.010788183 = product of:
          0.021576365 = sum of:
            0.021576365 = weight(_text_:22 in 2509) [ClassicSimilarity], result of:
              0.021576365 = score(doc=2509,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.1354154 = fieldWeight in 2509, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=2509)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Electronic library. 22(2004) no.2, S.112-120
  17. Cross-language information retrieval (1998) 0.00
    0.004061831 = product of:
      0.008123662 = sum of:
        0.008123662 = product of:
          0.016247325 = sum of:
            0.016247325 = weight(_text_:p in 6299) [ClassicSimilarity], result of:
              0.016247325 = score(doc=6299,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.099312946 = fieldWeight in 6299, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=6299)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Content
    Enthält die Beiträge: GREFENSTETTE, G.: The Problem of Cross-Language Information Retrieval; DAVIS, M.W.: On the Effective Use of Large Parallel Corpora in Cross-Language Text Retrieval; BALLESTEROS, L. u. W.B. CROFT: Statistical Methods for Cross-Language Information Retrieval; Distributed Cross-Lingual Information Retrieval; Automatic Cross-Language Information Retrieval Using Latent Semantic Indexing; EVANS, D.A. u.a.: Mapping Vocabularies Using Latent Semantics; PICCHI, E. u. C. PETERS: Cross-Language Information Retrieval: A System for Comparable Corpus Querying; YAMABANA, K. u.a.: A Language Conversion Front-End for Cross-Language Information Retrieval; GACHOT, D.A. u.a.: The Systran NLP Browser: An Application of Machine Translation Technology in Cross-Language Information Retrieval; HULL, D.: A Weighted Boolean Model for Cross-Language Text Retrieval; SHERIDAN, P. u.a. Building a Large Multilingual Test Collection from Comparable News Documents; OARD; D.W. u. B.J. DORR: Evaluating Cross-Language Text Filtering Effectiveness
  18. Mandl, T.: Tolerantes Information Retrieval : Neuronale Netze zur Erhöhung der Adaptivität und Flexibilität bei der Informationssuche (2001) 0.00
    0.003249465 = product of:
      0.00649893 = sum of:
        0.00649893 = product of:
          0.01299786 = sum of:
            0.01299786 = weight(_text_:p in 5965) [ClassicSimilarity], result of:
              0.01299786 = score(doc=5965,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.079450354 = fieldWeight in 5965, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.015625 = fieldNorm(doc=5965)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Footnote
    da nun nach fast 200 Seiten der Hauptteil der Dissertation folgt - die Vorstellung und Bewertung des bereits erwähnten COSIMIR Modells. Das COSIMIR Modell "berechnet die Ähnlichkeit zwischen den zwei anliegenden Input-Vektoren" (P.194). Der Output des Netzwerks wird an einem einzigen Knoten abgegriffen, an dem sich ein sogenannten Relevanzwert einstellt, wenn die Berechnungen der Gewichtungen interner Knoten zum Abschluss kommen. Diese Gewichtungen hängen von den angelegten Inputvektoren, aus denen die Gewichte der ersten Knotenschicht ermittelt werden, und den im Netzwerk vorgegebenen Kantengewichten ab. Die Gewichtung von Kanten ist der Kernpunkt des neuronalen Ansatzes: In Analogie zum biologischen Urbild (Dendrit mit Synapsen) wächst das Gewicht der Kante mit jeder Aktivierung während einer Trainingsphase. Legt man in dieser Phase zwei Inputvektoren, z.B. Dokumentvektor und Ouery gleichzeitig mit dem Relevanzurteil als Wert des Outputknoten an, verteilen sich durch den BackpropagationProzess die Gewichte entlang der Pfade, die zwischen den beteiligten Knoten bestehen. Da alle Knoten miteinander verbunden sind, entstehen nach mehreren Trainingsbeispielen bereits deutlich unterschiedliche Kantengewichte, weil die aktiv beteiligten Kanten die Änderungen akkumulativ speichern. Eine Variation des Verfahrens benutzt das NN als "Transformationsnetzwerk", wobei die beiden Inputvektoren mit einer Dokumentrepräsentation und einem dazugehörigen Indexat (von einem Experten bereitgestellt) belegt werden. Neben der schon aufgezeigten Trainingsnotwendigkeit weisen die Neuronalen Netze eine weitere intrinsische Problematik auf: Je mehr äußere Knoten benötigt werden, desto mehr interne Kanten (und bei der Verwendung von Zwischenschichten auch Knoten) sind zu verwalten, deren Anzahl nicht linear wächst. Dieser algorithmische Befund setzt naiven Einsätzen der NN-Modelle in der Praxis schnell Grenzen, deshalb ist es umso verdienstvoller, dass der Autor einen innovativen Weg zur Lösung des Problems mit den Mitteln des IR vorschlagen kann. Er verwendet das Latent Semantic Indexing, welches Dokumentrepräsentationen aus einem hochdimensionalen Vektorraum in einen niederdimensionalen abbildet, um die Anzahl der Knoten deutlich zu reduzieren. Damit ist eine sehr schöne Synthese gelungen, welche die eingangs angedeuteten formalen Übereinstimmungen zwischen Vektorraummodellen im IR und den NN aufzeigt und ausnutzt.

Languages

  • e 68
  • d 10

Types

  • a 68
  • m 3
  • r 3
  • p 2
  • s 2
  • el 1
  • x 1
  • More… Less…