Search (100 results, page 5 of 5)

  • × language_ss:"e"
  • × theme_ss:"Retrievalalgorithmen"
  1. Aigrain, P.; Longueville, V.: ¬A model for the evaluation of expansion techniques in information retrieval systems (1994) 0.00
    2.6145522E-4 = product of:
      0.006274925 = sum of:
        0.006274925 = product of:
          0.018824775 = sum of:
            0.018824775 = weight(_text_:p in 5331) [ClassicSimilarity], result of:
              0.018824775 = score(doc=5331,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.23835106 = fieldWeight in 5331, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5331)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
  2. Lewandowski, D.: How can library materials be ranked in the OPAC? (2009) 0.00
    2.5228495E-4 = product of:
      0.0060548387 = sum of:
        0.0060548387 = weight(_text_:der in 2810) [ClassicSimilarity], result of:
          0.0060548387 = score(doc=2810,freq=2.0), product of:
            0.049067024 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.021966046 = queryNorm
            0.12339935 = fieldWeight in 2810, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2810)
      0.041666668 = coord(1/24)
    
    Content
    Vortrag, anlässlich der 9. Internationalen Bielefeld-Konferenz 2009, 3. - 5. Februar 2009.
  3. Ayadi, H.; Torjmen-Khemakhem, M.; Daoud, M.; Xiangji Huang, J.; Ben Jemaa, M.: MF-Re-Rank : a modality feature-based re-ranking model for medical image retrieval (2018) 0.00
    2.5025924E-4 = product of:
      0.0060062218 = sum of:
        0.0060062218 = product of:
          0.0120124435 = sum of:
            0.0120124435 = weight(_text_:29 in 4459) [ClassicSimilarity], result of:
              0.0120124435 = score(doc=4459,freq=2.0), product of:
                0.07726968 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.021966046 = queryNorm
                0.15546128 = fieldWeight in 4459, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4459)
          0.5 = coord(1/2)
      0.041666668 = coord(1/24)
    
    Date
    29. 9.2018 11:43:31
  4. Information retrieval : data structures and algorithms (1992) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 3495) [ClassicSimilarity], result of:
              0.015687313 = score(doc=3495,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 3495, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3495)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Content
    An edited volume containing data structures and algorithms for information retrieval including a disk with examples written in C. for prgrammers and students interested in parsing text, automated indexing, its the first collection in book form of the basic data structures and algorithms that are critical to the storage and retrieval of documents. ------------------Enthält die Kapitel: FRAKES, W.B.: Introduction to information storage and retrieval systems; BAEZA-YATES, R.S.: Introduction to data structures and algorithms related to information retrieval; HARMAN, D. u.a.: Inverted files; FALOUTSOS, C.: Signature files; GONNET, G.H. u.a.: New indices for text: PAT trees and PAT arrays; FORD, D.A. u. S. CHRISTODOULAKIS: File organizations for optical disks; FOX, C.: Lexical analysis and stoplists; FRAKES, W.B.: Stemming algorithms; SRINIVASAN, P.: Thesaurus construction; BAEZA-YATES, R.A.: String searching algorithms; HARMAN, D.: Relevance feedback and other query modification techniques; WARTIK, S.: Boolean operators; WARTIK, S. u.a.: Hashing algorithms; HARMAN, D.: Ranking algorithms; FOX, E.: u.a.: Extended Boolean models; RASMUSSEN, E.: Clustering algorithms; HOLLAAR, L.: Special-purpose hardware for information retrieval; STANFILL, C.: Parallel information retrieval algorithms
  5. Kantor, P.; Kim, M.H.; Ibraev, U.; Atasoy, K.: Estimating the number of relevant documents in enormous collections (1999) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 6690) [ClassicSimilarity], result of:
              0.015687313 = score(doc=6690,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 6690, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=6690)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
  6. Crouch, C.J.; Crouch, D.B.; Chen, Q.; Holtz, S.J.: Improving the retrieval effectiveness of very short queries (2002) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 2572) [ClassicSimilarity], result of:
              0.015687313 = score(doc=2572,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 2572, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2572)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Abstract
    This paper describes an automatic approach designed to improve the retrieval effectiveness of very short queries such as those used in web searching. The method is based on the observation that stemming, which is designed to maximize recall, often results in depressed precision. Our approach is based on pseudo-feedback and attempts to increase the number of relevant documents in the pseudo-relevant set by reranking those documents based on the presence of unstemmed query terms in the document text. The original experiments underlying this work were carried out using Smart 11.0 and the lnc.ltc weighting scheme on three sets of documents from the TREC collection with corresponding TREC (title only) topics as queries. (The average length of these queries after stoplisting ranges from 2.4 to 4.5 terms.) Results, evaluated in terms of P@20 and non-interpolated average precision, showed clearly that pseudo-feedback (PF) based on this approach was effective in increasing the number of relevant documents in the top ranks. Subsequent experiments, performed on the same data sets using Smart 13.0 and the improved Lnu.ltu weighting scheme, indicate that these results hold up even over the much higher baseline provided by the new weights. Query drift analysis presents a more detailed picture of the improvements produced by this process.
  7. Quiroga, L.M.; Mostafa, J.: ¬An experiment in building profiles in information filtering : the role of context of user relevance feedback (2002) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 2579) [ClassicSimilarity], result of:
              0.015687313 = score(doc=2579,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 2579, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2579)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Abstract
    An experiment was conducted to see how relevance feedback could be used to build and adjust profiles to improve the performance of filtering systems. Data was collected during the system interaction of 18 graduate students with SIFTER (Smart Information Filtering Technology for Electronic Resources), a filtering system that ranks incoming information based on users' profiles. The data set came from a collection of 6000 records concerning consumer health. In the first phase of the study, three different modes of profile acquisition were compared. The explicit mode allowed users to directly specify the profile; the implicit mode utilized relevance feedback to create and refine the profile; and the combined mode allowed users to initialize the profile and to continuously refine it using relevance feedback. Filtering performance, measured in terms of Normalized Precision, showed that the three approaches were significantly different ( [small alpha, Greek] =0.05 and p =0.012). The explicit mode of profile acquisition consistently produced superior results. Exclusive reliance on relevance feedback in the implicit mode resulted in inferior performance. The low performance obtained by the implicit acquisition mode motivated the second phase of the study, which aimed to clarify the role of context in relevance feedback judgments. An inductive content analysis of thinking aloud protocols showed dimensions that were highly situational, establishing the importance context plays in feedback relevance assessments. Results suggest the need for better representation of documents, profiles, and relevance feedback mechanisms that incorporate dimensions identified in this research.
  8. Shah, B.; Raghavan, V.; Dhatric, P.; Zhao, X.: ¬A cluster-based approach for efficient content-based image retrieval using a similarity-preserving space transformation method (2006) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 6118) [ClassicSimilarity], result of:
              0.015687313 = score(doc=6118,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 6118, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=6118)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
  9. Li, J.; Willett, P.: ArticleRank : a PageRank-based alternative to numbers of citations for analysing citation networks (2009) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 751) [ClassicSimilarity], result of:
              0.015687313 = score(doc=751,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 751, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=751)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
  10. Cheng, C.-S.; Chung, C.-P.; Shann, J.J.-J.: Fast query evaluation through document identifier assignment for inverted file-based information retrieval systems (2006) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 979) [ClassicSimilarity], result of:
              0.015687313 = score(doc=979,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 979, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=979)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
  11. Schaefer, A.; Jordan, M.; Klas, C.-P.; Fuhr, N.: Active support for query formulation in virtual digital libraries : a case study with DAFFODIL (2005) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 4296) [ClassicSimilarity], result of:
              0.015687313 = score(doc=4296,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 4296, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4296)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
  12. Symonds, M.; Bruza, P.; Zuccon, G.; Koopman, B.; Sitbon, L.; Turner, I.: Automatic query expansion : a structural linguistic perspective (2014) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 1338) [ClassicSimilarity], result of:
              0.015687313 = score(doc=1338,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 1338, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1338)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
  13. Hoenkamp, E.; Bruza, P.: How everyday language can and will boost effective information retrieval (2015) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 2123) [ClassicSimilarity], result of:
              0.015687313 = score(doc=2123,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 2123, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2123)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
  14. Karisani, P.; Rahgozar, M.; Oroumchian, F.: Transforming LSA space dimensions into a rubric for an automatic assessment and feedback system (2016) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 2970) [ClassicSimilarity], result of:
              0.015687313 = score(doc=2970,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 2970, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2970)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
  15. Jiang, X.; Sun, X.; Yang, Z.; Zhuge, H.; Lapshinova-Koltunski, E.; Yao, J.: Exploiting heterogeneous scientific literature networks to combat ranking bias : evidence from the computational linguistics area (2016) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 3017) [ClassicSimilarity], result of:
              0.015687313 = score(doc=3017,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 3017, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3017)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Abstract
    It is important to help researchers find valuable papers from a large literature collection. To this end, many graph-based ranking algorithms have been proposed. However, most of these algorithms suffer from the problem of ranking bias. Ranking bias hurts the usefulness of a ranking algorithm because it returns a ranking list with an undesirable time distribution. This paper is a focused study on how to alleviate ranking bias by leveraging the heterogeneous network structure of the literature collection. We propose a new graph-based ranking algorithm, MutualRank, that integrates mutual reinforcement relationships among networks of papers, researchers, and venues to achieve a more synthetic, accurate, and less-biased ranking than previous methods. MutualRank provides a unified model that involves both intra- and inter-network information for ranking papers, researchers, and venues simultaneously. We use the ACL Anthology Network as the benchmark data set and construct the gold standard from computer linguistics course websites of well-known universities and two well-known textbooks. The experimental results show that MutualRank greatly outperforms the state-of-the-art competitors, including PageRank, HITS, CoRank, Future Rank, and P-Rank, in ranking papers in both improving ranking effectiveness and alleviating ranking bias. Rankings of researchers and venues by MutualRank are also quite reasonable.
  16. Jiang, J.-D.; Jiang, J.-Y.; Cheng, P.-J.: Cocluster hypothesis and ranking consistency for relevance ranking in web search (2019) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 5247) [ClassicSimilarity], result of:
              0.015687313 = score(doc=5247,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 5247, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5247)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
  17. Jacso, P.: Testing the calculation of a realistic h-index in Google Scholar, Scopus, and Web of Science for F. W. Lancaster (2008) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 5586) [ClassicSimilarity], result of:
              0.015687313 = score(doc=5586,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 5586, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5586)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
  18. Pan, M.; Huang, J.X.; He, T.; Mao, Z.; Ying, Z.; Tu, X.: ¬A simple kernel co-occurrence-based enhancement for pseudo-relevance feedback (2020) 0.00
    2.1787935E-4 = product of:
      0.0052291043 = sum of:
        0.0052291043 = product of:
          0.015687313 = sum of:
            0.015687313 = weight(_text_:p in 5678) [ClassicSimilarity], result of:
              0.015687313 = score(doc=5678,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.19862589 = fieldWeight in 5678, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5678)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Abstract
    Pseudo-relevance feedback is a well-studied query expansion technique in which it is assumed that the top-ranked documents in an initial set of retrieval results are relevant and expansion terms are then extracted from those documents. When selecting expansion terms, most traditional models do not simultaneously consider term frequency and the co-occurrence relationships between candidate terms and query terms. Intuitively, however, a term that has a higher co-occurrence with a query term is more likely to be related to the query topic. In this article, we propose a kernel co-occurrence-based framework to enhance retrieval performance by integrating term co-occurrence information into the Rocchio model and a relevance language model (RM3). Specifically, a kernel co-occurrence-based Rocchio method (KRoc) and a kernel co-occurrence-based RM3 method (KRM3) are proposed. In our framework, co-occurrence information is incorporated into both the factor of the term discrimination power and the factor of the within-document term weight to boost retrieval performance. The results of a series of experiments show that our proposed methods significantly outperform the corresponding strong baselines over all data sets in terms of the mean average precision and over most data sets in terms of P@10. A direct comparison of standard Text Retrieval Conference data sets indicates that our proposed methods are at least comparable to state-of-the-art approaches.
  19. Khoo, C.S.G.; Wan, K.-W.: ¬A simple relevancy-ranking strategy for an interface to Boolean OPACs (2004) 0.00
    2.1700701E-4 = product of:
      0.005208168 = sum of:
        0.005208168 = product of:
          0.010416336 = sum of:
            0.010416336 = weight(_text_:22 in 2509) [ClassicSimilarity], result of:
              0.010416336 = score(doc=2509,freq=2.0), product of:
                0.07692135 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.021966046 = queryNorm
                0.1354154 = fieldWeight in 2509, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=2509)
          0.5 = coord(1/2)
      0.041666668 = coord(1/24)
    
    Source
    Electronic library. 22(2004) no.2, S.112-120
  20. Cross-language information retrieval (1998) 0.00
    1.08939676E-4 = product of:
      0.0026145522 = sum of:
        0.0026145522 = product of:
          0.007843656 = sum of:
            0.007843656 = weight(_text_:p in 6299) [ClassicSimilarity], result of:
              0.007843656 = score(doc=6299,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.099312946 = fieldWeight in 6299, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=6299)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Content
    Enthält die Beiträge: GREFENSTETTE, G.: The Problem of Cross-Language Information Retrieval; DAVIS, M.W.: On the Effective Use of Large Parallel Corpora in Cross-Language Text Retrieval; BALLESTEROS, L. u. W.B. CROFT: Statistical Methods for Cross-Language Information Retrieval; Distributed Cross-Lingual Information Retrieval; Automatic Cross-Language Information Retrieval Using Latent Semantic Indexing; EVANS, D.A. u.a.: Mapping Vocabularies Using Latent Semantics; PICCHI, E. u. C. PETERS: Cross-Language Information Retrieval: A System for Comparable Corpus Querying; YAMABANA, K. u.a.: A Language Conversion Front-End for Cross-Language Information Retrieval; GACHOT, D.A. u.a.: The Systran NLP Browser: An Application of Machine Translation Technology in Cross-Language Information Retrieval; HULL, D.: A Weighted Boolean Model for Cross-Language Text Retrieval; SHERIDAN, P. u.a. Building a Large Multilingual Test Collection from Comparable News Documents; OARD; D.W. u. B.J. DORR: Evaluating Cross-Language Text Filtering Effectiveness

Types

  • a 91
  • el 3
  • m 3
  • p 2
  • s 2
  • d 1
  • r 1
  • More… Less…