Search (68 results, page 1 of 4)

  • × theme_ss:"Retrievalalgorithmen"
  1. Zhang, D.; Dong, Y.: ¬An effective algorithm to rank Web resources (2000) 0.05
    0.046643518 = product of:
      0.13993055 = sum of:
        0.09335803 = weight(_text_:computer in 3662) [ClassicSimilarity], result of:
          0.09335803 = score(doc=3662,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.56527805 = fieldWeight in 3662, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.109375 = fieldNorm(doc=3662)
        0.046572514 = product of:
          0.09314503 = sum of:
            0.09314503 = weight(_text_:resources in 3662) [ClassicSimilarity], result of:
              0.09314503 = score(doc=3662,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.56463283 = fieldWeight in 3662, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3662)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Source
    Computer networks. 33(2000) nos.1/6, S.449-455
  2. Smeaton, A.F.; Rijsbergen, C.J. van: ¬The retrieval effects of query expansion on a feedback document retrieval system (1983) 0.05
    0.04540606 = product of:
      0.13621818 = sum of:
        0.09335803 = weight(_text_:computer in 2134) [ClassicSimilarity], result of:
          0.09335803 = score(doc=2134,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.56527805 = fieldWeight in 2134, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.109375 = fieldNorm(doc=2134)
        0.04286014 = product of:
          0.08572028 = sum of:
            0.08572028 = weight(_text_:22 in 2134) [ClassicSimilarity], result of:
              0.08572028 = score(doc=2134,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.5416616 = fieldWeight in 2134, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2134)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Date
    30. 3.2001 13:32:22
    Source
    Computer journal. 26(1983), S.239-246
  3. Chen, H.; Zhang, Y.; Houston, A.L.: Semantic indexing and searching using a Hopfield net (1998) 0.04
    0.041345302 = product of:
      0.1240359 = sum of:
        0.040010586 = weight(_text_:computer in 5704) [ClassicSimilarity], result of:
          0.040010586 = score(doc=5704,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.24226204 = fieldWeight in 5704, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.046875 = fieldNorm(doc=5704)
        0.084025316 = weight(_text_:network in 5704) [ClassicSimilarity], result of:
          0.084025316 = score(doc=5704,freq=4.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.41750383 = fieldWeight in 5704, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.046875 = fieldNorm(doc=5704)
      0.33333334 = coord(2/6)
    
    Abstract
    Presents a neural network approach to document semantic indexing. Reports results of a study to apply a Hopfield net algorithm to simulate human associative memory for concept exploration in the domain of computer science and engineering. The INSPEC database, consisting of 320.000 abstracts from leading periodical articles was used as the document test bed. Benchmark tests conformed that 3 parameters: maximum number of activated nodes; maximum allowable error; and maximum number of iterations; were useful in positively influencing network convergence behaviour without negatively impacting central processing unit performance. Another series of benchmark tests was performed to determine the effectiveness of various filtering techniques in reducing the negative impact of noisy input terms. Preliminary user tests conformed expectations that the Hopfield net is potentially useful as an associative memory technique to improve document recall and precision by solving discrepancies between indexer vocabularies and end user vocabularies
  4. Campos, L.M. de; Fernández-Luna, J.M.; Huete, J.F.: Implementing relevance feedback in the Bayesian network retrieval model (2003) 0.04
    0.040426075 = product of:
      0.12127822 = sum of:
        0.10290959 = weight(_text_:network in 825) [ClassicSimilarity], result of:
          0.10290959 = score(doc=825,freq=6.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.51133573 = fieldWeight in 825, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.046875 = fieldNorm(doc=825)
        0.018368632 = product of:
          0.036737263 = sum of:
            0.036737263 = weight(_text_:22 in 825) [ClassicSimilarity], result of:
              0.036737263 = score(doc=825,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.23214069 = fieldWeight in 825, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=825)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Relevance Feedback consists in automatically formulating a new query according to the relevance judgments provided by the user after evaluating a set of retrieved documents. In this article, we introduce several relevance feedback methods for the Bayesian Network Retrieval ModeL The theoretical frame an which our methods are based uses the concept of partial evidences, which summarize the new pieces of information gathered after evaluating the results obtained by the original query. These partial evidences are inserted into the underlying Bayesian network and a new inference process (probabilities propagation) is run to compute the posterior relevance probabilities of the documents in the collection given the new query. The quality of the proposed methods is tested using a preliminary experimentation with different standard document collections.
    Date
    22. 3.2003 19:30:19
  5. Jiang, X.; Sun, X.; Yang, Z.; Zhuge, H.; Lapshinova-Koltunski, E.; Yao, J.: Exploiting heterogeneous scientific literature networks to combat ranking bias : evidence from the computational linguistics area (2016) 0.04
    0.039700046 = product of:
      0.11910014 = sum of:
        0.033342157 = weight(_text_:computer in 3017) [ClassicSimilarity], result of:
          0.033342157 = score(doc=3017,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.20188503 = fieldWeight in 3017, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3017)
        0.08575798 = weight(_text_:network in 3017) [ClassicSimilarity], result of:
          0.08575798 = score(doc=3017,freq=6.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.42611307 = fieldWeight in 3017, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3017)
      0.33333334 = coord(2/6)
    
    Abstract
    It is important to help researchers find valuable papers from a large literature collection. To this end, many graph-based ranking algorithms have been proposed. However, most of these algorithms suffer from the problem of ranking bias. Ranking bias hurts the usefulness of a ranking algorithm because it returns a ranking list with an undesirable time distribution. This paper is a focused study on how to alleviate ranking bias by leveraging the heterogeneous network structure of the literature collection. We propose a new graph-based ranking algorithm, MutualRank, that integrates mutual reinforcement relationships among networks of papers, researchers, and venues to achieve a more synthetic, accurate, and less-biased ranking than previous methods. MutualRank provides a unified model that involves both intra- and inter-network information for ranking papers, researchers, and venues simultaneously. We use the ACL Anthology Network as the benchmark data set and construct the gold standard from computer linguistics course websites of well-known universities and two well-known textbooks. The experimental results show that MutualRank greatly outperforms the state-of-the-art competitors, including PageRank, HITS, CoRank, Future Rank, and P-Rank, in ranking papers in both improving ranking effectiveness and alleviating ranking bias. Rankings of researchers and venues by MutualRank are also quite reasonable.
  6. Soulier, L.; Jabeur, L.B.; Tamine, L.; Bahsoun, W.: On ranking relevant entities in heterogeneous networks using a language-based model (2013) 0.04
    0.038110662 = product of:
      0.11433198 = sum of:
        0.09902479 = weight(_text_:network in 664) [ClassicSimilarity], result of:
          0.09902479 = score(doc=664,freq=8.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.492033 = fieldWeight in 664, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0390625 = fieldNorm(doc=664)
        0.015307193 = product of:
          0.030614385 = sum of:
            0.030614385 = weight(_text_:22 in 664) [ClassicSimilarity], result of:
              0.030614385 = score(doc=664,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.19345059 = fieldWeight in 664, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=664)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    A new challenge, accessing multiple relevant entities, arises from the availability of linked heterogeneous data. In this article, we address more specifically the problem of accessing relevant entities, such as publications and authors within a bibliographic network, given an information need. We propose a novel algorithm, called BibRank, that estimates a joint relevance of documents and authors within a bibliographic network. This model ranks each type of entity using a score propagation algorithm with respect to the query topic and the structure of the underlying bi-type information entity network. Evidence sources, namely content-based and network-based scores, are both used to estimate the topical similarity between connected entities. For this purpose, authorship relationships are analyzed through a language model-based score on the one hand and on the other hand, non topically related entities of the same type are detected through marginal citations. The article reports the results of experiments using the Bibrank algorithm for an information retrieval task. The CiteSeerX bibliographic data set forms the basis for the topical query automatic generation and evaluation. We show that a statistically significant improvement over closely related ranking models is achieved.
    Date
    22. 3.2013 19:34:49
  7. Chang, R.: Keyword searching and indexing (1993) 0.04
    0.035729438 = product of:
      0.107188314 = sum of:
        0.05384086 = weight(_text_:services in 7223) [ClassicSimilarity], result of:
          0.05384086 = score(doc=7223,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.3245064 = fieldWeight in 7223, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.0625 = fieldNorm(doc=7223)
        0.05334745 = weight(_text_:computer in 7223) [ClassicSimilarity], result of:
          0.05334745 = score(doc=7223,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.32301605 = fieldWeight in 7223, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0625 = fieldNorm(doc=7223)
      0.33333334 = coord(2/6)
    
    Abstract
    Explains how a computer indexing system works. Reviews fundamentals of how data are stored and retrieved by computers. Describes B-Tree and B+-Tree indexing structures. Gives basic keyword searching techniques that the user must apply to make use of the indexing programs. The demand for keyword retrieval is increasing and librarians should expect to see the keyword-indexing feature become commonly available
    Source
    Technical services quarterly. 10(1993) no.4, S.75-86
  8. Lalmas, M.: XML retrieval (2009) 0.03
    0.02693449 = product of:
      0.08080347 = sum of:
        0.03365054 = weight(_text_:services in 4998) [ClassicSimilarity], result of:
          0.03365054 = score(doc=4998,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.2028165 = fieldWeight in 4998, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4998)
        0.04715293 = weight(_text_:computer in 4998) [ClassicSimilarity], result of:
          0.04715293 = score(doc=4998,freq=4.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.28550854 = fieldWeight in 4998, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4998)
      0.33333334 = coord(2/6)
    
    LCSH
    Query languages (Computer science)
    Series
    Synthesis lectures on information concepts, retrieval & services; 7
    Subject
    Query languages (Computer science)
  9. MacFarlane, A.; McCann, J.A.; Robertson, S.E.: Parallel methods for the generation of partitioned inverted files (2005) 0.03
    0.026797077 = product of:
      0.08039123 = sum of:
        0.040380646 = weight(_text_:services in 651) [ClassicSimilarity], result of:
          0.040380646 = score(doc=651,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.2433798 = fieldWeight in 651, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.046875 = fieldNorm(doc=651)
        0.040010586 = weight(_text_:computer in 651) [ClassicSimilarity], result of:
          0.040010586 = score(doc=651,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.24226204 = fieldWeight in 651, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.046875 = fieldNorm(doc=651)
      0.33333334 = coord(2/6)
    
    Abstract
    Purpose - The generation of inverted indexes is one of the most computationally intensive activities for information retrieval systems: indexing large multi-gigabyte text databases can take many hours or even days to complete. We examine the generation of partitioned inverted files in order to speed up the process of indexing. Two types of index partitions are investigated: TermId and DocId. Design/methodology/approach - We use standard measures used in parallel computing such as speedup and efficiency to examine the computing results and also the space costs of our trial indexing experiments. Findings - The results from runs on both partitioning methods are compared and contrasted, concluding that DocId is the more efficient method. Practical implications - The practical implications are that the DocId partitioning method would in most circumstances be used for distributing inverted file data in a parallel computer, particularly if indexing speed is the primary consideration. Originality/value - The paper is of value to database administrators who manage large-scale text collections, and who need to use parallel computing to implement their text retrieval services.
  10. Furner, J.: ¬A unifying model of document relatedness for hybrid search engines (2003) 0.03
    0.025927832 = product of:
      0.077783495 = sum of:
        0.059414867 = weight(_text_:network in 2717) [ClassicSimilarity], result of:
          0.059414867 = score(doc=2717,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.29521978 = fieldWeight in 2717, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.046875 = fieldNorm(doc=2717)
        0.018368632 = product of:
          0.036737263 = sum of:
            0.036737263 = weight(_text_:22 in 2717) [ClassicSimilarity], result of:
              0.036737263 = score(doc=2717,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.23214069 = fieldWeight in 2717, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2717)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Previous work an search-engine design has indicated that information-seekers may benefit from being given the opportunity to exploit multiple sources of evidence of document relatedness. Few existing systems, however, give users more than minimal control over the selections that may be made among methods of exploitation. By applying the methods of "document network analysis" (DNA), a unifying, graph-theoretic model of content-, collaboration-, and context-based systems (CCC) may be developed in which the nature of the similarities between types of document relatedness and document ranking are clarified. The usefulness of the approach to system design suggested by this model may be tested by constructing and evaluating a prototype system (UCXtra) that allows searchers to maintain control over the multiple ways in which document collections may be ranked and re-ranked.
    Date
    11. 9.2004 17:32:22
  11. Dominich, S.: Mathematical foundations of information retrieval (2001) 0.02
    0.024352502 = product of:
      0.0730575 = sum of:
        0.057750307 = weight(_text_:computer in 1753) [ClassicSimilarity], result of:
          0.057750307 = score(doc=1753,freq=6.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.34967512 = fieldWeight in 1753, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1753)
        0.015307193 = product of:
          0.030614385 = sum of:
            0.030614385 = weight(_text_:22 in 1753) [ClassicSimilarity], result of:
              0.030614385 = score(doc=1753,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.19345059 = fieldWeight in 1753, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1753)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    This book offers a comprehensive and consistent mathematical approach to information retrieval (IR) without which no implementation is possible, and sheds an entirely new light upon the structure of IR models. It contains the descriptions of all IR models in a unified formal style and language, along with examples for each, thus offering a comprehensive overview of them. The book also creates mathematical foundations and a consistent mathematical theory (including all mathematical results achieved so far) of IR as a stand-alone mathematical discipline, which thus can be read and taught independently. Also, the book contains all necessary mathematical knowledge on which IR relies, to help the reader avoid searching different sources. The book will be of interest to computer or information scientists, librarians, mathematicians, undergraduate students and researchers whose work involves information retrieval.
    Date
    22. 3.2008 12:26:32
    LCSH
    Computer science / Mathematics
    Subject
    Computer science / Mathematics
  12. Chang, C.-H.; Hsu, C.-C.: Integrating query expansion and conceptual relevance feedback for personalized Web information retrieval (1998) 0.02
    0.02270303 = product of:
      0.06810909 = sum of:
        0.046679016 = weight(_text_:computer in 1319) [ClassicSimilarity], result of:
          0.046679016 = score(doc=1319,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.28263903 = fieldWeight in 1319, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1319)
        0.02143007 = product of:
          0.04286014 = sum of:
            0.04286014 = weight(_text_:22 in 1319) [ClassicSimilarity], result of:
              0.04286014 = score(doc=1319,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.2708308 = fieldWeight in 1319, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1319)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Date
    1. 8.1996 22:08:06
    Source
    Computer networks and ISDN systems. 30(1998) nos.1/7, S.621-623
  13. Joss, M.W.; Wszola, S.: ¬The engines that can : text search and retrieval software, their strategies, and vendors (1996) 0.02
    0.019583093 = product of:
      0.058749277 = sum of:
        0.040380646 = weight(_text_:services in 5123) [ClassicSimilarity], result of:
          0.040380646 = score(doc=5123,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.2433798 = fieldWeight in 5123, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.046875 = fieldNorm(doc=5123)
        0.018368632 = product of:
          0.036737263 = sum of:
            0.036737263 = weight(_text_:22 in 5123) [ClassicSimilarity], result of:
              0.036737263 = score(doc=5123,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.23214069 = fieldWeight in 5123, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5123)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Traces the development of text searching and retrieval software designed to cope with the increasing demands made by the storage and handling of large amounts of data, recorded on high data storage media, from CD-ROM to multi gigabyte storage media and online information services, with particular reference to the need to cope with graphics as well as conventional ASCII text. Includes details of: Boolean searching, fuzzy searching and matching; relevance ranking; proximity searching and improved strategies for dealing with text searching in very large databases. Concludes that the best searching tools for CD-ROM publishers are those optimized for searching and retrieval on CD-ROM. CD-ROM drives have relatively lower random seek times than hard discs and so the software most appropriate to the medium is that which can effectively arrange the indexes and text on the CD-ROM to avoid continuous random access searching. Lists and reviews a selection of software packages designed to achieve the sort of results required for rapid CD-ROM searching
    Date
    12. 9.1996 13:56:22
  14. Kelledy, F.; Smeaton, A.F.: Signature files and beyond (1996) 0.02
    0.01945974 = product of:
      0.058379218 = sum of:
        0.040010586 = weight(_text_:computer in 6973) [ClassicSimilarity], result of:
          0.040010586 = score(doc=6973,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.24226204 = fieldWeight in 6973, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.046875 = fieldNorm(doc=6973)
        0.018368632 = product of:
          0.036737263 = sum of:
            0.036737263 = weight(_text_:22 in 6973) [ClassicSimilarity], result of:
              0.036737263 = score(doc=6973,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.23214069 = fieldWeight in 6973, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=6973)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
  15. Klas, C.-P.; Fuhr, N.; Schaefer, A.: Evaluating strategic support for information access in the DAFFODIL system (2004) 0.02
    0.01945974 = product of:
      0.058379218 = sum of:
        0.040010586 = weight(_text_:computer in 2419) [ClassicSimilarity], result of:
          0.040010586 = score(doc=2419,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.24226204 = fieldWeight in 2419, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.046875 = fieldNorm(doc=2419)
        0.018368632 = product of:
          0.036737263 = sum of:
            0.036737263 = weight(_text_:22 in 2419) [ClassicSimilarity], result of:
              0.036737263 = score(doc=2419,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.23214069 = fieldWeight in 2419, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2419)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Date
    16.11.2008 16:22:48
    Series
    Lecture notes in computer science; vol.3232
  16. Rijsbergen, C.J. van: ¬A fast hierarchic clustering algorithm (1970) 0.02
    0.017782483 = product of:
      0.1066949 = sum of:
        0.1066949 = weight(_text_:computer in 3300) [ClassicSimilarity], result of:
          0.1066949 = score(doc=3300,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.6460321 = fieldWeight in 3300, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.125 = fieldNorm(doc=3300)
      0.16666667 = coord(1/6)
    
    Source
    Computer journal. 13(1970), S.324-326
  17. Kwok, K.L.: ¬A network approach to probabilistic information retrieval (1995) 0.02
    0.017151598 = product of:
      0.10290959 = sum of:
        0.10290959 = weight(_text_:network in 5696) [ClassicSimilarity], result of:
          0.10290959 = score(doc=5696,freq=6.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.51133573 = fieldWeight in 5696, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.046875 = fieldNorm(doc=5696)
      0.16666667 = coord(1/6)
    
    Abstract
    Shows how probabilistic information retrieval based on document components may be implemented as a feedforward (feedbackward) artificial neural network. The network supports adaptation of connection weights as well as the growing of new edges between queries and terms based on user relevance feedback data for training, and it reflects query modification and expansion in information retrieval. A learning rule is applied that can also be viewed as supporting sequential learning using a harmonic sequence learning rate. Experimental results with 4 standard small collections and a large Wall Street Journal collection show that small query expansion levels of about 30 terms can achieve most of the gains at the low-recall high-precision region, while larger expansion levels continue to provide gains at the high-recall low-precision region of a precision recall curve
  18. Ding, Y.; Yan, E.; Frazho, A.; Caverlee, J.: PageRank for ranking authors in co-citation networks (2009) 0.02
    0.017151598 = product of:
      0.10290959 = sum of:
        0.10290959 = weight(_text_:network in 3161) [ClassicSimilarity], result of:
          0.10290959 = score(doc=3161,freq=6.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.51133573 = fieldWeight in 3161, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.046875 = fieldNorm(doc=3161)
      0.16666667 = coord(1/6)
    
    Abstract
    This paper studies how varied damping factors in the PageRank algorithm influence the ranking of authors and proposes weighted PageRank algorithms. We selected the 108 most highly cited authors in the information retrieval (IR) area from the 1970s to 2008 to form the author co-citation network. We calculated the ranks of these 108 authors based on PageRank with the damping factor ranging from 0.05 to 0.95. In order to test the relationship between different measures, we compared PageRank and weighted PageRank results with the citation ranking, h-index, and centrality measures. We found that in our author co-citation network, citation rank is highly correlated with PageRank with different damping factors and also with different weighted PageRank algorithms; citation rank and PageRank are not significantly correlated with centrality measures; and h-index rank does not significantly correlate with centrality measures but does significantly correlate with other measures. The key factors that have impact on the PageRank of authors in the author co-citation network are being co-cited with important authors.
  19. Calegari, S.; Sanchez, E.: Object-fuzzy concept network : an enrichment of ontologies in semantic information retrieval (2008) 0.02
    0.016504131 = product of:
      0.09902479 = sum of:
        0.09902479 = weight(_text_:network in 2393) [ClassicSimilarity], result of:
          0.09902479 = score(doc=2393,freq=8.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.492033 = fieldWeight in 2393, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2393)
      0.16666667 = coord(1/6)
    
    Abstract
    This article shows how a fuzzy ontology-based approach can improve semantic documents retrieval. After formally defining a fuzzy ontology and a fuzzy knowledge base, a special type of new fuzzy relationship called (semantic) correlation, which links the concepts or entities in a fuzzy ontology, is discussed. These correlations, first assigned by experts, are updated after querying or when a document has been inserted into a database. Moreover, in order to define a dynamic knowledge of a domain adapting itself to the context, it is shown how to handle a tradeoff between the correct definition of an object, taken in the ontology structure, and the actual meaning assigned by individuals. The notion of a fuzzy concept network is extended, incorporating database objects so that entities and documents can similarly be represented in the network. Information retrieval (IR) algorithm, using an object-fuzzy concept network (O-FCN), is introduced and described. This algorithm allows us to derive a unique path among the entities involved in the query to obtain maxima semantic associations in the knowledge domain. Finally, the study has been validated by querying a database using fuzzy recall, fuzzy precision, and coefficient variant measures in the crisp and fuzzy cases.
  20. Finding anything in the billion page Web : are algorithms the key? (1999) 0.02
    0.015559672 = product of:
      0.09335803 = sum of:
        0.09335803 = weight(_text_:computer in 6248) [ClassicSimilarity], result of:
          0.09335803 = score(doc=6248,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.56527805 = fieldWeight in 6248, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.109375 = fieldNorm(doc=6248)
      0.16666667 = coord(1/6)
    
    Source
    Computer networks. 31(1999) no.11-16, S.1760-1761

Years

Languages

  • e 63
  • d 5
  • More… Less…

Types

  • a 59
  • m 6
  • el 3
  • r 1
  • s 1
  • More… Less…