Search (44 results, page 1 of 3)

  • × theme_ss:"Retrievalalgorithmen"
  1. Grossman, D.A.; Frieder, O.: Information retrieval : algorithms and heuristics (1998) 0.04
    0.037492394 = product of:
      0.07498479 = sum of:
        0.07498479 = product of:
          0.14996958 = sum of:
            0.14996958 = weight(_text_:o in 2182) [ClassicSimilarity], result of:
              0.14996958 = score(doc=2182,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.6652047 = fieldWeight in 2182, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2182)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  2. Wartik, S.; Fox, E.; Heath, L.; Chen, Q.-F.: Hashing algorithms (1992) 0.02
    0.024994928 = product of:
      0.049989857 = sum of:
        0.049989857 = product of:
          0.09997971 = sum of:
            0.09997971 = weight(_text_:o in 3510) [ClassicSimilarity], result of:
              0.09997971 = score(doc=3510,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.4434698 = fieldWeight in 3510, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3510)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Discusses hashing, an information storage and retrieval technique useful for implementing many of the other structures in this book. The concepts underlying hashing are presented, along with 2 implementation strategies. The chapter also contains an extensive discussion of perfect hashing, an important optimization in information retrieval, and an O(n) algorithm to find minimal perfect hash functions for a set of keys
  3. Vechtomova, O.; Karamuftuoglu, M.: Lexical cohesion and term proximity in document ranking (2008) 0.02
    0.024994928 = product of:
      0.049989857 = sum of:
        0.049989857 = product of:
          0.09997971 = sum of:
            0.09997971 = weight(_text_:o in 2101) [ClassicSimilarity], result of:
              0.09997971 = score(doc=2101,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.4434698 = fieldWeight in 2101, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2101)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  4. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.02
    0.024351954 = product of:
      0.04870391 = sum of:
        0.04870391 = product of:
          0.09740782 = sum of:
            0.09740782 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.09740782 = score(doc=402,freq=2.0), product of:
                0.15735255 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044934385 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  5. Cheng, C.-S.; Chung, C.-P.; Shann, J.J.-J.: Fast query evaluation through document identifier assignment for inverted file-based information retrieval systems (2006) 0.02
    0.022092605 = product of:
      0.04418521 = sum of:
        0.04418521 = product of:
          0.08837042 = sum of:
            0.08837042 = weight(_text_:o in 979) [ClassicSimilarity], result of:
              0.08837042 = score(doc=979,freq=4.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.39197564 = fieldWeight in 979, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=979)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Compressing an inverted file can greatly improve query performance of an information retrieval system (IRS) by reducing disk I/Os. We observe that a good document identifier assignment (DIA) can make the document identifiers in the posting lists more clustered, and result in better compression as well as shorter query processing time. In this paper, we tackle the NP-complete problem of finding an optimal DIA to minimize the average query processing time in an IRS when the probability distribution of query terms is given. We indicate that the greedy nearest neighbor (Greedy-NN) algorithm can provide excellent performance for this problem. However, the Greedy-NN algorithm is inappropriate if used in large-scale IRSs, due to its high complexity O(N2 × n), where N denotes the number of documents and n denotes the number of distinct terms. In real-world IRSs, the distribution of query terms is skewed. Based on this fact, we propose a fast O(N × n) heuristic, called partition-based document identifier assignment (PBDIA) algorithm, which can efficiently assign consecutive document identifiers to those documents containing frequently used query terms, and improve compression efficiency of the posting lists for those terms. This can result in reduced query processing time. The experimental results show that the PBDIA algorithm can yield a competitive performance versus the Greedy-NN for the DIA problem, and that this optimization problem has significant advantages for both long queries and parallel information retrieval (IR).
  6. Oberhauser, O.; Labner, J.: Relevance Ranking in Online-Katalogen : Informationsstand und Perspektiven (2003) 0.02
    0.021870563 = product of:
      0.043741126 = sum of:
        0.043741126 = product of:
          0.08748225 = sum of:
            0.08748225 = weight(_text_:o in 2188) [ClassicSimilarity], result of:
              0.08748225 = score(doc=2188,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.38803607 = fieldWeight in 2188, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2188)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  7. Vechtomova, O.; Karamuftuoglu, M.: Elicitation and use of relevance feedback information (2006) 0.02
    0.021870563 = product of:
      0.043741126 = sum of:
        0.043741126 = product of:
          0.08748225 = sum of:
            0.08748225 = weight(_text_:o in 966) [ClassicSimilarity], result of:
              0.08748225 = score(doc=966,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.38803607 = fieldWeight in 966, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=966)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  8. Smeaton, A.F.; Rijsbergen, C.J. van: ¬The retrieval effects of query expansion on a feedback document retrieval system (1983) 0.02
    0.02130796 = product of:
      0.04261592 = sum of:
        0.04261592 = product of:
          0.08523184 = sum of:
            0.08523184 = weight(_text_:22 in 2134) [ClassicSimilarity], result of:
              0.08523184 = score(doc=2134,freq=2.0), product of:
                0.15735255 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044934385 = queryNorm
                0.5416616 = fieldWeight in 2134, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2134)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    30. 3.2001 13:32:22
  9. Back, J.: ¬An evaluation of relevancy ranking techniques used by Internet search engines (2000) 0.02
    0.02130796 = product of:
      0.04261592 = sum of:
        0.04261592 = product of:
          0.08523184 = sum of:
            0.08523184 = weight(_text_:22 in 3445) [ClassicSimilarity], result of:
              0.08523184 = score(doc=3445,freq=2.0), product of:
                0.15735255 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044934385 = queryNorm
                0.5416616 = fieldWeight in 3445, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3445)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    25. 8.2005 17:42:22
  10. Beitzel, S.M.; Jensen, E.C.; Chowdhury, A.; Grossman, D.; Frieder, O; Goharian, N.: Fusion of effective retrieval strategies in the same information retrieval system (2004) 0.02
    0.018746197 = product of:
      0.037492394 = sum of:
        0.037492394 = product of:
          0.07498479 = sum of:
            0.07498479 = weight(_text_:o in 2502) [ClassicSimilarity], result of:
              0.07498479 = score(doc=2502,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.33260235 = fieldWeight in 2502, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2502)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  11. Herrera-Viedma, E.; Cordón, O.; Herrera, J.C.; Luqe, M.: ¬An IRS based on multi-granular lnguistic information (2003) 0.02
    0.018746197 = product of:
      0.037492394 = sum of:
        0.037492394 = product of:
          0.07498479 = sum of:
            0.07498479 = weight(_text_:o in 2740) [ClassicSimilarity], result of:
              0.07498479 = score(doc=2740,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.33260235 = fieldWeight in 2740, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2740)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  12. Oberhauser, O.: Relevance Ranking in den Online-Katalogen der "nächsten Generation" (2010) 0.02
    0.018746197 = product of:
      0.037492394 = sum of:
        0.037492394 = product of:
          0.07498479 = sum of:
            0.07498479 = weight(_text_:o in 4308) [ClassicSimilarity], result of:
              0.07498479 = score(doc=4308,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.33260235 = fieldWeight in 4308, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4308)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  13. Habernal, I.; Konopík, M.; Rohlík, O.: Question answering (2012) 0.02
    0.018746197 = product of:
      0.037492394 = sum of:
        0.037492394 = product of:
          0.07498479 = sum of:
            0.07498479 = weight(_text_:o in 101) [ClassicSimilarity], result of:
              0.07498479 = score(doc=101,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.33260235 = fieldWeight in 101, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.046875 = fieldNorm(doc=101)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  14. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.02
    0.018263966 = product of:
      0.03652793 = sum of:
        0.03652793 = product of:
          0.07305586 = sum of:
            0.07305586 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
              0.07305586 = score(doc=58,freq=2.0), product of:
                0.15735255 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044934385 = queryNorm
                0.46428138 = fieldWeight in 58, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=58)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:44
  15. Fuhr, N.: Rankingexperimente mit gewichteter Indexierung (1986) 0.02
    0.018263966 = product of:
      0.03652793 = sum of:
        0.03652793 = product of:
          0.07305586 = sum of:
            0.07305586 = weight(_text_:22 in 2051) [ClassicSimilarity], result of:
              0.07305586 = score(doc=2051,freq=2.0), product of:
                0.15735255 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044934385 = queryNorm
                0.46428138 = fieldWeight in 2051, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2051)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:56
  16. Chen, Z.; Fu, B.: On the complexity of Rocchio's similarity-based relevance feedback algorithm (2007) 0.02
    0.015621831 = product of:
      0.031243661 = sum of:
        0.031243661 = product of:
          0.062487323 = sum of:
            0.062487323 = weight(_text_:o in 578) [ClassicSimilarity], result of:
              0.062487323 = score(doc=578,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.27716863 = fieldWeight in 578, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=578)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Rocchio's similarity-based relevance feedback algorithm, one of the most important query reformation methods in information retrieval, is essentially an adaptive learning algorithm from examples in searching for documents represented by a linear classifier. Despite its popularity in various applications, there is little rigorous analysis of its learning complexity in literature. In this article, the authors prove for the first time that the learning complexity of Rocchio's algorithm is O(d + d**2(log d + log n)) over the discretized vector space {0, ... , n - 1 }**d when the inner product similarity measure is used. The upper bound on the learning complexity for searching for documents represented by a monotone linear classifier (q, 0) over {0, ... , n - 1 }d can be improved to, at most, 1 + 2k (n - 1) (log d + log(n - 1)), where k is the number of nonzero components in q. Several lower bounds on the learning complexity are also obtained for Rocchio's algorithm. For example, the authors prove that Rocchio's algorithm has a lower bound Omega((d über 2)log n) on its learning complexity over the Boolean vector space {0,1}**d.
  17. Urbain, J.; Goharian, N.; Frieder, O.: Probabilistic passage models for semantic search of genomics literature (2008) 0.02
    0.015621831 = product of:
      0.031243661 = sum of:
        0.031243661 = product of:
          0.062487323 = sum of:
            0.062487323 = weight(_text_:o in 2380) [ClassicSimilarity], result of:
              0.062487323 = score(doc=2380,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.27716863 = fieldWeight in 2380, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2380)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  18. Calegari, S.; Sanchez, E.: Object-fuzzy concept network : an enrichment of ontologies in semantic information retrieval (2008) 0.02
    0.015621831 = product of:
      0.031243661 = sum of:
        0.031243661 = product of:
          0.062487323 = sum of:
            0.062487323 = weight(_text_:o in 2393) [ClassicSimilarity], result of:
              0.062487323 = score(doc=2393,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.27716863 = fieldWeight in 2393, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2393)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This article shows how a fuzzy ontology-based approach can improve semantic documents retrieval. After formally defining a fuzzy ontology and a fuzzy knowledge base, a special type of new fuzzy relationship called (semantic) correlation, which links the concepts or entities in a fuzzy ontology, is discussed. These correlations, first assigned by experts, are updated after querying or when a document has been inserted into a database. Moreover, in order to define a dynamic knowledge of a domain adapting itself to the context, it is shown how to handle a tradeoff between the correct definition of an object, taken in the ontology structure, and the actual meaning assigned by individuals. The notion of a fuzzy concept network is extended, incorporating database objects so that entities and documents can similarly be represented in the network. Information retrieval (IR) algorithm, using an object-fuzzy concept network (O-FCN), is introduced and described. This algorithm allows us to derive a unique path among the entities involved in the query to obtain maxima semantic associations in the knowledge domain. Finally, the study has been validated by querying a database using fuzzy recall, fuzzy precision, and coefficient variant measures in the crisp and fuzzy cases.
  19. Jacucci, G.; Barral, O.; Daee, P.; Wenzel, M.; Serim, B.; Ruotsalo, T.; Pluchino, P.; Freeman, J.; Gamberini, L.; Kaski, S.; Blankertz, B.: Integrating neurophysiologic relevance feedback in intent modeling for information retrieval (2019) 0.02
    0.015621831 = product of:
      0.031243661 = sum of:
        0.031243661 = product of:
          0.062487323 = sum of:
            0.062487323 = weight(_text_:o in 5356) [ClassicSimilarity], result of:
              0.062487323 = score(doc=5356,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.27716863 = fieldWeight in 5356, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5356)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  20. Fuhr, N.: Modelle im Information Retrieval (2023) 0.02
    0.015621831 = product of:
      0.031243661 = sum of:
        0.031243661 = product of:
          0.062487323 = sum of:
            0.062487323 = weight(_text_:o in 800) [ClassicSimilarity], result of:
              0.062487323 = score(doc=800,freq=2.0), product of:
                0.22544876 = queryWeight, product of:
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.044934385 = queryNorm
                0.27716863 = fieldWeight in 800, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.017288 = idf(docFreq=795, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=800)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Information-Retrieval-Modelle -(IR-Modelle) spezifizieren, wie zu einer gegebenen Anfrage die Antwortdokumente aus einer Dokumentenkollektion bestimmt werden. Ausgangsbasis jedes Modells sind dabei zunächst bestimmte Annahmen über die Wissensrepräsentation (s. Teil B Methoden und Systeme der Inhaltserschließung) von Fragen und Dokumenten. Hier bezeichnen wir die Elemente dieser Repräsentationen als Terme, wobei es aus der Sicht des Modells egal ist, wie diese Terme aus dem Dokument (und analog aus der von Benutzenden eingegebenen Anfrage) abgeleitet werden: Bei Texten werden hierzu häufig computerlinguistische Methoden eingesetzt, aber auch komplexere automatische oder manuelle Erschließungsverfahren können zur Anwendung kommen. Repräsentationen besitzen ferner eine bestimmte Struktur. Ein Dokument wird meist als Menge oder Multimenge von Termen aufgefasst, wobei im zweiten Fall das Mehrfachvorkommen berücksichtigt wird. Diese Dokumentrepräsentation wird wiederum auf eine sogenannte Dokumentbeschreibung abgebildet, in der die einzelnen Terme gewichtet sein können. Im Folgenden unterscheiden wir nur zwischen ungewichteter (Gewicht eines Terms ist entweder 0 oder 1) und gewichteter Indexierung (das Gewicht ist eine nichtnegative reelle Zahl). Analog dazu gibt es eine Fragerepräsentation; legt man eine natürlichsprachige Anfrage zugrunde, so kann man die o. g. Verfahren für Dokumenttexte anwenden. Alternativ werden auch grafische oder formale Anfragesprachen verwendet, wobei aus Sicht der Modelle insbesondere deren logische Struktur (etwa beim Booleschen Retrieval) relevant ist. Die Fragerepräsentation wird dann in eine Fragebeschreibung überführt.

Years

Languages

  • e 37
  • d 7

Types

  • a 41
  • m 2
  • r 1
  • More… Less…