Search (107 results, page 1 of 6)

  • × year_i:[1990 TO 2000}
  • × theme_ss:"Retrievalalgorithmen"
  1. Wilhelmy, A.: Phonetische Ähnlichkeitssuche in Datenbanken (1991) 0.02
    0.02333379 = product of:
      0.10000195 = sum of:
        0.015706176 = weight(_text_:und in 5684) [ClassicSimilarity], result of:
          0.015706176 = score(doc=5684,freq=10.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.328536 = fieldWeight in 5684, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=5684)
        0.0052914224 = weight(_text_:in in 5684) [ClassicSimilarity], result of:
          0.0052914224 = score(doc=5684,freq=8.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.18034597 = fieldWeight in 5684, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=5684)
        0.015706176 = weight(_text_:und in 5684) [ClassicSimilarity], result of:
          0.015706176 = score(doc=5684,freq=10.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.328536 = fieldWeight in 5684, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=5684)
        0.020302603 = weight(_text_:bibliotheken in 5684) [ClassicSimilarity], result of:
          0.020302603 = score(doc=5684,freq=2.0), product of:
            0.08127756 = queryWeight, product of:
              3.768121 = idf(docFreq=2775, maxDocs=44218)
              0.021569785 = queryNorm
            0.24979347 = fieldWeight in 5684, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.768121 = idf(docFreq=2775, maxDocs=44218)
              0.046875 = fieldNorm(doc=5684)
        0.020302603 = weight(_text_:bibliotheken in 5684) [ClassicSimilarity], result of:
          0.020302603 = score(doc=5684,freq=2.0), product of:
            0.08127756 = queryWeight, product of:
              3.768121 = idf(docFreq=2775, maxDocs=44218)
              0.021569785 = queryNorm
            0.24979347 = fieldWeight in 5684, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.768121 = idf(docFreq=2775, maxDocs=44218)
              0.046875 = fieldNorm(doc=5684)
        0.020302603 = weight(_text_:bibliotheken in 5684) [ClassicSimilarity], result of:
          0.020302603 = score(doc=5684,freq=2.0), product of:
            0.08127756 = queryWeight, product of:
              3.768121 = idf(docFreq=2775, maxDocs=44218)
              0.021569785 = queryNorm
            0.24979347 = fieldWeight in 5684, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.768121 = idf(docFreq=2775, maxDocs=44218)
              0.046875 = fieldNorm(doc=5684)
        0.002390375 = weight(_text_:s in 5684) [ClassicSimilarity], result of:
          0.002390375 = score(doc=5684,freq=4.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.101928525 = fieldWeight in 5684, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.046875 = fieldNorm(doc=5684)
      0.23333333 = coord(7/30)
    
    Abstract
    In dialoggesteuerten Systemen zur Informationswiedergewinnung (Information Retrieval Systems, IRS) kann man - vergröbernd - das Wechselspiel zwischen Mensch und Computer als iterativen Prozess zur Erhöhung von Genauigkeit (Precision) auf der einen und Vollständigkeit (Recall) der Nachweise auf der anderen Seite verstehen. Vorgestellt wird ein maschinell anwendbares Verfahren, das auf phonologische Untersuchungen des Sprachwissenschaftlers Nikolaj S. Trubetzkoy (1890-1938) zurückgeht. In den Grundzügen kann es erheblich zur Verbesserung der Nachweisvollständigkeit beitragen. Dadurch, daß es die 'Ähnlichkeitsumgebungen' von Suchbegriffen in die Recherche mit einbezieht, zeigt es sich vor allem für Systeme mit koordinativer maschineller Indexierung als vorteilhaft. Bei alphabetischen Begriffen erweist sich die Einführung eines solchen zunächst nur auf den Benutzer hin orientierten Verfahrens auch aus technischer Sicht als günstig, da damit die Anzahl der Zugriffe bei den Suchvorgängen auch für große Datenvolumina niedrig gehalten werden kann
    Pages
    S.329-338
    Source
    Bibliotheken mit und ohne Grenzen: Informationsgesellschaft und Bibliothek. Der österreichische Bibliothekartag 1990, Bregenz, 4.-8.9.1990, Vorträge und Kommissionssitzungen
  2. Hüther, H.: Selix im DFG-Projekt Kascade (1998) 0.02
    0.016393315 = product of:
      0.12294986 = sum of:
        0.011706693 = weight(_text_:und in 5151) [ClassicSimilarity], result of:
          0.011706693 = score(doc=5151,freq=2.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.24487628 = fieldWeight in 5151, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=5151)
        0.096719384 = weight(_text_:informationswissenschaft in 5151) [ClassicSimilarity], result of:
          0.096719384 = score(doc=5151,freq=8.0), product of:
            0.09716552 = queryWeight, product of:
              4.504705 = idf(docFreq=1328, maxDocs=44218)
              0.021569785 = queryNorm
            0.99540854 = fieldWeight in 5151, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              4.504705 = idf(docFreq=1328, maxDocs=44218)
              0.078125 = fieldNorm(doc=5151)
        0.011706693 = weight(_text_:und in 5151) [ClassicSimilarity], result of:
          0.011706693 = score(doc=5151,freq=2.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.24487628 = fieldWeight in 5151, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=5151)
        0.0028170836 = weight(_text_:s in 5151) [ClassicSimilarity], result of:
          0.0028170836 = score(doc=5151,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.120123915 = fieldWeight in 5151, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.078125 = fieldNorm(doc=5151)
      0.13333334 = coord(4/30)
    
    Pages
    S.397-403
    Series
    Schriften zur Informationswissenschaft; Bd.34
    Source
    Knowledge Management und Kommunikationssysteme: Proceedings des 6. Internationalen Symposiums für Informationswissenschaft (ISI '98) Prag, 3.-7. November 1998 / Hochschulverband für Informationswissenschaft (HI) e.V. Konstanz ; Fachrichtung Informationswissenschaft der Universität des Saarlandes, Saarbrücken. Hrsg.: Harald H. Zimmermann u. Volker Schramm
  3. Pfeifer, U.; Pennekamp, S.: Incremental processing of vague queries in interactive retrieval systems (1997) 0.01
    0.011119273 = product of:
      0.066715635 = sum of:
        0.009365354 = weight(_text_:und in 735) [ClassicSimilarity], result of:
          0.009365354 = score(doc=735,freq=2.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.19590102 = fieldWeight in 735, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=735)
        0.038687754 = weight(_text_:informationswissenschaft in 735) [ClassicSimilarity], result of:
          0.038687754 = score(doc=735,freq=2.0), product of:
            0.09716552 = queryWeight, product of:
              4.504705 = idf(docFreq=1328, maxDocs=44218)
              0.021569785 = queryNorm
            0.3981634 = fieldWeight in 735, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.504705 = idf(docFreq=1328, maxDocs=44218)
              0.0625 = fieldNorm(doc=735)
        0.006110009 = weight(_text_:in in 735) [ClassicSimilarity], result of:
          0.006110009 = score(doc=735,freq=6.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.2082456 = fieldWeight in 735, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=735)
        0.009365354 = weight(_text_:und in 735) [ClassicSimilarity], result of:
          0.009365354 = score(doc=735,freq=2.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.19590102 = fieldWeight in 735, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=735)
        0.0031871665 = weight(_text_:s in 735) [ClassicSimilarity], result of:
          0.0031871665 = score(doc=735,freq=4.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.1359047 = fieldWeight in 735, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.0625 = fieldNorm(doc=735)
      0.16666667 = coord(5/30)
    
    Abstract
    The application of information retrieval techniques in interactive environments requires systems capable of effeciently processing vague queries. To reach reasonable response times, new data structures and algorithms have to be developed. In this paper we describe an approach taking advantage of the conditions of interactive usage and special access paths. To have a reference we investigate text queries and compared our algorithms to the well known 'Buckley/Lewit' algorithm. We achieved significant improvements for the response times
    Pages
    S.223-236
    Series
    Schriften zur Informationswissenschaft; Bd.30
    Source
    Hypertext - Information Retrieval - Multimedia '97: Theorien, Modelle und Implementierungen integrierter elektronischer Informationssysteme. Proceedings HIM '97. Hrsg.: N. Fuhr u.a
  4. Fuhr, N.: Zur Überwindung der Diskrepanz zwischen Retrievalforschung und -praxis (1990) 0.01
    0.0063552563 = product of:
      0.04766442 = sum of:
        0.020941569 = weight(_text_:und in 6625) [ClassicSimilarity], result of:
          0.020941569 = score(doc=6625,freq=10.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.438048 = fieldWeight in 6625, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=6625)
        0.003527615 = weight(_text_:in in 6625) [ClassicSimilarity], result of:
          0.003527615 = score(doc=6625,freq=2.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.120230645 = fieldWeight in 6625, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=6625)
        0.020941569 = weight(_text_:und in 6625) [ClassicSimilarity], result of:
          0.020941569 = score(doc=6625,freq=10.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.438048 = fieldWeight in 6625, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=6625)
        0.002253667 = weight(_text_:s in 6625) [ClassicSimilarity], result of:
          0.002253667 = score(doc=6625,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.09609913 = fieldWeight in 6625, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.0625 = fieldNorm(doc=6625)
      0.13333334 = coord(4/30)
    
    Abstract
    In diesem Beitrag werden einige Forschungsergebnisse des Information Retrieval vorgestellt, die unmittelbar zur Verbesserung der Retrievalqualität für bereits existierende Datenbanken eingesetzt werden können: Linguistische Algorithmen zur Grund- und Stammformreduktion unterstützen die Suche nach Flexions- und Derivationsformen von Suchtermen. Rankingalgorithmen, die Frage- und Dokumentterme gewichten, führen zu signifikant besseren Retrievalergebnissen als beim Booleschen Retrieval. Durch Relevance Feedback können die Retrievalqualität weiter gesteigert und außerdem der Benutzer bei der sukzessiven Modifikation seiner Frageformulierung unterstützt werden. Es wird eine benutzerfreundliche Bedienungsoberfläche für ein System vorgestellt, das auf diesen Konzepten basiert.
    Source
    Nachrichten für Dokumentation. 41(1990), S.3-7
  5. Ziegler, B.: ESS: ein schneller Algorithmus zur Mustersuche in Zeichenfolgen (1996) 0.01
    0.0057194647 = product of:
      0.042895984 = sum of:
        0.01638937 = weight(_text_:und in 7543) [ClassicSimilarity], result of:
          0.01638937 = score(doc=7543,freq=2.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.34282678 = fieldWeight in 7543, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.109375 = fieldNorm(doc=7543)
        0.0061733257 = weight(_text_:in in 7543) [ClassicSimilarity], result of:
          0.0061733257 = score(doc=7543,freq=2.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.21040362 = fieldWeight in 7543, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.109375 = fieldNorm(doc=7543)
        0.01638937 = weight(_text_:und in 7543) [ClassicSimilarity], result of:
          0.01638937 = score(doc=7543,freq=2.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.34282678 = fieldWeight in 7543, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.109375 = fieldNorm(doc=7543)
        0.003943917 = weight(_text_:s in 7543) [ClassicSimilarity], result of:
          0.003943917 = score(doc=7543,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.16817348 = fieldWeight in 7543, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.109375 = fieldNorm(doc=7543)
      0.13333334 = coord(4/30)
    
    Source
    Informatik: Forschung und Entwicklung. 11(1996) no.2, S.69-83
  6. Koopman, R.: ¬Ein OPAC mit Gewichtungsalgorithmen : Der PICA Micro OPC (1996) 0.00
    0.0035928607 = product of:
      0.035928607 = sum of:
        0.016555762 = weight(_text_:und in 4114) [ClassicSimilarity], result of:
          0.016555762 = score(doc=4114,freq=4.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.34630734 = fieldWeight in 4114, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=4114)
        0.016555762 = weight(_text_:und in 4114) [ClassicSimilarity], result of:
          0.016555762 = score(doc=4114,freq=4.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.34630734 = fieldWeight in 4114, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=4114)
        0.0028170836 = weight(_text_:s in 4114) [ClassicSimilarity], result of:
          0.0028170836 = score(doc=4114,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.120123915 = fieldWeight in 4114, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.078125 = fieldNorm(doc=4114)
      0.1 = coord(3/30)
    
    Imprint
    Düsseldorf : Universitäts- und Landesbibliothek
    Pages
    S.95-105
    Series
    Schriften der Universitäts- und Landesbibliothek Düsseldorf; Bd.25
  7. Chakrabarti, S.; Dom, B.; Kumar, S.R.; Raghavan, P.; Rajagopalan, S.; Tomkins, A.; Kleinberg, J.M.; Gibson, D.: Neue Pfade durch den Internet-Dschungel : Die zweite Generation von Web-Suchmaschinen (1999) 0.00
    0.0034882387 = product of:
      0.026161788 = sum of:
        0.009365354 = weight(_text_:und in 3) [ClassicSimilarity], result of:
          0.009365354 = score(doc=3,freq=2.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.19590102 = fieldWeight in 3, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=3)
        0.003527615 = weight(_text_:in in 3) [ClassicSimilarity], result of:
          0.003527615 = score(doc=3,freq=2.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.120230645 = fieldWeight in 3, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=3)
        0.009365354 = weight(_text_:und in 3) [ClassicSimilarity], result of:
          0.009365354 = score(doc=3,freq=2.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.19590102 = fieldWeight in 3, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=3)
        0.003903466 = weight(_text_:s in 3) [ClassicSimilarity], result of:
          0.003903466 = score(doc=3,freq=6.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.1664486 = fieldWeight in 3, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.0625 = fieldNorm(doc=3)
      0.13333334 = coord(4/30)
    
    Abstract
    Die im WWW verfügbare Datenmenge wächst mit atemberaubender Geschwindigkeit; entsprechend schwieriger wird es, relevante Informationen zu finden. ein neues Analyseverfahren stellt nahezu automatische Abhilfe in Aussicht
    Content
    Ausnutzen der Hyperlinks für verbesserte Such- und Findeverfahren; Darstellung des HITS-Algorithmus
    Source
    Spektrum der Wissenschaft. 1999, H.8, S.44-49
  8. Berry, M.W.; Browne, M.: Understanding search engines : mathematical modeling and text retrieval (1999) 0.00
    0.0033731693 = product of:
      0.025298769 = sum of:
        0.009933459 = weight(_text_:und in 5777) [ClassicSimilarity], result of:
          0.009933459 = score(doc=5777,freq=4.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.20778441 = fieldWeight in 5777, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=5777)
        0.003741601 = weight(_text_:in in 5777) [ClassicSimilarity], result of:
          0.003741601 = score(doc=5777,freq=4.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.12752387 = fieldWeight in 5777, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=5777)
        0.009933459 = weight(_text_:und in 5777) [ClassicSimilarity], result of:
          0.009933459 = score(doc=5777,freq=4.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.20778441 = fieldWeight in 5777, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=5777)
        0.0016902501 = weight(_text_:s in 5777) [ClassicSimilarity], result of:
          0.0016902501 = score(doc=5777,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.072074346 = fieldWeight in 5777, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.046875 = fieldNorm(doc=5777)
      0.13333334 = coord(4/30)
    
    Abstract
    This book discusses many of the key design issues for building search engines and emphazises the important role that applied mathematics can play in improving information retrieval. The authors discuss not only important data structures, algorithms, and software but also user-centered issues such as interfaces, manual indexing, and document preparation. They also present some of the current problems in information retrieval that many not be familiar to applied mathematicians and computer scientists and some of the driving computational methods (SVD, SDD) for automated conceptual indexing
    Classification
    ST 230 [Informatik # Monographien # Software und -entwicklung # Software allgemein, (Einführung, Lehrbücher, Methoden der Programmierung) Software engineering, Programmentwicklungssysteme, Softwarewerkzeuge]
    Pages
    XIII, 116 S
    RVK
    ST 230 [Informatik # Monographien # Software und -entwicklung # Software allgemein, (Einführung, Lehrbücher, Methoden der Programmierung) Software engineering, Programmentwicklungssysteme, Softwarewerkzeuge]
  9. Kleinberg, J.M.: Authoritative sources in a hyperlinked environment (1998) 0.00
    0.0028872362 = product of:
      0.02165427 = sum of:
        0.0070240153 = weight(_text_:und in 5) [ClassicSimilarity], result of:
          0.0070240153 = score(doc=5,freq=2.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.14692576 = fieldWeight in 5, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=5)
        0.005915991 = weight(_text_:in in 5) [ClassicSimilarity], result of:
          0.005915991 = score(doc=5,freq=10.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.20163295 = fieldWeight in 5, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=5)
        0.0070240153 = weight(_text_:und in 5) [ClassicSimilarity], result of:
          0.0070240153 = score(doc=5,freq=2.0), product of:
            0.04780656 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021569785 = queryNorm
            0.14692576 = fieldWeight in 5, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=5)
        0.0016902501 = weight(_text_:s in 5) [ClassicSimilarity], result of:
          0.0016902501 = score(doc=5,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.072074346 = fieldWeight in 5, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.046875 = fieldNorm(doc=5)
      0.13333334 = coord(4/30)
    
    Abstract
    The network structure of a hyperlinked environment can be a rich source of information about the content of the environment, provided we have effective means for understanding it. We develop a set of algorithmic tools for extracting information from the link structures of such environments, and report on experiments that demonstrate their effectiveness in a variety of contexts on the World Wide Web. The central issue we address within our framework is the distillation of broad search topics, through the discovery of "authoritative" information sources on such topics. We propose and test an algorithmic formulation of the notion of authority, based on the relationship between a set of relevant authoritative pages and the set of "hub pages" that join them together in the link structure. Our formulation has connections to the eigenvectors of certain matrices associated with the link graph; these connections in turn motivate additional heuristics for link-based analysis.
    Content
    Vorversionen auch in: Proceedings of the ACM-SIAM Symposium on Discrete Algorithms, 1998, und als IBM Research Report RJ 10076, May 1997.
    Source
    Journal of the Association for Computing Machinery. 46(1998) no.5, S.604-632
  10. Schamber, L.; Bateman, J.: Relevance criteria uses and importance : progress in development of a measurement scale (1999) 0.00
    0.0018254882 = product of:
      0.018254882 = sum of:
        0.005915991 = weight(_text_:in in 6691) [ClassicSimilarity], result of:
          0.005915991 = score(doc=6691,freq=10.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.20163295 = fieldWeight in 6691, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=6691)
        0.010648641 = product of:
          0.03194592 = sum of:
            0.03194592 = weight(_text_:l in 6691) [ClassicSimilarity], result of:
              0.03194592 = score(doc=6691,freq=4.0), product of:
                0.0857324 = queryWeight, product of:
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.021569785 = queryNorm
                0.37262368 = fieldWeight in 6691, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.046875 = fieldNorm(doc=6691)
          0.33333334 = coord(1/3)
        0.0016902501 = weight(_text_:s in 6691) [ClassicSimilarity], result of:
          0.0016902501 = score(doc=6691,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.072074346 = fieldWeight in 6691, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.046875 = fieldNorm(doc=6691)
      0.1 = coord(3/30)
    
    Abstract
    The criteria employed by end-users in making relevance judgments can be powerful and useful indicators of the values users ascribe to a variety of factors in their information seeking and use situations. This paper describes intermediate results in a long-term project intended to develop a measurement scale based on users' relevance criteria. The five tests that are reported here have involved 350 users in an effort to progressively refine and validate the scale content. The range of research questions and types of users and information environments have gradually been expanded to assess the adaptability and transferability of the instrument. The instrument provides quantitative data, notably criterion importance ratings that can be analyzed using several techniques. The substantive findings confirm those of previous studies on relevance evaluation behavior
    Pages
    S.381-389
    Source
    Knowledge: creation, organization and use. Proceedings of the 62nd Annual Meeting of the American Society for Information Science, 31.10.-4.11.1999. Ed.: L. Woods
  11. Wartik, S.; Fox, E.; Heath, L.; Chen, Q.-F.: Hashing algorithms (1992) 0.00
    0.0018215602 = product of:
      0.018215602 = sum of:
        0.004988801 = weight(_text_:in in 3510) [ClassicSimilarity], result of:
          0.004988801 = score(doc=3510,freq=4.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.17003182 = fieldWeight in 3510, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=3510)
        0.010039635 = product of:
          0.030118903 = sum of:
            0.030118903 = weight(_text_:l in 3510) [ClassicSimilarity], result of:
              0.030118903 = score(doc=3510,freq=2.0), product of:
                0.0857324 = queryWeight, product of:
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.021569785 = queryNorm
                0.35131297 = fieldWeight in 3510, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3510)
          0.33333334 = coord(1/3)
        0.0031871665 = weight(_text_:s in 3510) [ClassicSimilarity], result of:
          0.0031871665 = score(doc=3510,freq=4.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.1359047 = fieldWeight in 3510, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.0625 = fieldNorm(doc=3510)
      0.1 = coord(3/30)
    
    Abstract
    Discusses hashing, an information storage and retrieval technique useful for implementing many of the other structures in this book. The concepts underlying hashing are presented, along with 2 implementation strategies. The chapter also contains an extensive discussion of perfect hashing, an important optimization in information retrieval, and an O(n) algorithm to find minimal perfect hash functions for a set of keys
    Pages
    S.293-362
  12. Faloutsos, C.: Signature files (1992) 0.00
    0.0017470915 = product of:
      0.017470915 = sum of:
        0.003527615 = weight(_text_:in in 3499) [ClassicSimilarity], result of:
          0.003527615 = score(doc=3499,freq=2.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.120230645 = fieldWeight in 3499, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=3499)
        0.002253667 = weight(_text_:s in 3499) [ClassicSimilarity], result of:
          0.002253667 = score(doc=3499,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.09609913 = fieldWeight in 3499, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.0625 = fieldNorm(doc=3499)
        0.011689632 = product of:
          0.023379264 = sum of:
            0.023379264 = weight(_text_:22 in 3499) [ClassicSimilarity], result of:
              0.023379264 = score(doc=3499,freq=2.0), product of:
                0.07553371 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.021569785 = queryNorm
                0.30952093 = fieldWeight in 3499, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3499)
          0.5 = coord(1/2)
      0.1 = coord(3/30)
    
    Abstract
    Presents a survey and discussion on signature-based text retrieval methods. It describes the main idea behind the signature approach and its advantages over other text retrieval methods, it provides a classification of the signature methods that have appeared in the literature, it describes the main representatives of each class, together with the relative advantages and drawbacks, and it gives a list of applications as well as commercial or university prototypes that use the signature approach
    Date
    7. 5.1999 15:22:48
    Pages
    S.44-65
  13. Longshu, L.; Xia, Z.: On an aproximate fuzzy information retrieval agent (1998) 0.00
    0.0015820917 = product of:
      0.015820917 = sum of:
        0.003527615 = weight(_text_:in in 3294) [ClassicSimilarity], result of:
          0.003527615 = score(doc=3294,freq=2.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.120230645 = fieldWeight in 3294, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=3294)
        0.010039635 = product of:
          0.030118903 = sum of:
            0.030118903 = weight(_text_:l in 3294) [ClassicSimilarity], result of:
              0.030118903 = score(doc=3294,freq=2.0), product of:
                0.0857324 = queryWeight, product of:
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.021569785 = queryNorm
                0.35131297 = fieldWeight in 3294, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3294)
          0.33333334 = coord(1/3)
        0.002253667 = weight(_text_:s in 3294) [ClassicSimilarity], result of:
          0.002253667 = score(doc=3294,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.09609913 = fieldWeight in 3294, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.0625 = fieldNorm(doc=3294)
      0.1 = coord(3/30)
    
    Footnote
    [In Chinesisch]
    Source
    Journal of the China Society for Scientific and Technical Information. 17(1998) no.3, S.180-184
  14. Chang, C.-H.; Hsu, C.-C.: Integrating query expansion and conceptual relevance feedback for personalized Web information retrieval (1998) 0.00
    0.0015287049 = product of:
      0.015287049 = sum of:
        0.0030866629 = weight(_text_:in in 1319) [ClassicSimilarity], result of:
          0.0030866629 = score(doc=1319,freq=2.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.10520181 = fieldWeight in 1319, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1319)
        0.0019719584 = weight(_text_:s in 1319) [ClassicSimilarity], result of:
          0.0019719584 = score(doc=1319,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.08408674 = fieldWeight in 1319, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1319)
        0.010228428 = product of:
          0.020456856 = sum of:
            0.020456856 = weight(_text_:22 in 1319) [ClassicSimilarity], result of:
              0.020456856 = score(doc=1319,freq=2.0), product of:
                0.07553371 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.021569785 = queryNorm
                0.2708308 = fieldWeight in 1319, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1319)
          0.5 = coord(1/2)
      0.1 = coord(3/30)
    
    Date
    1. 8.1996 22:08:06
    Source
    Computer networks and ISDN systems. 30(1998) nos.1/7, S.621-623
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  15. Efthimiadis, E.N.: User choices : a new yardstick for the evaluation of ranking algorithms for interactive query expansion (1995) 0.00
    0.0014950562 = product of:
      0.014950562 = sum of:
        0.006236001 = weight(_text_:in in 5697) [ClassicSimilarity], result of:
          0.006236001 = score(doc=5697,freq=16.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.21253976 = fieldWeight in 5697, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5697)
        0.0014085418 = weight(_text_:s in 5697) [ClassicSimilarity], result of:
          0.0014085418 = score(doc=5697,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.060061958 = fieldWeight in 5697, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5697)
        0.0073060202 = product of:
          0.0146120405 = sum of:
            0.0146120405 = weight(_text_:22 in 5697) [ClassicSimilarity], result of:
              0.0146120405 = score(doc=5697,freq=2.0), product of:
                0.07553371 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.021569785 = queryNorm
                0.19345059 = fieldWeight in 5697, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5697)
          0.5 = coord(1/2)
      0.1 = coord(3/30)
    
    Abstract
    The performance of 8 ranking algorithms was evaluated with respect to their effectiveness in ranking terms for query expansion. The evaluation was conducted within an investigation of interactive query expansion and relevance feedback in a real operational environment. Focuses on the identification of algorithms that most effectively take cognizance of user preferences. user choices (i.e. the terms selected by the searchers for the query expansion search) provided the yardstick for the evaluation of the 8 ranking algorithms. This methodology introduces a user oriented approach in evaluating ranking algorithms for query expansion in contrast to the standard, system oriented approaches. Similarities in the performance of the 8 algorithms and the ways these algorithms rank terms were the main focus of this evaluation. The findings demonstrate that the r-lohi, wpq, enim, and porter algorithms have similar performance in bringing good terms to the top of a ranked list of terms for query expansion. However, further evaluation of the algorithms in different (e.g. full text) environments is needed before these results can be generalized beyond the context of the present study
    Date
    22. 2.1996 13:14:10
    Source
    Information processing and management. 31(1995) no.4, S.605-620
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  16. Joss, M.W.; Wszola, S.: ¬The engines that can : text search and retrieval software, their strategies, and vendors (1996) 0.00
    0.0014340535 = product of:
      0.014340535 = sum of:
        0.0026457112 = weight(_text_:in in 5123) [ClassicSimilarity], result of:
          0.0026457112 = score(doc=5123,freq=2.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.09017298 = fieldWeight in 5123, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=5123)
        0.0029275995 = weight(_text_:s in 5123) [ClassicSimilarity], result of:
          0.0029275995 = score(doc=5123,freq=6.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.124836445 = fieldWeight in 5123, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.046875 = fieldNorm(doc=5123)
        0.008767224 = product of:
          0.017534448 = sum of:
            0.017534448 = weight(_text_:22 in 5123) [ClassicSimilarity], result of:
              0.017534448 = score(doc=5123,freq=2.0), product of:
                0.07553371 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.021569785 = queryNorm
                0.23214069 = fieldWeight in 5123, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5123)
          0.5 = coord(1/2)
      0.1 = coord(3/30)
    
    Abstract
    Traces the development of text searching and retrieval software designed to cope with the increasing demands made by the storage and handling of large amounts of data, recorded on high data storage media, from CD-ROM to multi gigabyte storage media and online information services, with particular reference to the need to cope with graphics as well as conventional ASCII text. Includes details of: Boolean searching, fuzzy searching and matching; relevance ranking; proximity searching and improved strategies for dealing with text searching in very large databases. Concludes that the best searching tools for CD-ROM publishers are those optimized for searching and retrieval on CD-ROM. CD-ROM drives have relatively lower random seek times than hard discs and so the software most appropriate to the medium is that which can effectively arrange the indexes and text on the CD-ROM to avoid continuous random access searching. Lists and reviews a selection of software packages designed to achieve the sort of results required for rapid CD-ROM searching
    Date
    12. 9.1996 13:56:22
    Source
    CD-ROM professional. 9(1996) no.6, S.30+(14 S.)
  17. Information retrieval : data structures and algorithms (1992) 0.00
    0.00141345 = product of:
      0.0141345 = sum of:
        0.004409519 = weight(_text_:in in 3495) [ClassicSimilarity], result of:
          0.004409519 = score(doc=3495,freq=8.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.15028831 = fieldWeight in 3495, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3495)
        0.006274772 = product of:
          0.018824315 = sum of:
            0.018824315 = weight(_text_:l in 3495) [ClassicSimilarity], result of:
              0.018824315 = score(doc=3495,freq=2.0), product of:
                0.0857324 = queryWeight, product of:
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.021569785 = queryNorm
                0.2195706 = fieldWeight in 3495, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3495)
          0.33333334 = coord(1/3)
        0.0034502088 = weight(_text_:s in 3495) [ClassicSimilarity], result of:
          0.0034502088 = score(doc=3495,freq=12.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.14712115 = fieldWeight in 3495, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3495)
      0.1 = coord(3/30)
    
    Content
    An edited volume containing data structures and algorithms for information retrieval including a disk with examples written in C. for prgrammers and students interested in parsing text, automated indexing, its the first collection in book form of the basic data structures and algorithms that are critical to the storage and retrieval of documents. ------------------Enthält die Kapitel: FRAKES, W.B.: Introduction to information storage and retrieval systems; BAEZA-YATES, R.S.: Introduction to data structures and algorithms related to information retrieval; HARMAN, D. u.a.: Inverted files; FALOUTSOS, C.: Signature files; GONNET, G.H. u.a.: New indices for text: PAT trees and PAT arrays; FORD, D.A. u. S. CHRISTODOULAKIS: File organizations for optical disks; FOX, C.: Lexical analysis and stoplists; FRAKES, W.B.: Stemming algorithms; SRINIVASAN, P.: Thesaurus construction; BAEZA-YATES, R.A.: String searching algorithms; HARMAN, D.: Relevance feedback and other query modification techniques; WARTIK, S.: Boolean operators; WARTIK, S. u.a.: Hashing algorithms; HARMAN, D.: Ranking algorithms; FOX, E.: u.a.: Extended Boolean models; RASMUSSEN, E.: Clustering algorithms; HOLLAAR, L.: Special-purpose hardware for information retrieval; STANFILL, C.: Parallel information retrieval algorithms
    Footnote
    Rez. in: Computing reviews. July 1993, S.341-342 (G. Salton)
    Pages
    504 S
    Type
    s
  18. Lee, D.L.; Ren, L.: Document ranking on weight-partitioned signature files (1996) 0.00
    0.0013843302 = product of:
      0.013843302 = sum of:
        0.0030866629 = weight(_text_:in in 2417) [ClassicSimilarity], result of:
          0.0030866629 = score(doc=2417,freq=2.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.10520181 = fieldWeight in 2417, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2417)
        0.008784681 = product of:
          0.026354041 = sum of:
            0.026354041 = weight(_text_:l in 2417) [ClassicSimilarity], result of:
              0.026354041 = score(doc=2417,freq=2.0), product of:
                0.0857324 = queryWeight, product of:
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.021569785 = queryNorm
                0.30739886 = fieldWeight in 2417, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2417)
          0.33333334 = coord(1/3)
        0.0019719584 = weight(_text_:s in 2417) [ClassicSimilarity], result of:
          0.0019719584 = score(doc=2417,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.08408674 = fieldWeight in 2417, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2417)
      0.1 = coord(3/30)
    
    Abstract
    Proposes the weight partitioned signature file, a signature file organization for supporting document ranking. It uses multiple signature files each corresponding to one term frequency to represent terms with different term frequencies. Words with the same term frequency in a document are grouped together and hased into the signature file corresponding to that term frequency. Investigates the effect of false drops on retrieval effectiveness. Analyses the performance of the weight partitioned signature file under different search strategies and configurations. Obtains an optimal formula for storage allocation to minimise the effect of false drops on document ranks. Analytical results are supported by experiments on document collections
    Source
    ACM transactions on information systems. 14(1996) no.2, S.109-137
  19. Kelledy, F.; Smeaton, A.F.: Signature files and beyond (1996) 0.00
    0.0013103186 = product of:
      0.013103185 = sum of:
        0.0026457112 = weight(_text_:in in 6973) [ClassicSimilarity], result of:
          0.0026457112 = score(doc=6973,freq=2.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.09017298 = fieldWeight in 6973, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=6973)
        0.0016902501 = weight(_text_:s in 6973) [ClassicSimilarity], result of:
          0.0016902501 = score(doc=6973,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.072074346 = fieldWeight in 6973, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.046875 = fieldNorm(doc=6973)
        0.008767224 = product of:
          0.017534448 = sum of:
            0.017534448 = weight(_text_:22 in 6973) [ClassicSimilarity], result of:
              0.017534448 = score(doc=6973,freq=2.0), product of:
                0.07553371 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.021569785 = queryNorm
                0.23214069 = fieldWeight in 6973, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=6973)
          0.5 = coord(1/2)
      0.1 = coord(3/30)
    
    Abstract
    Proposes that signature files be used as a viable alternative to other indexing strategies such as inverted files for searching through large volumes of text. Demonstrates through simulation, that search times can be further reduced by enhancing the basic signature file concept using deterministic partitioning algorithms which eliminate the need for an exhaustive search of the entire signature file. Reports research to evaluate the performance of some deterministic partitioning algorithms in a non simulated environment using 276 MB of raw newspaper text (taken from the Wall Street Journal) and real user queries. Presents a selection of results to illustrate trends and highlight important aspects of the performance of these methods under realistic rather than simulated operating conditions. As a result of the research reported here certain aspects of this approach to signature files are shown to be found wanting and require improvement. Suggests lines of future research on the partitioning of signature files
    Pages
    S.124-144
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
  20. Kantor, P.; Kim, M.H.; Ibraev, U.; Atasoy, K.: Estimating the number of relevant documents in enormous collections (1999) 0.00
    0.0012613306 = product of:
      0.012613306 = sum of:
        0.004929992 = weight(_text_:in in 6690) [ClassicSimilarity], result of:
          0.004929992 = score(doc=6690,freq=10.0), product of:
            0.029340398 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.021569785 = queryNorm
            0.16802745 = fieldWeight in 6690, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6690)
        0.006274772 = product of:
          0.018824315 = sum of:
            0.018824315 = weight(_text_:l in 6690) [ClassicSimilarity], result of:
              0.018824315 = score(doc=6690,freq=2.0), product of:
                0.0857324 = queryWeight, product of:
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.021569785 = queryNorm
                0.2195706 = fieldWeight in 6690, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9746525 = idf(docFreq=2257, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=6690)
          0.33333334 = coord(1/3)
        0.0014085418 = weight(_text_:s in 6690) [ClassicSimilarity], result of:
          0.0014085418 = score(doc=6690,freq=2.0), product of:
            0.023451481 = queryWeight, product of:
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.021569785 = queryNorm
            0.060061958 = fieldWeight in 6690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.0872376 = idf(docFreq=40523, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6690)
      0.1 = coord(3/30)
    
    Abstract
    In assessing information retrieval systems, it is important to know not only the precision of the retrieved set, but also to compare the number of retrieved relevant items to the total number of relevant items. For large collections, such as the TREC test collections, or the World Wide Web, it is not possible to enumerate the entire set of relevant documents. If the retrieved documents are evaluated, a variant of the statistical "capture-recapture" method can be used to estimate the total number of relevant documents, providing the several retrieval systems used are sufficiently independent. We show that the underlying signal detection model supporting such an analysis can be extended in two ways. First, assuming that there are two distinct performance characteristics (corresponding to the chance of retrieving a relevant, and retrieving a given non-relevant document), we show that if there are three or more independent systems available it is possible to estimate the number of relevant documents without actually having to decide whether each individual document is relevant. We report applications of this 3-system method to the TREC data, leading to the conclusion that the independence assumptions are not satisfied. We then extend the model to a multi-system, multi-problem model, and show that it is possible to include statistical dependencies of all orders in the model, and determine the number of relevant documents for each of the problems in the set. Application to the TREC setting will be presented
    Pages
    S.507-514
    Source
    Knowledge: creation, organization and use. Proceedings of the 62nd Annual Meeting of the American Society for Information Science, 31.10.-4.11.1999. Ed.: L. Woods

Languages

  • e 100
  • d 5
  • chi 2
  • More… Less…

Types

  • a 97
  • m 4
  • s 3
  • p 2
  • r 2
  • el 1
  • More… Less…