Search (358 results, page 1 of 18)

  • × theme_ss:"Retrievalalgorithmen"
  1. Behnert, C.; Borst, T.: Neue Formen der Relevanz-Sortierung in bibliothekarischen Informationssystemen : das DFG-Projekt LibRank (2015) 0.02
    0.021775266 = product of:
      0.16331449 = sum of:
        0.09033074 = weight(_text_:wirtschaftswissenschaften in 5392) [ClassicSimilarity], result of:
          0.09033074 = score(doc=5392,freq=4.0), product of:
            0.11380646 = queryWeight, product of:
              6.3497796 = idf(docFreq=209, maxDocs=44218)
              0.017922899 = queryNorm
            0.79372245 = fieldWeight in 5392, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              6.3497796 = idf(docFreq=209, maxDocs=44218)
              0.0625 = fieldNorm(doc=5392)
        0.037188433 = weight(_text_:neue in 5392) [ClassicSimilarity], result of:
          0.037188433 = score(doc=5392,freq=4.0), product of:
            0.07302189 = queryWeight, product of:
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.017922899 = queryNorm
            0.5092779 = fieldWeight in 5392, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.0625 = fieldNorm(doc=5392)
        0.035093244 = weight(_text_:medien in 5392) [ClassicSimilarity], result of:
          0.035093244 = score(doc=5392,freq=2.0), product of:
            0.084356464 = queryWeight, product of:
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.017922899 = queryNorm
            0.4160113 = fieldWeight in 5392, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.0625 = fieldNorm(doc=5392)
        7.0206285E-4 = product of:
          0.0021061886 = sum of:
            0.0021061886 = weight(_text_:a in 5392) [ClassicSimilarity], result of:
              0.0021061886 = score(doc=5392,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.10191591 = fieldWeight in 5392, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5392)
          0.33333334 = coord(1/3)
      0.13333334 = coord(4/30)
    
    Abstract
    Das von der DFG geförderte Projekt LibRank erforscht neue Rankingverfahren für bibliothekarische Informationssysteme, die aufbauend auf Erkenntnissen aus dem Bereich Websuche qualitätsinduzierende Faktoren wie z. B. Aktualität, Popularität und Verfügbarkeit von einzelnen Medien berücksichtigen. Die konzipierten Verfahren werden im Kontext eines in den Wirtschaftswissenschaften häufig genutzten Rechercheportals (EconBiz) entwickelt und in einem Testsystem systematisch evaluiert. Es werden Rankingfaktoren, die für den Bibliotheksbereich von besonderem Interesse sind, vorgestellt und exemplarisch Probleme und Herausforderungen aufgezeigt.
    Field
    Wirtschaftswissenschaften
    Type
    a
  2. Maylein, L.; Langenstein, A.: Neues vom Relevanz-Ranking im HEIDI-Katalog der Universitätsbibliothek Heidelberg : Perspektiven für bibliothekarische Dienstleistungen (2013) 0.01
    0.012087138 = product of:
      0.09065353 = sum of:
        0.02860945 = product of:
          0.0572189 = sum of:
            0.0572189 = weight(_text_:dienstleistungen in 775) [ClassicSimilarity], result of:
              0.0572189 = score(doc=775,freq=2.0), product of:
                0.10771505 = queryWeight, product of:
                  6.009912 = idf(docFreq=294, maxDocs=44218)
                  0.017922899 = queryNorm
                0.5312062 = fieldWeight in 775, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  6.009912 = idf(docFreq=294, maxDocs=44218)
                  0.0625 = fieldNorm(doc=775)
          0.5 = coord(1/2)
        0.0098013915 = product of:
          0.019602783 = sum of:
            0.019602783 = weight(_text_:29 in 775) [ClassicSimilarity], result of:
              0.019602783 = score(doc=775,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.31092256 = fieldWeight in 775, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=775)
          0.5 = coord(1/2)
        0.037188433 = weight(_text_:neue in 775) [ClassicSimilarity], result of:
          0.037188433 = score(doc=775,freq=4.0), product of:
            0.07302189 = queryWeight, product of:
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.017922899 = queryNorm
            0.5092779 = fieldWeight in 775, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.0625 = fieldNorm(doc=775)
        0.015054256 = product of:
          0.022581384 = sum of:
            0.0029786006 = weight(_text_:a in 775) [ClassicSimilarity], result of:
              0.0029786006 = score(doc=775,freq=4.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.14413087 = fieldWeight in 775, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=775)
            0.019602783 = weight(_text_:29 in 775) [ClassicSimilarity], result of:
              0.019602783 = score(doc=775,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.31092256 = fieldWeight in 775, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=775)
          0.6666667 = coord(2/3)
      0.13333334 = coord(4/30)
    
    Abstract
    Das Relevanz-Ranking im Katalog der Universitätsbibliothek Heidelberg HEIDI, bereits 2009 in einem Beitrag in dieser Zeitschrift beschrieben, wurde in den letzten Jahren durch neue Entwicklungen und Methoden stark verbessert. Der Aufsatz beschreibt die Realisierung der bisherigen Rankingmaßnahmen unter der neu eingesetzten Suchmaschinenplattform SOLR. Weiter werden verschiedene neue Möglichkeiten für Rankinganpassungen unter SOLR sowie deren Einsatz im HEIDI-Katalog dargestellt.
    Date
    29. 6.2013 18:06:23
    Type
    a
  3. Weinstein, A.: Hochprozentig : Tipps and tricks für ein Top-Ranking (2002) 0.01
    0.010077579 = product of:
      0.15116368 = sum of:
        0.1499226 = weight(_text_:handwerk in 1083) [ClassicSimilarity], result of:
          0.1499226 = score(doc=1083,freq=2.0), product of:
            0.15594992 = queryWeight, product of:
              8.701155 = idf(docFreq=19, maxDocs=44218)
              0.017922899 = queryNorm
            0.96135086 = fieldWeight in 1083, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.701155 = idf(docFreq=19, maxDocs=44218)
              0.078125 = fieldNorm(doc=1083)
        0.0012410836 = product of:
          0.0037232507 = sum of:
            0.0037232507 = weight(_text_:a in 1083) [ClassicSimilarity], result of:
              0.0037232507 = score(doc=1083,freq=4.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.18016359 = fieldWeight in 1083, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1083)
          0.33333334 = coord(1/3)
      0.06666667 = coord(2/30)
    
    Abstract
    Die Suchmaschinen haben in den letzten Monaten an ihren Ranking-Algorithmen gefeilt, um Spamern das Handwerk zu erschweren. Internet Pro beleuchtet die Trends im Suchmaschinen-Marketing
    Type
    a
  4. Mayr, P.: Bradfordizing als Re-Ranking-Ansatz in Literaturinformationssystemen (2011) 0.01
    0.006611064 = product of:
      0.06611064 = sum of:
        0.047905106 = weight(_text_:wirtschaftswissenschaften in 4292) [ClassicSimilarity], result of:
          0.047905106 = score(doc=4292,freq=2.0), product of:
            0.11380646 = queryWeight, product of:
              6.3497796 = idf(docFreq=209, maxDocs=44218)
              0.017922899 = queryNorm
            0.4209349 = fieldWeight in 4292, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.3497796 = idf(docFreq=209, maxDocs=44218)
              0.046875 = fieldNorm(doc=4292)
        0.0073510436 = product of:
          0.014702087 = sum of:
            0.014702087 = weight(_text_:29 in 4292) [ClassicSimilarity], result of:
              0.014702087 = score(doc=4292,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23319192 = fieldWeight in 4292, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4292)
          0.5 = coord(1/2)
        0.010854486 = product of:
          0.01628173 = sum of:
            0.0015796415 = weight(_text_:a in 4292) [ClassicSimilarity], result of:
              0.0015796415 = score(doc=4292,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.07643694 = fieldWeight in 4292, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4292)
            0.014702087 = weight(_text_:29 in 4292) [ClassicSimilarity], result of:
              0.014702087 = score(doc=4292,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23319192 = fieldWeight in 4292, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4292)
          0.6666667 = coord(2/3)
      0.1 = coord(3/30)
    
    Abstract
    In diesem Artikel wird ein Re-Ranking-Ansatz für Suchsysteme vorgestellt, der die Recherche nach wissenschaftlicher Literatur messbar verbessern kann. Das nichttextorientierte Rankingverfahren Bradfordizing wird eingeführt und anschließend im empirischen Teil des Artikels bzgl. der Effektivität für typische fachbezogene Recherche-Topics evaluiert. Dem Bradford Law of Scattering (BLS), auf dem Bradfordizing basiert, liegt zugrunde, dass sich die Literatur zu einem beliebigen Fachgebiet bzw. -thema in Zonen unterschiedlicher Dokumentenkonzentration verteilt. Dem Kernbereich mit hoher Konzentration der Literatur folgen Bereiche mit mittlerer und geringer Konzentration. Bradfordizing sortiert bzw. rankt eine Dokumentmenge damit nach den sogenannten Kernzeitschriften. Der Retrievaltest mit 164 intellektuell bewerteten Fragestellungen in Fachdatenbanken aus den Bereichen Sozial- und Politikwissenschaften, Wirtschaftswissenschaften, Psychologie und Medizin zeigt, dass die Dokumente der Kernzeitschriften signifikant häufiger relevant bewertet werden als Dokumente der zweiten Dokumentzone bzw. den Peripherie-Zeitschriften. Die Implementierung von Bradfordizing und weiteren Re-Rankingverfahren liefert unmittelbare Mehrwerte für den Nutzer.
    Date
    9. 2.2011 17:47:29
    Type
    a
  5. Fan, W.; Fox, E.A.; Pathak, P.; Wu, H.: ¬The effects of fitness functions an genetic programming-based ranking discovery for Web search (2004) 0.01
    0.0065361825 = product of:
      0.098042734 = sum of:
        0.08622333 = weight(_text_:gas in 2239) [ClassicSimilarity], result of:
          0.08622333 = score(doc=2239,freq=2.0), product of:
            0.15268219 = queryWeight, product of:
              8.518833 = idf(docFreq=23, maxDocs=44218)
              0.017922899 = queryNorm
            0.5647242 = fieldWeight in 2239, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.518833 = idf(docFreq=23, maxDocs=44218)
              0.046875 = fieldNorm(doc=2239)
        0.01181941 = product of:
          0.017729115 = sum of:
            0.003159283 = weight(_text_:a in 2239) [ClassicSimilarity], result of:
              0.003159283 = score(doc=2239,freq=8.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.15287387 = fieldWeight in 2239, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2239)
            0.014569832 = weight(_text_:22 in 2239) [ClassicSimilarity], result of:
              0.014569832 = score(doc=2239,freq=2.0), product of:
                0.06276294 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23214069 = fieldWeight in 2239, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2239)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Abstract
    Genetic-based evolutionary learning algorithms, such as genetic algorithms (GAs) and genetic programming (GP), have been applied to information retrieval (IR) since the 1980s. Recently, GP has been applied to a new IR taskdiscovery of ranking functions for Web search-and has achieved very promising results. However, in our prior research, only one fitness function has been used for GP-based learning. It is unclear how other fitness functions may affect ranking function discovery for Web search, especially since it is weIl known that choosing a proper fitness function is very important for the effectiveness and efficiency of evolutionary algorithms. In this article, we report our experience in contrasting different fitness function designs an GP-based learning using a very large Web corpus. Our results indicate that the design of fitness functions is instrumental in performance improvement. We also give recommendations an the design of fitness functions for genetic-based information retrieval experiments.
    Date
    31. 5.2004 19:22:06
    Type
    a
  6. Martin-Bautista, M.J.; Vila, M.-A.; Larsen, H.L.: ¬A fuzzy genetic algorithm approach to an adaptive information retrieval agent (1999) 0.01
    0.0055726273 = product of:
      0.083589405 = sum of:
        0.08176539 = weight(_text_:760 in 3914) [ClassicSimilarity], result of:
          0.08176539 = score(doc=3914,freq=2.0), product of:
            0.1486828 = queryWeight, product of:
              8.29569 = idf(docFreq=29, maxDocs=44218)
              0.017922899 = queryNorm
            0.5499317 = fieldWeight in 3914, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.29569 = idf(docFreq=29, maxDocs=44218)
              0.046875 = fieldNorm(doc=3914)
        0.0018240128 = product of:
          0.0054720384 = sum of:
            0.0054720384 = weight(_text_:a in 3914) [ClassicSimilarity], result of:
              0.0054720384 = score(doc=3914,freq=24.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.26478532 = fieldWeight in 3914, product of:
                  4.8989797 = tf(freq=24.0), with freq of:
                    24.0 = termFreq=24.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3914)
          0.33333334 = coord(1/3)
      0.06666667 = coord(2/30)
    
    Abstract
    We present an approach to a Genetic Information Retrieval Agent filter (GIRAF) for documents from the Internet using a genetic algorithm (GA) with fuzzy set genes to learn the user's information needs. The population of chromosomes with fixed length represents such user's preferences. Each chromosome is associated with a fitness that may be considered the system's belief in the hypothesis that the chromosome, as a query, represents the user's information needs. In a chromosome, every gene characterizes documents by a keyword and an associated occurence frequency, represented by a certain type of a fuzzy subset of the set of positive integers. Based on the user's evaluation of the documents retrieved by the chromosome, compared to the scores computed by the system, the fitness of the chromosomes is adjusted. A prototype of GIRAF has been developed and tested. The results of the test are discussed, and some directions for further works are pointed out
    Source
    Journal of the American Society for Information Science. 50(1999) no.9, S.760-771
    Type
    a
  7. Hüther, H.: Selix im DFG-Projekt Kascade (1998) 0.01
    0.0051574437 = product of:
      0.051574435 = sum of:
        0.01225174 = product of:
          0.02450348 = sum of:
            0.02450348 = weight(_text_:29 in 5151) [ClassicSimilarity], result of:
              0.02450348 = score(doc=5151,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.38865322 = fieldWeight in 5151, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=5151)
          0.5 = coord(1/2)
        0.021231882 = weight(_text_:u in 5151) [ClassicSimilarity], result of:
          0.021231882 = score(doc=5151,freq=2.0), product of:
            0.058687534 = queryWeight, product of:
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.017922899 = queryNorm
            0.3617784 = fieldWeight in 5151, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.078125 = fieldNorm(doc=5151)
        0.01809081 = product of:
          0.027136216 = sum of:
            0.0026327355 = weight(_text_:a in 5151) [ClassicSimilarity], result of:
              0.0026327355 = score(doc=5151,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.12739488 = fieldWeight in 5151, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=5151)
            0.02450348 = weight(_text_:29 in 5151) [ClassicSimilarity], result of:
              0.02450348 = score(doc=5151,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.38865322 = fieldWeight in 5151, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=5151)
          0.6666667 = coord(2/3)
      0.1 = coord(3/30)
    
    Date
    25. 8.2000 19:55:29
    Source
    Knowledge Management und Kommunikationssysteme: Proceedings des 6. Internationalen Symposiums für Informationswissenschaft (ISI '98) Prag, 3.-7. November 1998 / Hochschulverband für Informationswissenschaft (HI) e.V. Konstanz ; Fachrichtung Informationswissenschaft der Universität des Saarlandes, Saarbrücken. Hrsg.: Harald H. Zimmermann u. Volker Schramm
    Type
    a
  8. Archuby, C.G.: Interfaces se recuperacion para catalogos en linea con salidas ordenadas por probable relevancia (2000) 0.01
    0.0051303366 = product of:
      0.051303364 = sum of:
        0.017326577 = product of:
          0.034653153 = sum of:
            0.034653153 = weight(_text_:29 in 5727) [ClassicSimilarity], result of:
              0.034653153 = score(doc=5727,freq=4.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.5496386 = fieldWeight in 5727, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=5727)
          0.5 = coord(1/2)
        0.009119529 = product of:
          0.018239059 = sum of:
            0.018239059 = weight(_text_:online in 5727) [ClassicSimilarity], result of:
              0.018239059 = score(doc=5727,freq=2.0), product of:
                0.05439423 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.017922899 = queryNorm
                0.33531237 = fieldWeight in 5727, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.078125 = fieldNorm(doc=5727)
          0.5 = coord(1/2)
        0.02485726 = product of:
          0.03728589 = sum of:
            0.0026327355 = weight(_text_:a in 5727) [ClassicSimilarity], result of:
              0.0026327355 = score(doc=5727,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.12739488 = fieldWeight in 5727, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=5727)
            0.034653153 = weight(_text_:29 in 5727) [ClassicSimilarity], result of:
              0.034653153 = score(doc=5727,freq=4.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.5496386 = fieldWeight in 5727, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=5727)
          0.6666667 = coord(2/3)
      0.1 = coord(3/30)
    
    Date
    29. 1.1996 18:23:13
    Footnote
    Übers. d. Titels: Interface for retrieval from online access catalogues with ranked results according to their relevance
    Source
    Ciencia da informacao. 29(2000) no.3, S.5-13
    Type
    a
  9. Chakrabarti, S.; Dom, B.; Kumar, S.R.; Raghavan, P.; Rajagopalan, S.; Tomkins, A.; Kleinberg, J.M.; Gibson, D.: Neue Pfade durch den Internet-Dschungel : Die zweite Generation von Web-Suchmaschinen (1999) 0.01
    0.0051151845 = product of:
      0.051151842 = sum of:
        0.0098013915 = product of:
          0.019602783 = sum of:
            0.019602783 = weight(_text_:29 in 3) [ClassicSimilarity], result of:
              0.019602783 = score(doc=3,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.31092256 = fieldWeight in 3, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3)
          0.5 = coord(1/2)
        0.026296193 = weight(_text_:neue in 3) [ClassicSimilarity], result of:
          0.026296193 = score(doc=3,freq=2.0), product of:
            0.07302189 = queryWeight, product of:
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.017922899 = queryNorm
            0.36011383 = fieldWeight in 3, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.0625 = fieldNorm(doc=3)
        0.015054256 = product of:
          0.022581384 = sum of:
            0.0029786006 = weight(_text_:a in 3) [ClassicSimilarity], result of:
              0.0029786006 = score(doc=3,freq=4.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.14413087 = fieldWeight in 3, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3)
            0.019602783 = weight(_text_:29 in 3) [ClassicSimilarity], result of:
              0.019602783 = score(doc=3,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.31092256 = fieldWeight in 3, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3)
          0.6666667 = coord(2/3)
      0.1 = coord(3/30)
    
    Date
    31.12.1996 19:29:41
    Type
    a
  10. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.00
    0.004065128 = product of:
      0.06097692 = sum of:
        0.039444286 = weight(_text_:neue in 58) [ClassicSimilarity], result of:
          0.039444286 = score(doc=58,freq=2.0), product of:
            0.07302189 = queryWeight, product of:
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.017922899 = queryNorm
            0.5401707 = fieldWeight in 58, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.09375 = fieldNorm(doc=58)
        0.021532632 = product of:
          0.03229895 = sum of:
            0.003159283 = weight(_text_:a in 58) [ClassicSimilarity], result of:
              0.003159283 = score(doc=58,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.15287387 = fieldWeight in 58, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=58)
            0.029139664 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
              0.029139664 = score(doc=58,freq=2.0), product of:
                0.06276294 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.017922899 = queryNorm
                0.46428138 = fieldWeight in 58, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=58)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Date
    14. 6.2015 22:12:44
    Source
    Deutscher Dokumentartag 1985, Nürnberg, 1.-4.10.1985: Fachinformation: Methodik - Management - Markt; neue Entwicklungen, Berufe, Produkte. Bearb.: H. Strohl-Goebel
    Type
    a
  11. Smeaton, A.F.; Rijsbergen, C.J. van: ¬The retrieval effects of query expansion on a feedback document retrieval system (1983) 0.00
    0.0037242575 = product of:
      0.055863857 = sum of:
        0.029724635 = weight(_text_:u in 2134) [ClassicSimilarity], result of:
          0.029724635 = score(doc=2134,freq=2.0), product of:
            0.058687534 = queryWeight, product of:
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.017922899 = queryNorm
            0.50648975 = fieldWeight in 2134, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.109375 = fieldNorm(doc=2134)
        0.02613922 = product of:
          0.03920883 = sum of:
            0.005212551 = weight(_text_:a in 2134) [ClassicSimilarity], result of:
              0.005212551 = score(doc=2134,freq=4.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.25222903 = fieldWeight in 2134, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2134)
            0.033996277 = weight(_text_:22 in 2134) [ClassicSimilarity], result of:
              0.033996277 = score(doc=2134,freq=2.0), product of:
                0.06276294 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.017922899 = queryNorm
                0.5416616 = fieldWeight in 2134, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2134)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Date
    30. 3.2001 13:32:22
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
    Type
    a
  12. Marcus, S.: Textvergleich mit mehreren Mustern (2005) 0.00
    0.0035607335 = product of:
      0.035607334 = sum of:
        0.027439507 = weight(_text_:einzelne in 862) [ClassicSimilarity], result of:
          0.027439507 = score(doc=862,freq=2.0), product of:
            0.10548963 = queryWeight, product of:
              5.885746 = idf(docFreq=333, maxDocs=44218)
              0.017922899 = queryNorm
            0.26011568 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.885746 = idf(docFreq=333, maxDocs=44218)
              0.03125 = fieldNorm(doc=862)
        0.0049006958 = product of:
          0.0098013915 = sum of:
            0.0098013915 = weight(_text_:29 in 862) [ClassicSimilarity], result of:
              0.0098013915 = score(doc=862,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.15546128 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03125 = fieldNorm(doc=862)
          0.5 = coord(1/2)
        0.0032671306 = product of:
          0.0098013915 = sum of:
            0.0098013915 = weight(_text_:29 in 862) [ClassicSimilarity], result of:
              0.0098013915 = score(doc=862,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.15546128 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03125 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
      0.1 = coord(3/30)
    
    Abstract
    Das Gebiet des Pattern-Matching besitzt in vielen wissenschaftlichen Bereichen eine hohe Relevanz. Aufgrund unterschiedlicher Einsatzgebiete sind auch Umsetzung und Anwendung des Pattern-Matching sehr verschieden. Die allen Anwendungen des Pattern-Matching inhärente Aufgabe besteht darin, in einer Vielzahl von Eingabedaten bestimmte Muster wieder zu erkennen. Dies ist auch der deutschen Bezeichnung Mustererkennung zu entnehmen. In der Medizin findet Pattern-Matching zum Beispiel bei der Untersuchung von Chromosomensträngen auf bestimmte Folgen von Chromosomen Verwendung. Auf dem Gebiet der Bildverarbeitung können mit Hilfe des Pattern-Matching ganze Bilder verglichen oder einzelne Bildpunkte betrachtet werden, die durch ein Muster identifizierbar sind. Ein weiteres Einsatzgebiet des Pattern-Matching ist das Information-Retrieval, bei dem in gespeicherten Daten nach relevanten Informationen gesucht wird. Die Relevanz der zu suchenden Daten wird auch hier anhand eines Musters, zum Beispiel einem bestimmten Schlagwort, beurteilt. Ein vergleichbares Verfahren findet auch im Internet Anwendung. Internet-Benutzer, die mittels einer Suchmaschine nach bedeutsamen Informationen suchen, erhalten diese durch den Einsatz eines Pattern-Matching-Automaten. Die in diesem Zusammenhang an den Pattern-Matching-Automaten gestellten Anforderungen variieren mit der Suchanfrage, die an eine Suchmaschine gestellt wird. Eine solche Suchanfrage kann im einfachsten Fall aus genau einem Schlüsselwort bestehen. Im komplexeren Fall enthält die Anfrage mehrere Schlüsselwörter. Dabei muss für eine erfolgreiche Suche eine Konkatenation der in der Anfrage enthaltenen Wörter erfolgen. Zu Beginn dieser Arbeit wird in Kapitel 2 eine umfassende Einführung in die Thematik des Textvergleichs gegeben, wobei die Definition einiger grundlegender Begriffe vorgenommen wird. Anschließend werden in Kapitel 3 Verfahren zum Textvergleich mit mehreren Mustern vorgestellt. Dabei wird zunächst ein einfaches Vorgehen erläutert, um einen Einsteig in das Thema des Textvergleichs mit mehreren Mustern zu erleichtern. Danach wird eine komplexe Methode des Textvergleichs vorgestellt und anhand von Beispielen verdeutlicht.
    Date
    13. 2.2007 20:56:29
  13. Kwok, K.L.: ¬A network approach to probabilistic information retrieval (1995) 0.00
    0.0032677792 = product of:
      0.032677792 = sum of:
        0.0073510436 = product of:
          0.014702087 = sum of:
            0.014702087 = weight(_text_:29 in 5696) [ClassicSimilarity], result of:
              0.014702087 = score(doc=5696,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23319192 = fieldWeight in 5696, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5696)
          0.5 = coord(1/2)
        0.012739129 = weight(_text_:u in 5696) [ClassicSimilarity], result of:
          0.012739129 = score(doc=5696,freq=2.0), product of:
            0.058687534 = queryWeight, product of:
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.017922899 = queryNorm
            0.21706703 = fieldWeight in 5696, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.046875 = fieldNorm(doc=5696)
        0.012587617 = product of:
          0.018881425 = sum of:
            0.0041793385 = weight(_text_:a in 5696) [ClassicSimilarity], result of:
              0.0041793385 = score(doc=5696,freq=14.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.20223314 = fieldWeight in 5696, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5696)
            0.014702087 = weight(_text_:29 in 5696) [ClassicSimilarity], result of:
              0.014702087 = score(doc=5696,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23319192 = fieldWeight in 5696, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5696)
          0.6666667 = coord(2/3)
      0.1 = coord(3/30)
    
    Abstract
    Shows how probabilistic information retrieval based on document components may be implemented as a feedforward (feedbackward) artificial neural network. The network supports adaptation of connection weights as well as the growing of new edges between queries and terms based on user relevance feedback data for training, and it reflects query modification and expansion in information retrieval. A learning rule is applied that can also be viewed as supporting sequential learning using a harmonic sequence learning rate. Experimental results with 4 standard small collections and a large Wall Street Journal collection show that small query expansion levels of about 30 terms can achieve most of the gains at the low-recall high-precision region, while larger expansion levels continue to provide gains at the high-recall low-precision region of a precision recall curve
    Date
    29. 1.1996 18:42:14
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
    Type
    a
  14. Abdelkareem, M.A.A.: In terms of publication index, what indicator is the best for researchers indexing, Google Scholar, Scopus, Clarivate or others? (2018) 0.00
    0.0031582252 = product of:
      0.047373377 = sum of:
        0.046759073 = weight(_text_:post in 4548) [ClassicSimilarity], result of:
          0.046759073 = score(doc=4548,freq=2.0), product of:
            0.10409636 = queryWeight, product of:
              5.808009 = idf(docFreq=360, maxDocs=44218)
              0.017922899 = queryNorm
            0.4491903 = fieldWeight in 4548, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.808009 = idf(docFreq=360, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4548)
        6.14305E-4 = product of:
          0.001842915 = sum of:
            0.001842915 = weight(_text_:a in 4548) [ClassicSimilarity], result of:
              0.001842915 = score(doc=4548,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.089176424 = fieldWeight in 4548, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4548)
          0.33333334 = coord(1/3)
      0.06666667 = coord(2/30)
    
    Abstract
    I believe that Google Scholar is the most popular academic indexing way for researchers and citations. However, some other indexing institutions may be more professional than Google Scholar but not as popular as Google Scholar. Other indexing websites like Scopus and Clarivate are providing more statistical figures for scholars, institutions or even journals. On account of publication citations, always Google Scholar shows higher citations for a paper than other indexing websites since Google Scholar consider most of the publication platforms so he can easily count the citations. While other databases just consider the citations come from those journals that are already indexed in their database
    Source
    https://www.researchgate.net/post/In_terms_of_publication_index_what_indicator_is_the_best_for_researchers_indexing_Google_Scholar_Scopus_Clarivate_or_others
  15. Effektive Information Retrieval Verfahren in Theorie und Praxis : ausgewählte und erweiterte Beiträge des Vierten Hildesheimer Evaluierungs- und Retrievalworkshop (HIER 2005), Hildesheim, 20.7.2005 (2006) 0.00
    0.0029352715 = product of:
      0.029352715 = sum of:
        0.013719753 = weight(_text_:einzelne in 5973) [ClassicSimilarity], result of:
          0.013719753 = score(doc=5973,freq=2.0), product of:
            0.10548963 = queryWeight, product of:
              5.885746 = idf(docFreq=333, maxDocs=44218)
              0.017922899 = queryNorm
            0.13005784 = fieldWeight in 5973, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.885746 = idf(docFreq=333, maxDocs=44218)
              0.015625 = fieldNorm(doc=5973)
        0.011386585 = weight(_text_:neue in 5973) [ClassicSimilarity], result of:
          0.011386585 = score(doc=5973,freq=6.0), product of:
            0.07302189 = queryWeight, product of:
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.017922899 = queryNorm
            0.15593387 = fieldWeight in 5973, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.015625 = fieldNorm(doc=5973)
        0.0042463765 = weight(_text_:u in 5973) [ClassicSimilarity], result of:
          0.0042463765 = score(doc=5973,freq=2.0), product of:
            0.058687534 = queryWeight, product of:
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.017922899 = queryNorm
            0.07235568 = fieldWeight in 5973, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.015625 = fieldNorm(doc=5973)
      0.1 = coord(3/30)
    
    Abstract
    Information Retrieval hat sich zu einer Schlüsseltechnologie in der Wissensgesellschaft entwickelt. Die Anzahl der täglichen Anfragen an Internet-Suchmaschinen bildet nur einen Indikator für die große Bedeutung dieses Themas. Der Sammelbandband informiert über Themen wie Information Retrieval-Grundlagen, Retrieval Systeme, Digitale Bibliotheken, Evaluierung und Multilinguale Systeme, beschreibt Anwendungsszenarien und setzt sich mit neuen Herausforderungen an das Information Retrieval auseinander. Die Beiträge behandeln aktuelle Themen und neue Herausforderungen an das Information Retrieval. Die intensive Beteiligung der Informationswissenschaft der Universität Hildesheim am Cross Language Evaluation Forum (CLEF), einer europäischen Evaluierungsinitiative zur Erforschung mehrsprachiger Retrieval Systeme, berührt mehrere der Beiträge. Ebenso spielen Anwendungsszenarien und die Auseinandersetzung mit aktuellen und praktischen Fragestellungen eine große Rolle.
    Content
    Inhalt: Jan-Hendrik Scheufen: RECOIN: Modell offener Schnittstellen für Information-Retrieval-Systeme und -Komponenten Markus Nick, Klaus-Dieter Althoff: Designing Maintainable Experience-based Information Systems Gesine Quint, Steffen Weichert: Die benutzerzentrierte Entwicklung des Produkt- Retrieval-Systems EIKON der Blaupunkt GmbH Claus-Peter Klas, Sascha Kriewel, André Schaefer, Gudrun Fischer: Das DAFFODIL System - Strategische Literaturrecherche in Digitalen Bibliotheken Matthias Meiert: Entwicklung eines Modells zur Integration digitaler Dokumente in die Universitätsbibliothek Hildesheim Daniel Harbig, René Schneider: Ontology Learning im Rahmen von MyShelf Michael Kluck, Marco Winter: Topic-Entwicklung und Relevanzbewertung bei GIRT: ein Werkstattbericht Thomas Mandl: Neue Entwicklungen bei den Evaluierungsinitiativen im Information Retrieval Joachim Pfister: Clustering von Patent-Dokumenten am Beispiel der Datenbanken des Fachinformationszentrums Karlsruhe Ralph Kölle, Glenn Langemeier, Wolfgang Semar: Programmieren lernen in kollaborativen Lernumgebungen Olga Tartakovski, Margaryta Shramko: Implementierung eines Werkzeugs zur Sprachidentifikation in mono- und multilingualen Texten Nina Kummer: Indexierungstechniken für das japanische Retrieval Suriya Na Nhongkai, Hans-Joachim Bentz: Bilinguale Suche mittels Konzeptnetzen Robert Strötgen, Thomas Mandl, René Schneider: Entwicklung und Evaluierung eines Question Answering Systems im Rahmen des Cross Language Evaluation Forum (CLEF) Niels Jensen: Evaluierung von mehrsprachigem Web-Retrieval: Experimente mit dem EuroGOV-Korpus im Rahmen des Cross Language Evaluation Forum (CLEF)
    Editor
    Mandl, T. u. C. Womser-Hacker
    Footnote
    Rez. in: Information - Wissenschaft und Praxis 57(2006) H.5, S.290-291 (C. Schindler): "Weniger als ein Jahr nach dem "Vierten Hildesheimer Evaluierungs- und Retrievalworkshop" (HIER 2005) im Juli 2005 ist der dazugehörige Tagungsband erschienen. Eingeladen hatte die Hildesheimer Informationswissenschaft um ihre Forschungsergebnisse und die einiger externer Experten zum Thema Information Retrieval einem Fachpublikum zu präsentieren und zur Diskussion zu stellen. Unter dem Titel "Effektive Information Retrieval Verfahren in Theorie und Praxis" sind nahezu sämtliche Beiträge des Workshops in dem nun erschienenen, 15 Beiträge umfassenden Band gesammelt. Mit dem Schwerpunkt Information Retrieval (IR) wird ein Teilgebiet der Informationswissenschaft vorgestellt, das schon immer im Zentrum informationswissenschaftlicher Forschung steht. Ob durch den Leistungsanstieg von Prozessoren und Speichermedien, durch die Verbreitung des Internet über nationale Grenzen hinweg oder durch den stetigen Anstieg der Wissensproduktion, festzuhalten ist, dass in einer zunehmend wechselseitig vernetzten Welt die Orientierung und das Auffinden von Dokumenten in großen Wissensbeständen zu einer zentralen Herausforderung geworden sind. Aktuelle Verfahrensweisen zu diesem Thema, dem Information Retrieval, präsentiert der neue Band anhand von praxisbezogenen Projekten und theoretischen Diskussionen. Das Kernthema Information Retrieval wird in dem Sammelband in die Bereiche Retrieval-Systeme, Digitale Bibliothek, Evaluierung und Multilinguale Systeme untergliedert. Die Artikel der einzelnen Sektionen sind insgesamt recht heterogen und bieten daher keine Überschneidungen inhaltlicher Art. Jedoch ist eine vollkommene thematische Abdeckung der unterschiedlichen Bereiche ebenfalls nicht gegeben, was bei der Präsentation von Forschungsergebnissen eines Institutes und seiner Kooperationspartner auch nur bedingt erwartet werden kann. So lässt sich sowohl in der Gliederung als auch in den einzelnen Beiträgen eine thematische Verdichtung erkennen, die das spezielle Profil und die Besonderheit der Hildesheimer Informationswissenschaft im Feld des Information Retrieval wiedergibt. Teil davon ist die mehrsprachige und interdisziplinäre Ausrichtung, die die Schnittstellen zwischen Informationswissenschaft, Sprachwissenschaft und Informatik in ihrer praxisbezogenen und internationalen Forschung fokussiert.
    "Evaluierung", das Thema des dritten Kapitels, ist in seiner Breite nicht auf das Information Retrieval beschränkt sondern beinhaltet ebenso einzelne Aspekte der Bereiche Mensch-Maschine-Interaktion sowie des E-Learning. Michael Muck und Marco Winter von der Stiftung Wissenschaft und Politik sowie dem Informationszentrum Sozialwissenschaften thematisieren in ihrem Beitrag den Einfluss der Fragestellung (Topic) auf die Bewertung von Relevanz und zeigen Verfahrensweisen für die Topic-Erstellung auf, die beim Cross Language Evaluation Forum (CLEF) Anwendung finden. Im darauf folgenden Aufsatz stellt Thomas Mandl verschiedene Evaluierungsinitiativen im Information Retrieval und aktuelle Entwicklungen dar. Joachim Pfister erläutert in seinem Beitrag das automatisierte Gruppieren, das sogenannte Clustering, von Patent-Dokumenten in den Datenbanken des Fachinformationszentrums Karlsruhe und evaluiert unterschiedliche Clusterverfahren auf Basis von Nutzerbewertungen. Ralph Kölle, Glenn Langemeier und Wolfgang Semar widmen sich dem kollaborativen Lernen unter den speziellen Bedingungen des Programmierens. Dabei werden das System VitaminL zur synchronen Bearbeitung von Programmieraufgaben und das Kennzahlensystem K-3 für die Bewertung kollaborativer Zusammenarbeit in einer Lehrveranstaltung angewendet. Der aktuelle Forschungsschwerpunkt der Hildesheimer Informationswissenschaft zeichnet sich im vierten Kapitel unter dem Thema "Multilinguale Systeme" ab. Hier finden sich die meisten Beiträge des Tagungsbandes wieder. Olga Tartakovski und Margaryta Shramko beschreiben und prüfen das System Langldent, das die Sprache von mono- und multilingualen Texten identifiziert. Die Eigenheiten der japanischen Schriftzeichen stellt Nina Kummer dar und vergleicht experimentell die unterschiedlichen Techniken der Indexierung. Suriya Na Nhongkai und Hans-Joachim Bentz präsentieren und prüfen eine bilinguale Suche auf Basis von Konzeptnetzen, wobei die Konzeptstruktur das verbindende Elemente der beiden Textsammlungen darstellt. Das Entwickeln und Evaluieren eines mehrsprachigen Question-Answering-Systems im Rahmen des Cross Language Evaluation Forum (CLEF), das die alltagssprachliche Formulierung von konkreten Fragestellungen ermöglicht, wird im Beitrag von Robert Strötgen, Thomas Mandl und Rene Schneider thematisiert. Den Schluss bildet der Aufsatz von Niels Jensen, der ein mehrsprachiges Web-Retrieval-System ebenfalls im Zusammenhang mit dem CLEF anhand des multilingualen EuroGOVKorpus evaluiert.
  16. Crestani, F.: Combination of similarity measures for effective spoken document retrieval (2003) 0.00
    0.0028319713 = product of:
      0.042479567 = sum of:
        0.017152434 = product of:
          0.03430487 = sum of:
            0.03430487 = weight(_text_:29 in 4690) [ClassicSimilarity], result of:
              0.03430487 = score(doc=4690,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.5441145 = fieldWeight in 4690, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4690)
          0.5 = coord(1/2)
        0.025327131 = product of:
          0.037990697 = sum of:
            0.00368583 = weight(_text_:a in 4690) [ClassicSimilarity], result of:
              0.00368583 = score(doc=4690,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.17835285 = fieldWeight in 4690, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4690)
            0.03430487 = weight(_text_:29 in 4690) [ClassicSimilarity], result of:
              0.03430487 = score(doc=4690,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.5441145 = fieldWeight in 4690, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4690)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Source
    Journal of information science. 29(2003) no.2, S.87-96
    Type
    a
  17. Calegari, S.; Sanchez, E.: Object-fuzzy concept network : an enrichment of ontologies in semantic information retrieval (2008) 0.00
    0.002819324 = product of:
      0.02819324 = sum of:
        0.00612587 = product of:
          0.01225174 = sum of:
            0.01225174 = weight(_text_:29 in 2393) [ClassicSimilarity], result of:
              0.01225174 = score(doc=2393,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.19432661 = fieldWeight in 2393, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2393)
          0.5 = coord(1/2)
        0.010615941 = weight(_text_:u in 2393) [ClassicSimilarity], result of:
          0.010615941 = score(doc=2393,freq=2.0), product of:
            0.058687534 = queryWeight, product of:
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.017922899 = queryNorm
            0.1808892 = fieldWeight in 2393, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2393)
        0.011451426 = product of:
          0.017177138 = sum of:
            0.004925397 = weight(_text_:a in 2393) [ClassicSimilarity], result of:
              0.004925397 = score(doc=2393,freq=28.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23833402 = fieldWeight in 2393, product of:
                  5.2915025 = tf(freq=28.0), with freq of:
                    28.0 = termFreq=28.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2393)
            0.01225174 = weight(_text_:29 in 2393) [ClassicSimilarity], result of:
              0.01225174 = score(doc=2393,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.19432661 = fieldWeight in 2393, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2393)
          0.6666667 = coord(2/3)
      0.1 = coord(3/30)
    
    Abstract
    This article shows how a fuzzy ontology-based approach can improve semantic documents retrieval. After formally defining a fuzzy ontology and a fuzzy knowledge base, a special type of new fuzzy relationship called (semantic) correlation, which links the concepts or entities in a fuzzy ontology, is discussed. These correlations, first assigned by experts, are updated after querying or when a document has been inserted into a database. Moreover, in order to define a dynamic knowledge of a domain adapting itself to the context, it is shown how to handle a tradeoff between the correct definition of an object, taken in the ontology structure, and the actual meaning assigned by individuals. The notion of a fuzzy concept network is extended, incorporating database objects so that entities and documents can similarly be represented in the network. Information retrieval (IR) algorithm, using an object-fuzzy concept network (O-FCN), is introduced and described. This algorithm allows us to derive a unique path among the entities involved in the query to obtain maxima semantic associations in the knowledge domain. Finally, the study has been validated by querying a database using fuzzy recall, fuzzy precision, and coefficient variant measures in the crisp and fuzzy cases.
    Date
    9.11.2008 13:07:29
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
    Type
    a
  18. Cole, C.: Intelligent information retrieval: diagnosing information need : Part II: uncertainty expansion in a prototype of a diagnostic IR tool (1998) 0.00
    0.0025301932 = product of:
      0.037952896 = sum of:
        0.014702087 = product of:
          0.029404175 = sum of:
            0.029404175 = weight(_text_:29 in 6432) [ClassicSimilarity], result of:
              0.029404175 = score(doc=6432,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.46638384 = fieldWeight in 6432, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6432)
          0.5 = coord(1/2)
        0.023250809 = product of:
          0.034876212 = sum of:
            0.0054720384 = weight(_text_:a in 6432) [ClassicSimilarity], result of:
              0.0054720384 = score(doc=6432,freq=6.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.26478532 = fieldWeight in 6432, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6432)
            0.029404175 = weight(_text_:29 in 6432) [ClassicSimilarity], result of:
              0.029404175 = score(doc=6432,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.46638384 = fieldWeight in 6432, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6432)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Date
    11. 8.2001 14:48:29
    Type
    a
  19. Okada, M.; Ando, K.; Lee, S.S.; Hayashi, Y.; Aoe, J.I.: ¬An efficient substring search method by using delayed keyword extraction (2001) 0.00
    0.0024274043 = product of:
      0.036411062 = sum of:
        0.014702087 = product of:
          0.029404175 = sum of:
            0.029404175 = weight(_text_:29 in 6415) [ClassicSimilarity], result of:
              0.029404175 = score(doc=6415,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.46638384 = fieldWeight in 6415, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6415)
          0.5 = coord(1/2)
        0.021708973 = product of:
          0.03256346 = sum of:
            0.003159283 = weight(_text_:a in 6415) [ClassicSimilarity], result of:
              0.003159283 = score(doc=6415,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.15287387 = fieldWeight in 6415, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6415)
            0.029404175 = weight(_text_:29 in 6415) [ClassicSimilarity], result of:
              0.029404175 = score(doc=6415,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.46638384 = fieldWeight in 6415, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6415)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Date
    29. 3.2002 17:24:03
    Type
    a
  20. Hancock-Beaulieu, M.; Walker, S.: ¬An evaluation of automatic query expansion in an online library catalogue (1992) 0.00
    0.0022309998 = product of:
      0.022309996 = sum of:
        0.0063836705 = product of:
          0.012767341 = sum of:
            0.012767341 = weight(_text_:online in 2731) [ClassicSimilarity], result of:
              0.012767341 = score(doc=2731,freq=2.0), product of:
                0.05439423 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23471867 = fieldWeight in 2731, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2731)
          0.5 = coord(1/2)
        0.014862318 = weight(_text_:u in 2731) [ClassicSimilarity], result of:
          0.014862318 = score(doc=2731,freq=2.0), product of:
            0.058687534 = queryWeight, product of:
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.017922899 = queryNorm
            0.25324488 = fieldWeight in 2731, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2731)
        0.0010640075 = product of:
          0.0031920224 = sum of:
            0.0031920224 = weight(_text_:a in 2731) [ClassicSimilarity], result of:
              0.0031920224 = score(doc=2731,freq=6.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.1544581 = fieldWeight in 2731, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2731)
          0.33333334 = coord(1/3)
      0.1 = coord(3/30)
    
    Abstract
    An automatic query expansion (AQE) facility in anonline catalogue was evaluated in an operational library setting. The OKAPI experimental system had other features including: ranked output 'best match' keyword searching, automatic stemming, spelling normalisation and cross referencing as well as relevance feedback. A combination of transaction log analysis, search replays, questionnaires and interviews was used for data collection. Findings show that contrary to previous results, AQE was beneficial in a substantial number of searches. Use intentions, the effectiveness of the 'best match' search and user interaction were identified as the main factors affecting the take-up of the query expansion facility
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
    Type
    a

Years

Languages

Types

  • a 337
  • m 9
  • el 8
  • s 4
  • p 2
  • r 2
  • x 2
  • More… Less…