Search (173 results, page 1 of 9)

  • × theme_ss:"Automatisches Klassifizieren"
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.21
    0.20508106 = product of:
      0.41016212 = sum of:
        0.057200458 = product of:
          0.17160137 = sum of:
            0.17160137 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.17160137 = score(doc=562,freq=2.0), product of:
                0.30533072 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.036014426 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.17160137 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.17160137 = score(doc=562,freq=2.0), product of:
            0.30533072 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.036014426 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.17160137 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.17160137 = score(doc=562,freq=2.0), product of:
            0.30533072 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.036014426 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.009758915 = product of:
          0.029276744 = sum of:
            0.029276744 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.029276744 = score(doc=562,freq=2.0), product of:
                0.12611638 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.036014426 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
      0.5 = coord(4/8)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  2. Panyr, J.: Vektorraum-Modell und Clusteranalyse in Information-Retrieval-Systemen (1987) 0.11
    0.11161989 = product of:
      0.29765305 = sum of:
        0.016991155 = weight(_text_:information in 2322) [ClassicSimilarity], result of:
          0.016991155 = score(doc=2322,freq=6.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.2687516 = fieldWeight in 2322, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=2322)
        0.05044969 = weight(_text_:retrieval in 2322) [ClassicSimilarity], result of:
          0.05044969 = score(doc=2322,freq=6.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.46309367 = fieldWeight in 2322, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=2322)
        0.2302122 = weight(_text_:modell in 2322) [ClassicSimilarity], result of:
          0.2302122 = score(doc=2322,freq=8.0), product of:
            0.21656582 = queryWeight, product of:
              6.0133076 = idf(docFreq=293, maxDocs=44218)
              0.036014426 = queryNorm
            1.0630126 = fieldWeight in 2322, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              6.0133076 = idf(docFreq=293, maxDocs=44218)
              0.0625 = fieldNorm(doc=2322)
      0.375 = coord(3/8)
    
    Abstract
    Ausgehend von theoretischen Indexierungsansätzen wird das klassische Vektorraum-Modell für automatische Indexierung (mit dem Trennschärfen-Modell) erläutert. Das Clustering in Information-Retrieval-Systemem wird als eine natürliche logische Folge aus diesem Modell aufgefaßt und in allen seinen Ausprägungen (d.h. als Dokumenten-, Term- oder Dokumenten- und Termklassifikation) behandelt. Anschließend werden die Suchstrategien in vorklassifizierten Dokumentenbeständen (Clustersuche) detailliert beschrieben. Zum Schluß wird noch die sinnvolle Anwendung der Clusteranalyse in Information-Retrieval-Systemen kurz diskutiert
  3. Reiner, U.: Automatische DDC-Klassifizierung von bibliografischen Titeldatensätzen (2009) 0.02
    0.024351038 = product of:
      0.0649361 = sum of:
        0.012262309 = weight(_text_:information in 611) [ClassicSimilarity], result of:
          0.012262309 = score(doc=611,freq=2.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.19395474 = fieldWeight in 611, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=611)
        0.036408927 = weight(_text_:retrieval in 611) [ClassicSimilarity], result of:
          0.036408927 = score(doc=611,freq=2.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.33420905 = fieldWeight in 611, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.078125 = fieldNorm(doc=611)
        0.01626486 = product of:
          0.048794575 = sum of:
            0.048794575 = weight(_text_:22 in 611) [ClassicSimilarity], result of:
              0.048794575 = score(doc=611,freq=2.0), product of:
                0.12611638 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.036014426 = queryNorm
                0.38690117 = fieldWeight in 611, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=611)
          0.33333334 = coord(1/3)
      0.375 = coord(3/8)
    
    Content
    Präsentation zum Vortrag anlässlich des 98. Deutscher Bibliothekartag in Erfurt: Ein neuer Blick auf Bibliotheken; TK10: Information erschließen und recherchieren Inhalte erschließen - mit neuen Tools
    Date
    22. 8.2009 12:54:24
    Theme
    Klassifikationssysteme im Online-Retrieval
  4. Mengle, S.; Goharian, N.: Passage detection using text classification (2009) 0.02
    0.022912933 = product of:
      0.061101153 = sum of:
        0.012262309 = weight(_text_:information in 2765) [ClassicSimilarity], result of:
          0.012262309 = score(doc=2765,freq=8.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.19395474 = fieldWeight in 2765, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2765)
        0.04070642 = weight(_text_:retrieval in 2765) [ClassicSimilarity], result of:
          0.04070642 = score(doc=2765,freq=10.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.37365708 = fieldWeight in 2765, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2765)
        0.00813243 = product of:
          0.024397288 = sum of:
            0.024397288 = weight(_text_:22 in 2765) [ClassicSimilarity], result of:
              0.024397288 = score(doc=2765,freq=2.0), product of:
                0.12611638 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.036014426 = queryNorm
                0.19345059 = fieldWeight in 2765, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2765)
          0.33333334 = coord(1/3)
      0.375 = coord(3/8)
    
    Abstract
    Passages can be hidden within a text to circumvent their disallowed transfer. Such release of compartmentalized information is of concern to all corporate and governmental organizations. Passage retrieval is well studied; we posit, however, that passage detection is not. Passage retrieval is the determination of the degree of relevance of blocks of text, namely passages, comprising a document. Rather than determining the relevance of a document in its entirety, passage retrieval determines the relevance of the individual passages. As such, modified traditional information-retrieval techniques compare terms found in user queries with the individual passages to determine a similarity score for passages of interest. In passage detection, passages are classified into predetermined categories. More often than not, passage detection techniques are deployed to detect hidden paragraphs in documents. That is, to hide information, documents are injected with hidden text into passages. Rather than matching query terms against passages to determine their relevance, using text-mining techniques, the passages are classified. Those documents with hidden passages are defined as infected. Thus, simply stated, passage retrieval is the search for passages relevant to a user query, while passage detection is the classification of passages. That is, in passage detection, passages are labeled with one or more categories from a set of predetermined categories. We present a keyword-based dynamic passage approach (KDP) and demonstrate that KDP outperforms statistically significantly (99% confidence) the other document-splitting approaches by 12% to 18% in the passage detection and passage category-prediction tasks. Furthermore, we evaluate the effects of the feature selection, passage length, ambiguous passages, and finally training-data category distribution on passage-detection accuracy.
    Date
    22. 3.2009 19:14:43
    Source
    Journal of the American Society for Information Science and Technology. 60(2009) no.4, S.814-825
  5. Schiminovich, S.: Automatic classification and retrieval of documents by means of a bibliographic pattern discovery algorithm (1971) 0.02
    0.022313308 = product of:
      0.08925323 = sum of:
        0.017167233 = weight(_text_:information in 4846) [ClassicSimilarity], result of:
          0.017167233 = score(doc=4846,freq=2.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.27153665 = fieldWeight in 4846, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=4846)
        0.072086 = weight(_text_:retrieval in 4846) [ClassicSimilarity], result of:
          0.072086 = score(doc=4846,freq=4.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.6617001 = fieldWeight in 4846, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.109375 = fieldNorm(doc=4846)
      0.25 = coord(2/8)
    
    Source
    Information storage and retrieval. 6(1971), S.417-435
  6. Panyr, J.: Automatische Klassifikation und Information Retrieval : Anwendung und Entwicklung komplexer Verfahren in Information-Retrieval-Systemen und ihre Evaluierung (1986) 0.02
    0.021818683 = product of:
      0.08727473 = sum of:
        0.025486732 = weight(_text_:information in 32) [ClassicSimilarity], result of:
          0.025486732 = score(doc=32,freq=6.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.40312737 = fieldWeight in 32, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=32)
        0.061788 = weight(_text_:retrieval in 32) [ClassicSimilarity], result of:
          0.061788 = score(doc=32,freq=4.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.5671716 = fieldWeight in 32, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.09375 = fieldNorm(doc=32)
      0.25 = coord(2/8)
    
    Series
    Sprache und Information; Bd.12
  7. Rijsbergen, C.J. van: Automatic classification in information retrieval (1978) 0.02
    0.019468494 = product of:
      0.077873975 = sum of:
        0.019619694 = weight(_text_:information in 2412) [ClassicSimilarity], result of:
          0.019619694 = score(doc=2412,freq=2.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.3103276 = fieldWeight in 2412, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=2412)
        0.058254283 = weight(_text_:retrieval in 2412) [ClassicSimilarity], result of:
          0.058254283 = score(doc=2412,freq=2.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.5347345 = fieldWeight in 2412, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.125 = fieldNorm(doc=2412)
      0.25 = coord(2/8)
    
  8. Jenkins, C.: Automatic classification of Web resources using Java and Dewey Decimal Classification (1998) 0.02
    0.018379018 = product of:
      0.049010716 = sum of:
        0.012139066 = weight(_text_:information in 1673) [ClassicSimilarity], result of:
          0.012139066 = score(doc=1673,freq=4.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.1920054 = fieldWeight in 1673, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1673)
        0.02548625 = weight(_text_:retrieval in 1673) [ClassicSimilarity], result of:
          0.02548625 = score(doc=1673,freq=2.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.23394634 = fieldWeight in 1673, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1673)
        0.0113854 = product of:
          0.0341562 = sum of:
            0.0341562 = weight(_text_:22 in 1673) [ClassicSimilarity], result of:
              0.0341562 = score(doc=1673,freq=2.0), product of:
                0.12611638 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.036014426 = queryNorm
                0.2708308 = fieldWeight in 1673, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1673)
          0.33333334 = coord(1/3)
      0.375 = coord(3/8)
    
    Abstract
    The Wolverhampton Web Library (WWLib) is a WWW search engine that provides access to UK based information. The experimental version developed in 1995, was a success but highlighted the need for a much higher degree of automation. An interesting feature of the experimental WWLib was that it organised information according to DDC. Discusses the advantages of classification and describes the automatic classifier that is being developed in Java as part of the new, fully automated WWLib
    Date
    1. 8.1996 22:08:06
    Theme
    Klassifikationssysteme im Online-Retrieval
  9. Oberhauser, O.: Automatisches Klassifizieren : Entwicklungsstand - Methodik - Anwendungsbereiche (2005) 0.02
    0.018051922 = product of:
      0.048138462 = sum of:
        0.0030655772 = weight(_text_:information in 38) [ClassicSimilarity], result of:
          0.0030655772 = score(doc=38,freq=2.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.048488684 = fieldWeight in 38, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.01953125 = fieldNorm(doc=38)
        0.009102232 = weight(_text_:retrieval in 38) [ClassicSimilarity], result of:
          0.009102232 = score(doc=38,freq=2.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.08355226 = fieldWeight in 38, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.01953125 = fieldNorm(doc=38)
        0.035970654 = weight(_text_:modell in 38) [ClassicSimilarity], result of:
          0.035970654 = score(doc=38,freq=2.0), product of:
            0.21656582 = queryWeight, product of:
              6.0133076 = idf(docFreq=293, maxDocs=44218)
              0.036014426 = queryNorm
            0.16609572 = fieldWeight in 38, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.0133076 = idf(docFreq=293, maxDocs=44218)
              0.01953125 = fieldNorm(doc=38)
      0.375 = coord(3/8)
    
    Abstract
    Automatisches Klassifizieren von Textdokumenten bedeutet die maschinelle Zuordnung jeweils einer oder mehrerer Notationen eines vorgegebenen Klassifikationssystems zu natürlich-sprachlichen Texten mithilfe eines geeigneten Algorithmus. In der vorliegenden Arbeit wird in Form einer umfassenden Literaturstudie ein aktueller Kenntnisstand zu den Ein-satzmöglichkeiten des automatischen Klassifizierens für die sachliche Erschliessung von elektronischen Dokumenten, insbesondere von Web-Ressourcen, erarbeitet. Dies betrifft zum einen den methodischen Aspekt und zum anderen die in relevanten Projekten und Anwendungen gewonnenen Erfahrungen. In methodischer Hinsicht gelten heute statistische Verfahren, die auf dem maschinellen Lernen basieren und auf der Grundlage bereits klassifizierter Beispieldokumente ein Modell - einen "Klassifikator" - erstellen, das zur Klassifizierung neuer Dokumente verwendet werden kann, als "state-of-the-art". Die vier in den 1990er Jahren an den Universitäten Lund, Wolverhampton und Oldenburg sowie bei OCLC (Dublin, OH) durchgeführten "grossen" Projekte zum automatischen Klassifizieren von Web-Ressourcen, die in dieser Arbeit ausführlich analysiert werden, arbeiteten allerdings noch mit einfacheren bzw. älteren methodischen Ansätzen. Diese Projekte bedeuten insbesondere aufgrund ihrer Verwendung etablierter bibliothekarischer Klassifikationssysteme einen wichtigen Erfahrungsgewinn, selbst wenn sie bisher nicht zu permanenten und qualitativ zufriedenstellenden Diensten für die Erschliessung elektronischer Ressourcen geführt haben. Die Analyse der weiteren einschlägigen Anwendungen und Projekte lässt erkennen, dass derzeit in den Bereichen Patent- und Mediendokumentation die aktivsten Bestrebungen bestehen, Systeme für die automatische klassifikatorische Erschliessung elektronischer Dokumente im laufenden operativen Betrieb einzusetzen. Dabei dominieren jedoch halbautomatische Systeme, die menschliche Bearbeiter durch Klassifizierungsvorschläge unterstützen, da die gegenwärtig erreichbare Klassifizierungsgüte für eine Vollautomatisierung meist noch nicht ausreicht. Weitere interessante Anwendungen und Projekte finden sich im Bereich von Web-Portalen, Suchmaschinen und (kommerziellen) Informationsdiensten, während sich etwa im Bibliothekswesen kaum nennenswertes Interesse an einer automatischen Klassifizierung von Büchern bzw. bibliographischen Datensätzen registrieren lässt. Die Studie schliesst mit einer Diskussion der wichtigsten Projekte und Anwendungen sowie einiger im Zusammenhang mit dem automatischen Klassifizieren relevanter Fragestellungen und Themen.
    Footnote
    Die am Anfang des Werkes gestellte Frage, ob »die Techniken des automatischen Klassifizierens heute bereits so weit [sind], dass damit grosse Mengen elektronischer Dokumente [-] zufrieden stellend erschlossen werden können? « (S. 13), beantwortet der Verfasser mit einem eindeutigen »nein«, was Salton und McGills Aussage von 1983, »daß einfache automatische Indexierungsverfahren schnell und kostengünstig arbeiten, und daß sie Recall- und Precisionwerte erreichen, die mindestens genauso gut sind wie bei der manuellen Indexierung mit kontrolliertem Vokabular « (Gerard Salton und Michael J. McGill: Information Retrieval. Hamburg u.a. 1987, S. 64 f.) kräftig relativiert. Über die Gründe, warum drei der großen Projekte nicht weiter verfolgt werden, will Oberhauser nicht spekulieren, nennt aber mangelnden Erfolg, Verlagerung der Arbeit in den beteiligten Institutionen sowie Finanzierungsprobleme als mögliche Ursachen. Das größte Entwicklungspotenzial beim automatischen Erschließen großer Dokumentenmengen sieht der Verfasser heute in den Bereichen der Patentund Mediendokumentation. Hier solle man im bibliothekarischen Bereich die Entwicklung genau verfolgen, da diese »sicherlich mittelfristig auf eine qualitativ zufrieden stellende Vollautomatisierung« abziele (S. 146). Oberhausers Darstellung ist ein rundum gelungenes Werk, das zum Handapparat eines jeden, der sich für automatische Erschließung interessiert, gehört."
  10. Schek, M.: Automatische Klassifizierung in Erschließung und Recherche eines Pressearchivs (2006) 0.02
    0.018029155 = product of:
      0.07211662 = sum of:
        0.014563571 = weight(_text_:retrieval in 6043) [ClassicSimilarity], result of:
          0.014563571 = score(doc=6043,freq=2.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.13368362 = fieldWeight in 6043, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=6043)
        0.05755305 = weight(_text_:modell in 6043) [ClassicSimilarity], result of:
          0.05755305 = score(doc=6043,freq=2.0), product of:
            0.21656582 = queryWeight, product of:
              6.0133076 = idf(docFreq=293, maxDocs=44218)
              0.036014426 = queryNorm
            0.26575315 = fieldWeight in 6043, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.0133076 = idf(docFreq=293, maxDocs=44218)
              0.03125 = fieldNorm(doc=6043)
      0.25 = coord(2/8)
    
    Abstract
    Die Süddeutsche Zeitung (SZ) verfügt seit ihrer Gründung 1945 über ein Pressearchiv, das die Texte der eigenen Redakteure und zahlreicher nationaler und internationaler Publikationen dokumentiert und für Recherchezwecke bereitstellt. Die DIZ-Pressedatenbank (www.medienport.de) ermöglicht die browserbasierte Recherche für Redakteure und externe Kunden im Intra- und Internet und die kundenspezifischen Content Feeds für Verlage, Rundfunkanstalten und Portale. Die DIZ-Pressedatenbank enthält z. Zt. 7,8 Millionen Artikel, die jeweils als HTML oder PDF abrufbar sind. Täglich kommen ca. 3.500 Artikel hinzu, von denen ca. 1.000 durch Dokumentare inhaltlich erschlossen werden. Die Informationserschließung erfolgt im DIZ nicht durch die Vergabe von Schlagwörtern am Dokument, sondern durch die Verlinkung der Artikel mit "virtuellen Mappen", den Dossiers. Insgesamt enthält die DIZ-Pressedatenbank ca. 90.000 Dossiers, die untereinander zum "DIZ-Wissensnetz" verlinkt sind. DIZ definiert das Wissensnetz als Alleinstellungsmerkmal und wendet beträchtliche personelle Ressourcen für die Aktualisierung und Qualitätssicherung der Dossiers auf. Im Zuge der Medienkrise mussten sich DIZ der Herausforderung stellen, bei sinkenden Lektoratskapazitäten die Qualität der Informationserschließung im Input zu erhalten. Auf der Outputseite gilt es, eine anspruchsvolle Zielgruppe - u.a. die Redakteure der Süddeutschen Zeitung - passgenau und zeitnah mit den Informationen zu versorgen, die sie für ihre tägliche Arbeit benötigt. Bezogen auf die Ausgangssituation in der Dokumentation der Süddeutschen Zeitung identifizierte DIZ drei Ansatzpunkte, wie die Aufwände auf der Inputseite (Lektorat) zu optimieren sind und gleichzeitig auf der Outputseite (Recherche) das Wissensnetz besser zu vermarkten ist: - (Teil-)Automatische Klassifizierung von Pressetexten (Vorschlagwesen) - Visualisierung des Wissensnetzes - Neue Retrievalmöglichkeiten (Ähnlichkeitssuche, Clustering) Im Bereich "Visualisierung" setzt DIZ auf den Net-Navigator von intelligent views, eine interaktive Visualisierung allgemeiner Graphen, basierend auf einem physikalischen Modell. In den Bereichen automatische Klassifizierung, Ähnlichkeitssuche und Clustering hat DIZ sich für das Produkt nextBot der Firma Brainbot entschieden.
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  11. Wu, M.; Fuller, M.; Wilkinson, R.: Using clustering and classification approaches in interactive retrieval (2001) 0.02
    0.017034933 = product of:
      0.06813973 = sum of:
        0.017167233 = weight(_text_:information in 2666) [ClassicSimilarity], result of:
          0.017167233 = score(doc=2666,freq=2.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.27153665 = fieldWeight in 2666, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=2666)
        0.0509725 = weight(_text_:retrieval in 2666) [ClassicSimilarity], result of:
          0.0509725 = score(doc=2666,freq=2.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.46789268 = fieldWeight in 2666, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.109375 = fieldNorm(doc=2666)
      0.25 = coord(2/8)
    
    Source
    Information processing and management. 37(2001) no.3, S.459-484
  12. Ko, Y.: ¬A new term-weighting scheme for text classification using the odds of positive and negative class probabilities (2015) 0.02
    0.016125135 = product of:
      0.06450054 = sum of:
        0.020809827 = weight(_text_:information in 2339) [ClassicSimilarity], result of:
          0.020809827 = score(doc=2339,freq=16.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.3291521 = fieldWeight in 2339, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2339)
        0.043690715 = weight(_text_:retrieval in 2339) [ClassicSimilarity], result of:
          0.043690715 = score(doc=2339,freq=8.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.40105087 = fieldWeight in 2339, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=2339)
      0.25 = coord(2/8)
    
    Abstract
    Text classification (TC) is a core technique for text mining and information retrieval. It has been applied to many applications in many different research and industrial areas. Term-weighting schemes assign an appropriate weight to each term to obtain a high TC performance. Although term weighting is one of the important modules for TC and TC has different peculiarities from those in information retrieval, many term-weighting schemes used in information retrieval, such as term frequency-inverse document frequency (tf-idf), have been used in TC in the same manner. The peculiarity of TC that differs most from information retrieval is the existence of class information. This article proposes a new term-weighting scheme that uses class information using positive and negative class distributions. As a result, the proposed scheme, log tf-TRR, consistently performs better than do other schemes using class information as well as traditional schemes such as tf-idf.
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.12, S.2553-2565
  13. Ribeiro-Neto, B.; Laender, A.H.F.; Lima, L.R.S. de: ¬An experimental study in automatically categorizing medical documents (2001) 0.02
    0.015983254 = product of:
      0.04262201 = sum of:
        0.008670762 = weight(_text_:information in 5702) [ClassicSimilarity], result of:
          0.008670762 = score(doc=5702,freq=4.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.13714671 = fieldWeight in 5702, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5702)
        0.025744999 = weight(_text_:retrieval in 5702) [ClassicSimilarity], result of:
          0.025744999 = score(doc=5702,freq=4.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.23632148 = fieldWeight in 5702, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5702)
        0.008206251 = product of:
          0.024618752 = sum of:
            0.024618752 = weight(_text_:29 in 5702) [ClassicSimilarity], result of:
              0.024618752 = score(doc=5702,freq=2.0), product of:
                0.1266875 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.036014426 = queryNorm
                0.19432661 = fieldWeight in 5702, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5702)
          0.33333334 = coord(1/3)
      0.375 = coord(3/8)
    
    Abstract
    In this article, we evaluate the retrieval performance of an algorithm that automatically categorizes medical documents. The categorization, which consists in assigning an International Code of Disease (ICD) to the medical document under examination, is based on wellknown information retrieval techniques. The algorithm, which we proposed, operates in a fully automatic mode and requires no supervision or training data. Using a database of 20,569 documents, we verify that the algorithm attains levels of average precision in the 70-80% range for category coding and in the 60-70% range for subcategory coding. We also carefully analyze the case of those documents whose categorization is not in accordance with the one provided by the human specialists. The vast majority of them represent cases that can only be fully categorized with the assistance of a human subject (because, for instance, they require specific knowledge of a given pathology). For a slim fraction of all documents (0.77% for category coding and 1.4% for subcategory coding), the algorithm makes assignments that are clearly incorrect. However, this fraction corresponds to only one-fourth of the mistakes made by the human specialists
    Date
    29. 9.2001 13:59:42
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.5, S.391-401
  14. Liu, R.-L.: ¬A passage extractor for classification of disease aspect information (2013) 0.02
    0.0159594 = product of:
      0.0425584 = sum of:
        0.01622151 = weight(_text_:information in 1107) [ClassicSimilarity], result of:
          0.01622151 = score(doc=1107,freq=14.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.256578 = fieldWeight in 1107, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1107)
        0.018204464 = weight(_text_:retrieval in 1107) [ClassicSimilarity], result of:
          0.018204464 = score(doc=1107,freq=2.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.16710453 = fieldWeight in 1107, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1107)
        0.00813243 = product of:
          0.024397288 = sum of:
            0.024397288 = weight(_text_:22 in 1107) [ClassicSimilarity], result of:
              0.024397288 = score(doc=1107,freq=2.0), product of:
                0.12611638 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.036014426 = queryNorm
                0.19345059 = fieldWeight in 1107, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1107)
          0.33333334 = coord(1/3)
      0.375 = coord(3/8)
    
    Abstract
    Retrieval of disease information is often based on several key aspects such as etiology, diagnosis, treatment, prevention, and symptoms of diseases. Automatic identification of disease aspect information is thus essential. In this article, I model the aspect identification problem as a text classification (TC) problem in which a disease aspect corresponds to a category. The disease aspect classification problem poses two challenges to classifiers: (a) a medical text often contains information about multiple aspects of a disease and hence produces noise for the classifiers and (b) text classifiers often cannot extract the textual parts (i.e., passages) about the categories of interest. I thus develop a technique, PETC (Passage Extractor for Text Classification), that extracts passages (from medical texts) for the underlying text classifiers to classify. Case studies on thousands of Chinese and English medical texts show that PETC enhances a support vector machine (SVM) classifier in classifying disease aspect information. PETC also performs better than three state-of-the-art classifier enhancement techniques, including two passage extraction techniques for text classifiers and a technique that employs term proximity information to enhance text classifiers. The contribution is of significance to evidence-based medicine, health education, and healthcare decision support. PETC can be used in those application domains in which a text to be classified may have several parts about different categories.
    Date
    28.10.2013 19:22:57
    Source
    Journal of the American Society for Information Science and Technology. 64(2013) no.11, S.2265-2277
  15. Yu, W.; Gong, Y.: Document clustering by concept factorization (2004) 0.01
    0.014601371 = product of:
      0.058405485 = sum of:
        0.014714771 = weight(_text_:information in 4084) [ClassicSimilarity], result of:
          0.014714771 = score(doc=4084,freq=2.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.23274569 = fieldWeight in 4084, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=4084)
        0.043690715 = weight(_text_:retrieval in 4084) [ClassicSimilarity], result of:
          0.043690715 = score(doc=4084,freq=2.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.40105087 = fieldWeight in 4084, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.09375 = fieldNorm(doc=4084)
      0.25 = coord(2/8)
    
    Source
    SIGIR'04: Proceedings of the 27th Annual International ACM-SIGIR Conference an Research and Development in Information Retrieval. Ed.: K. Järvelin, u.a
  16. Guerrero-Bote, V.P.; Moya Anegón, F. de; Herrero Solana, V.: Document organization using Kohonen's algorithm (2002) 0.01
    0.014545789 = product of:
      0.058183156 = sum of:
        0.016991155 = weight(_text_:information in 2564) [ClassicSimilarity], result of:
          0.016991155 = score(doc=2564,freq=6.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.2687516 = fieldWeight in 2564, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=2564)
        0.041192 = weight(_text_:retrieval in 2564) [ClassicSimilarity], result of:
          0.041192 = score(doc=2564,freq=4.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.37811437 = fieldWeight in 2564, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=2564)
      0.25 = coord(2/8)
    
    Abstract
    The classification of documents from a bibliographic database is a task that is linked to processes of information retrieval based on partial matching. A method is described of vectorizing reference documents from LISA which permits their topological organization using Kohonen's algorithm. As an example a map is generated of 202 documents from LISA, and an analysis is made of the possibilities of this type of neural network with respect to the development of information retrieval systems based on graphical browsing.
    Source
    Information processing and management. 38(2002) no.1, S.79-89
  17. Chung, Y.M.; Lee, J.Y.: ¬A corpus-based approach to comparative evaluation of statistical term association measures (2001) 0.01
    0.014502384 = product of:
      0.038673025 = sum of:
        0.012262309 = weight(_text_:information in 5769) [ClassicSimilarity], result of:
          0.012262309 = score(doc=5769,freq=8.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.19395474 = fieldWeight in 5769, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5769)
        0.018204464 = weight(_text_:retrieval in 5769) [ClassicSimilarity], result of:
          0.018204464 = score(doc=5769,freq=2.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.16710453 = fieldWeight in 5769, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5769)
        0.008206251 = product of:
          0.024618752 = sum of:
            0.024618752 = weight(_text_:29 in 5769) [ClassicSimilarity], result of:
              0.024618752 = score(doc=5769,freq=2.0), product of:
                0.1266875 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.036014426 = queryNorm
                0.19432661 = fieldWeight in 5769, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5769)
          0.33333334 = coord(1/3)
      0.375 = coord(3/8)
    
    Abstract
    Statistical association measures have been widely applied in information retrieval research, usually employing a clustering of documents or terms on the basis of their relationships. Applications of the association measures for term clustering include automatic thesaurus construction and query expansion. This research evaluates the similarity of six association measures by comparing the relationship and behavior they demonstrate in various analyses of a test corpus. Analysis techniques include comparisons of highly ranked term pairs and term clusters, analyses of the correlation among the association measures using Pearson's correlation coefficient and MDS mapping, and an analysis of the impact of a term frequency on the association values by means of z-score. The major findings of the study are as follows: First, the most similar association measures are mutual information and Yule's coefficient of colligation Y, whereas cosine and Jaccard coefficients, as well as X**2 statistic and likelihood ratio, demonstrate quite similar behavior for terms with high frequency. Second, among all the measures, the X**2 statistic is the least affected by the frequency of terms. Third, although cosine and Jaccard coefficients tend to emphasize high frequency terms, mutual information and Yule's Y seem to overestimate rare terms
    Date
    29. 9.2001 14:01:18
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.4, S.283-296
  18. Möller, G.: Automatic classification of the World Wide Web using Universal Decimal Classification (1999) 0.01
    0.014411967 = product of:
      0.05764787 = sum of:
        0.021238944 = weight(_text_:information in 494) [ClassicSimilarity], result of:
          0.021238944 = score(doc=494,freq=6.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.3359395 = fieldWeight in 494, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=494)
        0.036408927 = weight(_text_:retrieval in 494) [ClassicSimilarity], result of:
          0.036408927 = score(doc=494,freq=2.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.33420905 = fieldWeight in 494, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.078125 = fieldNorm(doc=494)
      0.25 = coord(2/8)
    
    Imprint
    Hinskey Hill : Learned Information
    Source
    Online information 99: 23rd International Online Information Meeting, Proceedings, London, 7-9 December 1999. Ed.: D. Raitt et al
    Theme
    Klassifikationssysteme im Online-Retrieval
  19. Miyamoto, S.: Information clustering based an fuzzy multisets (2003) 0.01
    0.0142671205 = product of:
      0.057068482 = sum of:
        0.02102548 = weight(_text_:information in 1071) [ClassicSimilarity], result of:
          0.02102548 = score(doc=1071,freq=12.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.3325631 = fieldWeight in 1071, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1071)
        0.036043 = weight(_text_:retrieval in 1071) [ClassicSimilarity], result of:
          0.036043 = score(doc=1071,freq=4.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.33085006 = fieldWeight in 1071, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1071)
      0.25 = coord(2/8)
    
    Abstract
    A fuzzy multiset model for information clustering is proposed with application to information retrieval on the World Wide Web. Noting that a search engine retrieves multiple occurrences of the same subjects with possibly different degrees of relevance, we observe that fuzzy multisets provide an appropriate model of information retrieval on the WWW. Information clustering which means both term clustering and document clustering is considered. Three methods of the hard c-means, fuzzy c-means, and an agglomerative method using cluster centers are proposed. Two distances between fuzzy multisets and algorithms for calculating cluster centers are defined. Theoretical properties concerning the clustering algorithms are studied. Illustrative examples are given to show how the algorithms work.
    Source
    Information processing and management. 39(2003) no.2, S.195-213
  20. Ingwersen, P.; Wormell, I.: Ranganathan in the perspective of advanced information retrieval (1992) 0.01
    0.013766305 = product of:
      0.05506522 = sum of:
        0.013873219 = weight(_text_:information in 7695) [ClassicSimilarity], result of:
          0.013873219 = score(doc=7695,freq=4.0), product of:
            0.06322253 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.036014426 = queryNorm
            0.21943474 = fieldWeight in 7695, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=7695)
        0.041192 = weight(_text_:retrieval in 7695) [ClassicSimilarity], result of:
          0.041192 = score(doc=7695,freq=4.0), product of:
            0.10894058 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.036014426 = queryNorm
            0.37811437 = fieldWeight in 7695, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=7695)
      0.25 = coord(2/8)
    
    Abstract
    Examnines Ranganathan's approach to knowledge organisation and its relevance to intellectual accessibility in libraries. Discusses the current and future developments of his methodology and theories in knowledge-based systems. Topics covered include: semi-automatic classification and structure of thesauri; user-intermediary interactions in information retrieval (IR); semantic value-theory and uncertainty principles in IR; and case grammar

Years

Languages

  • e 146
  • d 25
  • a 1
  • chi 1
  • More… Less…

Types

  • a 150
  • el 20
  • x 5
  • m 3
  • r 2
  • s 2
  • d 1
  • More… Less…