Search (82 results, page 1 of 5)

  • × theme_ss:"Automatisches Indexieren"
  1. Groß, T.: Automatische Indexierung von wirtschaftswissenschaftlichen Dokumenten : Implementierung und Evaluierung am Beispiel der Deutschen Zentralbibliothek für Wirtschaftswissenschaften (2010) 0.01
    0.0061221616 = product of:
      0.030610807 = sum of:
        0.010176711 = product of:
          0.03053013 = sum of:
            0.03053013 = weight(_text_:problem in 2078) [ClassicSimilarity], result of:
              0.03053013 = score(doc=2078,freq=2.0), product of:
                0.1302053 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.03067635 = queryNorm
                0.23447686 = fieldWeight in 2078, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2078)
          0.33333334 = coord(1/3)
        0.020434096 = product of:
          0.06130229 = sum of:
            0.06130229 = weight(_text_:2010 in 2078) [ClassicSimilarity], result of:
              0.06130229 = score(doc=2078,freq=5.0), product of:
                0.14672957 = queryWeight, product of:
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.03067635 = queryNorm
                0.41779095 = fieldWeight in 2078, product of:
                  2.236068 = tf(freq=5.0), with freq of:
                    5.0 = termFreq=5.0
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2078)
          0.33333334 = coord(1/3)
      0.2 = coord(2/10)
    
    Abstract
    Die Bewertung der Indexierungsqualität bzw. -güte ist ein grundlegendes Problem von manuellen und automatischen Indexierungsverfahren. Letztere werden aber gerade im digitalen Zeitalter als einzige Möglichkeit angesehen, den zunehmenden Schwierigkeiten bibliothekarischer Informationsstrukturierung gerecht zu werden. Diese Arbeit befasst sich mit der Funktionsweise, Implementierung und Evaluierung der Sacherschließungssoftware MindServer Categorizer, der Firma Recommind, an der Deutschen Zentralbibliothek für Wirtschaftswissenschaften (ZBW). Grundlage der maschinellen Sacherschließung und anschließenden quantitativen und qualitativen Auswertung bilden rund 39.000 wirtschaftswissenschaftliche Dokumente aus den Datenbanken Econis und EconStor. Unter Zuhilfenahme des rund 6.000 Deskriptoren umfassenden Standard-Thesaurus Wirtschaft (STW) wird der ursprünglich rein statistische Indexierungsansatz des MindServer Categorizer zu einem begriffsorientierten Verfahren weiterentwickelt und zur Inhaltserschließung digitaler Informationsressourcen eingesetzt. Der zentrale Fokus dieser Arbeit liegt vor allem auf der Evaluierung der maschinell beschlagworteten Titel, in Anlehnung und entsprechender Anpassung der von Stock (2008) und Lancaster (2003) hierzu vorgeschlagenen Kriterien: Indexierungskonsistenz, -tiefe, -breite, -spezifität, -effektivität. Zusätzlich wird die Belegungsbilanz des STW evaluiert und es erfolgt ferner eine stichprobenartige, qualitative Bewertung der Ergebnisse seitens der zuständigen Fachreferenten und -referentinnen.
    Content
    Vgl. unter: http://edoc.hu-berlin.de/series/berliner-handreichungen/2010-284/PDF/284.pdf. Auch als: Automatische Indexierung von Dokumenten in einer wissenschaftlichen Bibliothek: Implementierung und Evaluierung am Beispiel der Deutschen Zentralbibliothek für Wirtschaftswissenschaften. Diplomica Verlag, 2011.
    Year
    2010
  2. Jones, K.P.: Natural-language processing and automatic indexing : a reply (1990) 0.01
    0.005805583 = product of:
      0.058055833 = sum of:
        0.058055833 = product of:
          0.1741675 = sum of:
            0.1741675 = weight(_text_:1990 in 394) [ClassicSimilarity], result of:
              0.1741675 = score(doc=394,freq=5.0), product of:
                0.13825724 = queryWeight, product of:
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.03067635 = queryNorm
                1.2597351 = fieldWeight in 394, product of:
                  2.236068 = tf(freq=5.0), with freq of:
                    5.0 = termFreq=5.0
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.125 = fieldNorm(doc=394)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Indexer. 17(1990), S.114-115
    Year
    1990
  3. Correa, C.A.; Kobashi, N.Y.: ¬A hybrid model of automatic indexing based on paraconsitent logic 0.01
    0.0055440883 = product of:
      0.027720442 = sum of:
        0.012212053 = product of:
          0.03663616 = sum of:
            0.03663616 = weight(_text_:problem in 3537) [ClassicSimilarity], result of:
              0.03663616 = score(doc=3537,freq=2.0), product of:
                0.1302053 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.03067635 = queryNorm
                0.28137225 = fieldWeight in 3537, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3537)
          0.33333334 = coord(1/3)
        0.015508389 = product of:
          0.046525165 = sum of:
            0.046525165 = weight(_text_:2010 in 3537) [ClassicSimilarity], result of:
              0.046525165 = score(doc=3537,freq=2.0), product of:
                0.14672957 = queryWeight, product of:
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.03067635 = queryNorm
                0.31708103 = fieldWeight in 3537, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3537)
          0.33333334 = coord(1/3)
      0.2 = coord(2/10)
    
    Abstract
    In the processes of information organization, information retrieval and information visualization one point in common can be found: they are strongly connected by the procedures associated to the indexing of texts or documents. Indexing is an essential component of text analysis, and the indexing process have equal importance both for retrieval and for the visualization of information. In this context, it is worth mentioning the solutions that use automatic indexing. The researches that proposes solutions for automatic indexing are based on different theoretical assumptions such as: statistics, linguistics and controlled vocabulary (Leiva 1999). Most solutions developed hybrid models combining these theoretical assumptions. Other solutions to the problem of automatic indexing are based on theories that allow the treatment of uncertainty, imprecision and vagueness. The aim of this paper is to argue the theoretical potential for use in hybrid models of automatic indexing, the paraconsistent logic, a non-classical logic, with capacity to handle situations that involve uncertainty, imprecision and vagueness.
    Source
    Paradigms and conceptual systems in knowledge organization: Proceedings of the Eleventh International ISKO conference, Rome, 23-26 February 2010, ed. Claudio Gnoli, Indeks, Frankfurt M
  4. Bordoni, L.; Pazienza, M.T.: Documents automatic indexing in an environmental domain (1997) 0.00
    0.0047890493 = product of:
      0.023945246 = sum of:
        0.014247394 = product of:
          0.04274218 = sum of:
            0.04274218 = weight(_text_:problem in 530) [ClassicSimilarity], result of:
              0.04274218 = score(doc=530,freq=2.0), product of:
                0.1302053 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.03067635 = queryNorm
                0.3282676 = fieldWeight in 530, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=530)
          0.33333334 = coord(1/3)
        0.009697851 = product of:
          0.029093552 = sum of:
            0.029093552 = weight(_text_:22 in 530) [ClassicSimilarity], result of:
              0.029093552 = score(doc=530,freq=2.0), product of:
                0.10742335 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03067635 = queryNorm
                0.2708308 = fieldWeight in 530, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=530)
          0.33333334 = coord(1/3)
      0.2 = coord(2/10)
    
    Abstract
    Describes an application of Natural Language Processing (NLP) techniques, in HIRMA (Hypertextual Information Retrieval Managed by ARIOSTO), to the problem of document indexing by referring to a system which incorporates natural language processing techniques to determine the subject of the text of documents and to associate them with relevant semantic indexes. Describes briefly the overall system, details of its implementation on a corpus of scientific abstracts related to environmental topics and experimental evidence of the system's behaviour. Analyzes in detail an experiment designed to evaluate the system's retrieval ability in terms of recall and precision
    Source
    International forum on information and documentation. 22(1997) no.1, S.17-28
  5. Groß, T.: Automatische Indexierung von Dokumenten in einer wissenschaftlichen Bibliothek : Implementierung und Evaluierung am Beispiel der Deutschen Zentralbibliothek für Wirtschaftswissenschaften (2011) 0.00
    0.004620074 = product of:
      0.023100369 = sum of:
        0.010176711 = product of:
          0.03053013 = sum of:
            0.03053013 = weight(_text_:problem in 1083) [ClassicSimilarity], result of:
              0.03053013 = score(doc=1083,freq=2.0), product of:
                0.1302053 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.03067635 = queryNorm
                0.23447686 = fieldWeight in 1083, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1083)
          0.33333334 = coord(1/3)
        0.012923657 = product of:
          0.03877097 = sum of:
            0.03877097 = weight(_text_:2010 in 1083) [ClassicSimilarity], result of:
              0.03877097 = score(doc=1083,freq=2.0), product of:
                0.14672957 = queryWeight, product of:
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.03067635 = queryNorm
                0.2642342 = fieldWeight in 1083, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1083)
          0.33333334 = coord(1/3)
      0.2 = coord(2/10)
    
    Abstract
    Die Bewertung der Indexierungsqualität bzw. -güte ist ein grundlegendes Problem von intellektuellen und automatischen Indexierungsverfahren. Letztere werden aber gerade im digitalen Zeitalter als einzige Möglichkeit angesehen, den zunehmenden Schwierigkeiten bibliothekarischer Informationsstrukturierung gerecht zu werden. Diese Studie befasst sich mit der Funktionsweise, Implementierung und Evaluierung der Sacherschließungssoftware MindServer Categorizer der Firma Recommind an der Deutschen Zentralbibliothek für Wirtschaftswissenschaften. Grundlage der maschinellen Sacherschließung und anschließenden quantitativen und qualitativen Auswertung bilden rund 39.000 wirtschaftswissenschaftliche Dokumente aus den Datenbanken Econis und EconStor. Unter Zuhilfenahme des rund 6.000 Schlagwörter umfassenden Standard-Thesaurus Wirtschaft wird der ursprünglich rein statistische Indexierungsansatz des MindServer Categorizer zu einem begriffsorientierten Verfahren weiterentwickelt und zur Inhaltserschließung digitaler Informationsressourcen eingesetzt. Der zentrale Fokus dieser Studie liegt vor allem auf der Evaluierung der maschinell beschlagworteten Titel, in Anlehnung an die hierzu von Stock und Lancaster vorgeschlagenen Kriterien: Indexierungskonsistenz, -tiefe, -breite, -spezifität, -effektivität. Weiterhin wird die Belegungsbilanz des STW evaluiert und es erfolgt zusätzlich eine qualitative, stichprobenartige Bewertung der Ergebnisse seitens der zuständigen Fachreferenten und -referentinnen.
    Content
    Vgl. die zugrunde liegende Masterarbeit unter: http://edoc.hu-berlin.de/series/berliner-handreichungen/2010-284/PDF/284.pdf.
  6. Plaunt, C.; Norgard, B.A.: ¬An association-based method for automatic indexing with a controlled vocabulary (1998) 0.00
    0.004263816 = product of:
      0.02131908 = sum of:
        0.0143920425 = product of:
          0.043176126 = sum of:
            0.043176126 = weight(_text_:problem in 1794) [ClassicSimilarity], result of:
              0.043176126 = score(doc=1794,freq=4.0), product of:
                0.1302053 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.03067635 = queryNorm
                0.33160037 = fieldWeight in 1794, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1794)
          0.33333334 = coord(1/3)
        0.0069270367 = product of:
          0.02078111 = sum of:
            0.02078111 = weight(_text_:22 in 1794) [ClassicSimilarity], result of:
              0.02078111 = score(doc=1794,freq=2.0), product of:
                0.10742335 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03067635 = queryNorm
                0.19345059 = fieldWeight in 1794, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1794)
          0.33333334 = coord(1/3)
      0.2 = coord(2/10)
    
    Abstract
    In this article, we describe and test a two-stage algorithm based on a lexical collocation technique which maps from the lexical clues contained in a document representation into a controlled vocabulary list of subject headings. Using a collection of 4.626 INSPEC documents, we create a 'dictionary' of associations between the lexical items contained in the titles, authors, and abstracts, and controlled vocabulary subject headings assigned to those records by human indexers using a likelihood ratio statistic as the measure of association. In the deployment stage, we use the dictiony to predict which of the controlled vocabulary subject headings best describe new documents when they are presented to the system. Our evaluation of this algorithm, in which we compare the automatically assigned subject headings to the subject headings assigned to the test documents by human catalogers, shows that we can obtain results comparable to, and consistent with, human cataloging. In effect we have cast this as a classic partial match information retrieval problem. We consider the problem to be one of 'retrieving' (or assigning) the most probably 'relevant' (or correct) controlled vocabulary subject headings to a document based on the clues contained in that document
    Date
    11. 9.2000 19:53:22
  7. Garfield, E.: KeyWords Plus takes you beyond title words (1990) 0.00
    0.003934862 = product of:
      0.03934862 = sum of:
        0.03934862 = product of:
          0.11804586 = sum of:
            0.11804586 = weight(_text_:1990 in 4344) [ClassicSimilarity], result of:
              0.11804586 = score(doc=4344,freq=3.0), product of:
                0.13825724 = queryWeight, product of:
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.03067635 = queryNorm
                0.85381323 = fieldWeight in 4344, product of:
                  1.7320508 = tf(freq=3.0), with freq of:
                    3.0 = termFreq=3.0
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4344)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Year
    1990
  8. Garfield, E.: KeyWords Plus : ISI's breakthrough retrieval method (1990) 0.00
    0.003934862 = product of:
      0.03934862 = sum of:
        0.03934862 = product of:
          0.11804586 = sum of:
            0.11804586 = weight(_text_:1990 in 4345) [ClassicSimilarity], result of:
              0.11804586 = score(doc=4345,freq=3.0), product of:
                0.13825724 = queryWeight, product of:
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.03067635 = queryNorm
                0.85381323 = fieldWeight in 4345, product of:
                  1.7320508 = tf(freq=3.0), with freq of:
                    3.0 = termFreq=3.0
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4345)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Year
    1990
  9. Salton, G.: Future prospects for text-based information retrieval (1990) 0.00
    0.0033727388 = product of:
      0.03372739 = sum of:
        0.03372739 = product of:
          0.10118216 = sum of:
            0.10118216 = weight(_text_:1990 in 2327) [ClassicSimilarity], result of:
              0.10118216 = score(doc=2327,freq=3.0), product of:
                0.13825724 = queryWeight, product of:
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.03067635 = queryNorm
                0.7318399 = fieldWeight in 2327, product of:
                  1.7320508 = tf(freq=3.0), with freq of:
                    3.0 = termFreq=3.0
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2327)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Year
    1990
  10. Salton, G.; Araya, J.: On the use of clustered file organizations in information search and retrieval (1990) 0.00
    0.0033727388 = product of:
      0.03372739 = sum of:
        0.03372739 = product of:
          0.10118216 = sum of:
            0.10118216 = weight(_text_:1990 in 2409) [ClassicSimilarity], result of:
              0.10118216 = score(doc=2409,freq=3.0), product of:
                0.13825724 = queryWeight, product of:
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.03067635 = queryNorm
                0.7318399 = fieldWeight in 2409, product of:
                  1.7320508 = tf(freq=3.0), with freq of:
                    3.0 = termFreq=3.0
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2409)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Year
    1990
  11. Williams, R.V.: Hans Peter Luhn and Herbert M. Ohlman : their roles in the origins of keyword-in-context/permutation automatic indexing (2010) 0.00
    0.0032694556 = product of:
      0.032694556 = sum of:
        0.032694556 = product of:
          0.09808366 = sum of:
            0.09808366 = weight(_text_:2010 in 3440) [ClassicSimilarity], result of:
              0.09808366 = score(doc=3440,freq=5.0), product of:
                0.14672957 = queryWeight, product of:
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.03067635 = queryNorm
                0.6684655 = fieldWeight in 3440, product of:
                  2.236068 = tf(freq=5.0), with freq of:
                    5.0 = termFreq=5.0
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3440)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.4, S.835-849
    Year
    2010
  12. Koryconski, C.; Newell, A.F.: Natural-language processing and automatic indexing (1990) 0.00
    0.0029027916 = product of:
      0.029027916 = sum of:
        0.029027916 = product of:
          0.08708375 = sum of:
            0.08708375 = weight(_text_:1990 in 2313) [ClassicSimilarity], result of:
              0.08708375 = score(doc=2313,freq=5.0), product of:
                0.13825724 = queryWeight, product of:
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.03067635 = queryNorm
                0.62986755 = fieldWeight in 2313, product of:
                  2.236068 = tf(freq=5.0), with freq of:
                    5.0 = termFreq=5.0
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2313)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Indexer. 17(1990), S.21-29
    Year
    1990
  13. Krause, J.; Womser-Hacker, C.: PADOK-II : Retrievaltests zur Bewertung von Volltextindexierungsvarianten für das deutsche Patentinformationssystem (1990) 0.00
    0.0029027916 = product of:
      0.029027916 = sum of:
        0.029027916 = product of:
          0.08708375 = sum of:
            0.08708375 = weight(_text_:1990 in 2653) [ClassicSimilarity], result of:
              0.08708375 = score(doc=2653,freq=5.0), product of:
                0.13825724 = queryWeight, product of:
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.03067635 = queryNorm
                0.62986755 = fieldWeight in 2653, product of:
                  2.236068 = tf(freq=5.0), with freq of:
                    5.0 = termFreq=5.0
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2653)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Nachrichten für Dokumentation. 41(1990) H.1, S.13-19
    Year
    1990
  14. Lepsky, K.; Müller, T.; Wille, J.: Metadata improvement for image information retrieval (2010) 0.00
    0.0028607734 = product of:
      0.028607734 = sum of:
        0.028607734 = product of:
          0.0858232 = sum of:
            0.0858232 = weight(_text_:2010 in 4995) [ClassicSimilarity], result of:
              0.0858232 = score(doc=4995,freq=5.0), product of:
                0.14672957 = queryWeight, product of:
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.03067635 = queryNorm
                0.5849073 = fieldWeight in 4995, product of:
                  2.236068 = tf(freq=5.0), with freq of:
                    5.0 = termFreq=5.0
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4995)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Paradigms and conceptual systems in knowledge organization: Proceedings of the Eleventh International ISKO Conference, 23-26 February 2010 Rome, Italy. Edited by Claudio Gnoli and Fulvio Mazzocchi
    Year
    2010
  15. Srinivasan, P.: On generalizing the Two-Poisson Model (1990) 0.00
    0.0025399427 = product of:
      0.025399428 = sum of:
        0.025399428 = product of:
          0.07619828 = sum of:
            0.07619828 = weight(_text_:1990 in 2880) [ClassicSimilarity], result of:
              0.07619828 = score(doc=2880,freq=5.0), product of:
                0.13825724 = queryWeight, product of:
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.03067635 = queryNorm
                0.5511341 = fieldWeight in 2880, product of:
                  2.236068 = tf(freq=5.0), with freq of:
                    5.0 = termFreq=5.0
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2880)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Journal of the American Society for Information Science. 41(1990) no.1, S.61-66
    Year
    1990
  16. Cui, H.; Boufford, D.; Selden, P.: Semantic annotation of biosystematics literature without training examples (2010) 0.00
    0.0024520915 = product of:
      0.024520915 = sum of:
        0.024520915 = product of:
          0.07356274 = sum of:
            0.07356274 = weight(_text_:2010 in 3422) [ClassicSimilarity], result of:
              0.07356274 = score(doc=3422,freq=5.0), product of:
                0.14672957 = queryWeight, product of:
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.03067635 = queryNorm
                0.5013491 = fieldWeight in 3422, product of:
                  2.236068 = tf(freq=5.0), with freq of:
                    5.0 = termFreq=5.0
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3422)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.3, S.522-542
    Year
    2010
  17. Salton, G.: Another look at automatic text-retrieval systems (1986) 0.00
    0.0022948582 = product of:
      0.022948582 = sum of:
        0.022948582 = product of:
          0.06884574 = sum of:
            0.06884574 = weight(_text_:1990 in 1356) [ClassicSimilarity], result of:
              0.06884574 = score(doc=1356,freq=2.0), product of:
                0.13825724 = queryWeight, product of:
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.03067635 = queryNorm
                0.497954 = fieldWeight in 1356, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1356)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Footnote
    Bezugnahme auf: Blair, D.C.: An evaluation of retrieval effectiveness for a full-text document-retrieval system. Comm. ACM 28(1985) S.280-299. - Vgl. auch: Blair, D.C.: Full text retrieval ... Int. Class. 13(1986) S.18-23; Blair, D.C., M.E. Maron: full-text information retrieval ... Inf. Proc. Man. 26(1990) S.437-447.
  18. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.00
    0.0022166518 = product of:
      0.022166518 = sum of:
        0.022166518 = product of:
          0.06649955 = sum of:
            0.06649955 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.06649955 = score(doc=402,freq=2.0), product of:
                0.10742335 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03067635 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  19. Malone, L.C.; Wildman-Pepe, J.; Driscoll, J.R.: Evaluation of an automated keywording system (1990) 0.00
    0.0021770939 = product of:
      0.021770937 = sum of:
        0.021770937 = product of:
          0.06531281 = sum of:
            0.06531281 = weight(_text_:1990 in 4999) [ClassicSimilarity], result of:
              0.06531281 = score(doc=4999,freq=5.0), product of:
                0.13825724 = queryWeight, product of:
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.03067635 = queryNorm
                0.47240067 = fieldWeight in 4999, product of:
                  2.236068 = tf(freq=5.0), with freq of:
                    5.0 = termFreq=5.0
                  4.506965 = idf(docFreq=1325, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4999)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Microcomputers for information management. 7(1990) no. 2, S.127-148
    Year
    1990
  20. Beyer, C.; Trunk, D.: Automatische Verfahren für die Formalerschließung im Projekt PETRUS (2011) 0.00
    0.0020677852 = product of:
      0.020677852 = sum of:
        0.020677852 = product of:
          0.062033553 = sum of:
            0.062033553 = weight(_text_:2010 in 1712) [ClassicSimilarity], result of:
              0.062033553 = score(doc=1712,freq=2.0), product of:
                0.14672957 = queryWeight, product of:
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.03067635 = queryNorm
                0.4227747 = fieldWeight in 1712, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.7831497 = idf(docFreq=1005, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1712)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Abstract
    Die Deutsche Nationalbibliothek (DNB) erstellt für Netzpublikationen seit Anfang 2010 keine intellektuell per Autopsie und nach gültigem Regelwerk (RAK-WB bzw. RAK-NBM) erschlossen Titeldaten mehr. Da das stetig wachsende Kontingent an Netzpublikationen durch intellektuelle Bearbeitung nicht mehr zu bewältigen ist, wurde mit dem Projekt PETRUS damit begonnen, die Erschließungsprozesse für Netzpublikationen zu automatisieren.

Years

Languages

  • e 53
  • d 27
  • f 1
  • ru 1
  • More… Less…

Types

  • a 73
  • x 4
  • el 3
  • m 2
  • s 1
  • More… Less…