Search (37 results, page 1 of 2)

  • × theme_ss:"Klassifikationssysteme im Online-Retrieval"
  1. Hill, J.S.: Online classification number access : some practical considerations (1984) 0.02
    0.017032763 = product of:
      0.051098287 = sum of:
        0.051098287 = product of:
          0.102196574 = sum of:
            0.102196574 = weight(_text_:22 in 7684) [ClassicSimilarity], result of:
              0.102196574 = score(doc=7684,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.61904186 = fieldWeight in 7684, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=7684)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    Journal of academic librarianship. 10(1984), S.17-22
  2. Ménard, E.; Mas, S.; Alberts, I.: Faceted classification for museum artefacts : a methodology to support web site development of large cultural organizations (2010) 0.02
    0.015756248 = product of:
      0.04726874 = sum of:
        0.04726874 = product of:
          0.09453748 = sum of:
            0.09453748 = weight(_text_:methodology in 3945) [ClassicSimilarity], result of:
              0.09453748 = score(doc=3945,freq=10.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.44516024 = fieldWeight in 3945, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3945)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - This research project aims to provide a new visual representation of the Artefacts Canada digital collection, as well as a means for users to browse this content. Artefacts Canada Humanities is a database containing approximately 3.5 million records describing the different collections of Canadian museums. Design/methodology/approach - A four-step methodology was adopted for the development of the faceted taxonomy model. First, a best practice review consisting of an extensive analysis of existing terminology standards in museum communities and public web interfaces of large cultural organizations was performed. The second step of the methodology entailed a domain analysis; this involved extracting and comparing relevant concepts from terminological authoritative sources. The third step proceeded to term clustering and entity listing,which involved the breaking-up of the taxonomy domains into potential facets. An incremental user testing was also realized in order to validate and refine the taxonomy components (facets, values, and relationships). Findings - The project resulted in a bilingual and expandable vocabulary structure that will further be used to describe the Artefacts Canada database records. The new taxonomy simplifies the representation of complex content by grouping objects into similar facets to classify all records of the Artefacts Canada database. The user-friendly bilingual taxonomy provides worldwide visitors with the means to better access Canadian virtual museum collections. Originality/value - Few methodological tools are available for museums which wish to adopt a faceted approach in the development of their web sites. For practitioners, the methodology developed within this project is a direct contribution to support web site development of large cultural organizations.
  3. Lim, E.: Southeast Asian subject gateways : an examination of their classification practices (2000) 0.01
    0.012774572 = product of:
      0.038323715 = sum of:
        0.038323715 = product of:
          0.07664743 = sum of:
            0.07664743 = weight(_text_:22 in 6040) [ClassicSimilarity], result of:
              0.07664743 = score(doc=6040,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.46428138 = fieldWeight in 6040, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6040)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 6.2002 19:42:47
  4. Golub, K.; Lykke, M.: Automated classification of web pages in hierarchical browsing (2009) 0.01
    0.012456408 = product of:
      0.03736922 = sum of:
        0.03736922 = product of:
          0.07473844 = sum of:
            0.07473844 = weight(_text_:methodology in 3614) [ClassicSimilarity], result of:
              0.07473844 = score(doc=3614,freq=4.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.35193008 = fieldWeight in 3614, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3614)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The purpose of this study is twofold: to investigate whether it is meaningful to use the Engineering Index (Ei) classification scheme for browsing, and then, if proven useful, to investigate the performance of an automated classification algorithm based on the Ei classification scheme. Design/methodology/approach - A user study was conducted in which users solved four controlled searching tasks. The users browsed the Ei classification scheme in order to examine the suitability of the classification systems for browsing. The classification algorithm was evaluated by the users who judged the correctness of the automatically assigned classes. Findings - The study showed that the Ei classification scheme is suited for browsing. Automatically assigned classes were on average partly correct, with some classes working better than others. Success of browsing showed to be correlated and dependent on classification correctness. Research limitations/implications - Further research should address problems of disparate evaluations of one and the same web page. Additional reasons behind browsing failures in the Ei classification scheme also need further investigation. Practical implications - Improvements for browsing were identified: describing class captions and/or listing their subclasses from start; allowing for searching for words from class captions with synonym search (easily provided for Ei since the classes are mapped to thesauri terms); when searching for class captions, returning the hierarchical tree expanded around the class in which caption the search term is found. The need for improvements of classification schemes was also indicated. Originality/value - A user-based evaluation of automated subject classification in the context of browsing has not been conducted before; hence the study also presents new findings concerning methodology.
  5. Frost, C.O.; Janes, J.: ¬An empirical test of gopher searching using three organizational schemes : background and methods (1994) 0.01
    0.012331214 = product of:
      0.03699364 = sum of:
        0.03699364 = product of:
          0.07398728 = sum of:
            0.07398728 = weight(_text_:methodology in 3031) [ClassicSimilarity], result of:
              0.07398728 = score(doc=3031,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.348393 = fieldWeight in 3031, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3031)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This paper describes the background underlying and the methods to be used in an investigation of use of the gopher protocol for browsing and searching for information via the Internet. Networked information resources are notoriously difficult to find; gopher is one attempt to facilitate the processes of organization and retrieval in that environment. Our project will design classification schemes based on traditional library models (i.e. Dewey and Library of Congress) and compare use of these to use of an existing gopher. We describe the project, the research questions, literature on related issues, the construction of the classification schemes, and the experimental methodology used
  6. Sparck Jones, K.: Revisiting classification for retrieval (2005) 0.01
    0.012331214 = product of:
      0.03699364 = sum of:
        0.03699364 = product of:
          0.07398728 = sum of:
            0.07398728 = weight(_text_:methodology in 4328) [ClassicSimilarity], result of:
              0.07398728 = score(doc=4328,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.348393 = fieldWeight in 4328, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4328)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - This short note seeks to respond to Hjørland and Pederson's paper "A substantive theory of classification for information retrieval" which starts from Sparck Jones's, "Some thoughts on classification for retrieval", originally published in 1970. Design/methodology/approach - The note comments on the context in which the 1970 paper was written, and on Hjørland and Pedersen's views, emphasising the need for well-grounded classification theory and application. Findings - The note maintains that text-based, a posteriori, classification, as increasingly found in applications, is likely to be more useful, in general, than a priori classification. Originality/value - The note elaborates on points made in a well-received earlier paper.
  7. Broughton, V.; Slavic, A.: Building a faceted classification for the humanities : principles and procedures (2007) 0.01
    0.012204738 = product of:
      0.036614213 = sum of:
        0.036614213 = product of:
          0.07322843 = sum of:
            0.07322843 = weight(_text_:methodology in 2875) [ClassicSimilarity], result of:
              0.07322843 = score(doc=2875,freq=6.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.34481966 = fieldWeight in 2875, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2875)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - This paper aims to provide an overview of principles and procedures involved in creating a faceted classification scheme for use in resource discovery in an online environment. Design/methodology/approach - Facet analysis provides an established rigorous methodology for the conceptual organization of a subject field, and the structuring of an associated classification or controlled vocabulary. This paper explains how that methodology was applied to the humanities in the FATKS project, where the objective was to explore the potential of facet analytical theory for creating a controlled vocabulary for the humanities, and to establish the requirements of a faceted classification appropriate to an online environment. A detailed faceted vocabulary was developed for two areas of the humanities within a broader facet framework for the whole of knowledge. Research issues included how to create a data model which made the faceted structure explicit and machine-readable and provided for its further development and use. Findings - In order to support easy facet combination in indexing, and facet searching and browsing on the interface, faceted classification requires a formalized data structure and an appropriate tool for its management. The conceptual framework of a faceted system proper can be applied satisfactorily to humanities, and fully integrated within a vocabulary management system. Research limitations/implications - The procedures described in this paper are concerned only with the structuring of the classification, and do not extend to indexing, retrieval and application issues. Practical implications - Many stakeholders in the domain of resource discovery consider developing their own classification system and supporting tools. The methods described in this paper may clarify the process of building a faceted classification and may provide some useful ideas with respect to the vocabulary maintenance tool. Originality/value - As far as the authors are aware there is no comparable research in this area.
  8. Comaromi, C.L.: Summation of classification as an enhancement of intellectual access to information in an online environment (1990) 0.01
    0.010645477 = product of:
      0.03193643 = sum of:
        0.03193643 = product of:
          0.06387286 = sum of:
            0.06387286 = weight(_text_:22 in 3576) [ClassicSimilarity], result of:
              0.06387286 = score(doc=3576,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.38690117 = fieldWeight in 3576, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3576)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    8. 1.2007 12:22:40
  9. Reiner, U.: Automatische DDC-Klassifizierung von bibliografischen Titeldatensätzen (2009) 0.01
    0.010645477 = product of:
      0.03193643 = sum of:
        0.03193643 = product of:
          0.06387286 = sum of:
            0.06387286 = weight(_text_:22 in 611) [ClassicSimilarity], result of:
              0.06387286 = score(doc=611,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.38690117 = fieldWeight in 611, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=611)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 8.2009 12:54:24
  10. Neelameghan, A.: S.R. Ranganathan's general theory of knowledge classification in designing, indexing and retrieving from specialised databases (1997) 0.01
    0.010569612 = product of:
      0.031708833 = sum of:
        0.031708833 = product of:
          0.063417666 = sum of:
            0.063417666 = weight(_text_:methodology in 3) [ClassicSimilarity], result of:
              0.063417666 = score(doc=3,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.29862255 = fieldWeight in 3, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Summarizes some experiences of the application of the priciples and postulates of S.R. Ranganathan's General Theory of Knowledge Classification, incorporating the freely faceted approach and analytico synthetic methods, to the design and development of specialized databases, including indexing, user interfaces and retrieval. Enumerates some of the earlier instances of the facet method in machine based systems, beginning with Hollerith's punched card system for the data processing of the US Census. Elaborates on Ranganathan's holistic approach to information systems and services provided by his normative principles. Notes similarities between the design of databases and faceted classification systems. Examples from working systems are given to demonstrate the usefulness of selected canons and principles of classification and the analytico synthetic methodology to database design. The examples are mostly operational database systems developed using Unesco's Micro CDS-ISIS software
  11. Williamson, N.J.: ¬The Library of Congress Classification in the computer age (1989) 0.01
    0.010569612 = product of:
      0.031708833 = sum of:
        0.031708833 = product of:
          0.063417666 = sum of:
            0.063417666 = weight(_text_:methodology in 2074) [ClassicSimilarity], result of:
              0.063417666 = score(doc=2074,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.29862255 = fieldWeight in 2074, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2074)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Research leading to a detailed content analysis and profile of the Library of Congress Classification schedules. A description of the design, methodology, and preliminary findings of research in progress. Final results of the study are expected to aid the Library of Congress in the development of a MARC format for its classification schedules and in the conversion of the schedules into machine-readable form. LC's ultimate goal is the design of an integrated system which will permit on-line editing, maintenance, and control of the schedules, on-line classification and shelflisting procedures, the production of the schedules in two or more physical formats, and the use of LCC to enhance search capabilities in on-line catalogues.
  12. Lösse, M.; Svensson, L.: "Classification at a Crossroad" : Internationales UDC-Seminar 2009 in Den Haag, Niederlande (2010) 0.01
    0.009032987 = product of:
      0.027098961 = sum of:
        0.027098961 = product of:
          0.054197922 = sum of:
            0.054197922 = weight(_text_:22 in 4379) [ClassicSimilarity], result of:
              0.054197922 = score(doc=4379,freq=4.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.32829654 = fieldWeight in 4379, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4379)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Am 29. und 30. Oktober 2009 fand in der Königlichen Bibliothek in Den Haag das zweite internationale UDC-Seminar zum Thema "Classification at a Crossroad" statt. Organisiert wurde diese Konferenz - wie auch die erste Konferenz dieser Art im Jahr 2007 - vom UDC-Konsortium (UDCC). Im Mittelpunkt der diesjährigen Veranstaltung stand die Erschließung des World Wide Web unter besserer Nutzung von Klassifikationen (im Besonderen natürlich der UDC), einschließlich benutzerfreundlicher Repräsentationen von Informationen und Wissen. Standards, neue Technologien und Dienste, semantische Suche und der multilinguale Zugriff spielten ebenfalls eine Rolle. 135 Teilnehmer aus 35 Ländern waren dazu nach Den Haag gekommen. Das Programm umfasste mit 22 Vorträgen aus 14 verschiedenen Ländern eine breite Palette, wobei Großbritannien mit fünf Beiträgen am stärksten vertreten war. Die Tagesschwerpunkte wurden an beiden Konferenztagen durch die Eröffnungsvorträge gesetzt, die dann in insgesamt sechs thematischen Sitzungen weiter vertieft wurden.
    Date
    22. 1.2010 15:06:54
  13. Sparck Jones, K.: Some thoughts on classification for retrieval (2005) 0.01
    0.00880801 = product of:
      0.02642403 = sum of:
        0.02642403 = product of:
          0.05284806 = sum of:
            0.05284806 = weight(_text_:methodology in 4392) [ClassicSimilarity], result of:
              0.05284806 = score(doc=4392,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.24885213 = fieldWeight in 4392, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4392)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - This paper was originally published in 1970 (Journal of documentation. 26(1970), S.89-101), considered the suggestion that classifications for retrieval should be constructed automatically and raised some serious problems concerning the sorts of classification which were required, and the way in which formal classification theories should be exploited, given that a retrieval classification is required for a purpose. These difficulties had not been sufficiently considered, and the paper, therefore, aims to attempt an analysis of them, though no solutions of immediate application could be suggested. Design/methodology/approach - Starting with the illustrative proposition that a polythetic, multiple, unordered classification is required in automatic thesaurus construction, this is considered in the context of classification in general, where eight sorts of classification can be distinguished, each covering a range of class definitions and class-finding algorithms. Findings - Since there is generally no natural or best classification of a set of objects as such, the evaluation of alternative classifications requires either formal criteria of goodness of fit, or, if a classification is required for a purpose, a precise statement of that purpose. In any case a substantive theory of classification is needed, which does not exist; and, since sufficiently precise specifications of retrieval requirements are also lacking, the only currently available approach to automatic classification experiments for information retrieval is to do enough of them. Originality/value - Gives insights into the classification of material for information retrieval.
  14. Hjoerland, B.; Pedersen, K.N.: ¬A substantive theory of classification for information retrieval (2005) 0.01
    0.00880801 = product of:
      0.02642403 = sum of:
        0.02642403 = product of:
          0.05284806 = sum of:
            0.05284806 = weight(_text_:methodology in 1892) [ClassicSimilarity], result of:
              0.05284806 = score(doc=1892,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.24885213 = fieldWeight in 1892, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1892)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - To suggest that a theory of classification for information retrieval (IR), asked for by Spärck Jones in a 1970 paper, presupposes a full implementation of a pragmatic understanding. Part of the Journal of Documentation celebration, "60 years of the best in information research". Design/methodology/approach - Literature-based conceptual analysis, taking Sparck Jones as its starting-point. Analysis involves distinctions between "positivism" and "pragmatism" and "classical" versus Kuhnian understandings of concepts. Findings - Classification, both manual and automatic, for retrieval benefits from drawing upon a combination of qualitative and quantitative techniques, a consideration of theories of meaning, and the adding of top-down approaches to IR in which divisions of labour, domains, traditions, genres, document architectures etc. are included as analytical elements and in which specific IR algorithms are based on the examination of specific literatures. Introduces an example illustrating the consequences of a full implementation of a pragmatist understanding when handling homonyms. Practical implications - Outlines how to classify from a pragmatic-philosophical point of view. Originality/value - Provides, emphasizing a pragmatic understanding, insights of importance to classification for retrieval, both manual and automatic. - Vgl. auch: Szostak, R.: Classification, interdisciplinarity, and the study of science. In: Journal of documentation. 64(2008) no.3, S.319-332.
  15. Lima, G.A. de; Castro, I.R.: Uso da classificacao decimal universal para a recuperacao da informacao em ambientes digitas : uma revisao sistematica da literatura (2021) 0.01
    0.00880801 = product of:
      0.02642403 = sum of:
        0.02642403 = product of:
          0.05284806 = sum of:
            0.05284806 = weight(_text_:methodology in 760) [ClassicSimilarity], result of:
              0.05284806 = score(doc=760,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.24885213 = fieldWeight in 760, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=760)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Knowledge Organization Systems, even traditional ones, such as the Universal Decimal Classification, have been studied to improve the retrieval of information online, although the potential of using knowledge structures in the user interface has not yet been widespread. Objective: This study presents a mapping of scientific production on information retrieval methodologies, which make use of the Universal Decimal Classification. Methodology: Systematic Literature Review, conducted in two stages, with a selection of 44 publications, resulting in the time interval from 1964 to 2017, whose categories analyzed were: most productive authors, languages of publications, types of document, year of publication, most cited work, major impact journal, and thematic categories covered in the publications. Results: A total of nine more productive authors and co-authors were found; predominance of the English language (42 publications); works published in the format of journal articles (33); and highlight to the year 2007 (eight publications). In addition, it was identified that the most cited work was by Mcilwaine (1997), with 61 citations, and the journal Extensions & Corrections to the UDC was the one with the largest number of publications, in addition to the incidence of the theme Universal Automation linked to a thesaurus for information retrieval, present in 19 works. Conclusions: Shortage of studies that explore the potential of the Decimal Classification, especially in Brazilian literature, which highlights the need for further study on the topic, involving research at the national and international levels.
  16. Doyle, B.: ¬The classification and evaluation of Content Management Systems (2003) 0.01
    0.0085163815 = product of:
      0.025549144 = sum of:
        0.025549144 = product of:
          0.051098287 = sum of:
            0.051098287 = weight(_text_:22 in 2871) [ClassicSimilarity], result of:
              0.051098287 = score(doc=2871,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.30952093 = fieldWeight in 2871, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2871)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    30. 7.2004 12:22:52
  17. Peereboom, M.: DutchESS : Dutch Electronic Subject Service - a Dutch national collaborative effort (2000) 0.01
    0.0085163815 = product of:
      0.025549144 = sum of:
        0.025549144 = product of:
          0.051098287 = sum of:
            0.051098287 = weight(_text_:22 in 4869) [ClassicSimilarity], result of:
              0.051098287 = score(doc=4869,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.30952093 = fieldWeight in 4869, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4869)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 6.2002 19:39:23
  18. Van Dijck, P.: Introduction to XFML (2003) 0.01
    0.0085163815 = product of:
      0.025549144 = sum of:
        0.025549144 = product of:
          0.051098287 = sum of:
            0.051098287 = weight(_text_:22 in 2474) [ClassicSimilarity], result of:
              0.051098287 = score(doc=2474,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.30952093 = fieldWeight in 2474, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2474)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    http://www.xml.com/lpt/a/2003/01/22/xfml.html
  19. Dack, D.: Australian attends conference on Dewey (1989) 0.01
    0.0074518337 = product of:
      0.0223555 = sum of:
        0.0223555 = product of:
          0.044711 = sum of:
            0.044711 = weight(_text_:22 in 2509) [ClassicSimilarity], result of:
              0.044711 = score(doc=2509,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.2708308 = fieldWeight in 2509, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2509)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    8.11.1995 11:52:22
  20. Vizine-Goetz, D.: OCLC investigates using classification tools to organize Internet data (1998) 0.01
    0.0074518337 = product of:
      0.0223555 = sum of:
        0.0223555 = product of:
          0.044711 = sum of:
            0.044711 = weight(_text_:22 in 2342) [ClassicSimilarity], result of:
              0.044711 = score(doc=2342,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.2708308 = fieldWeight in 2342, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2342)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 9.1997 19:16:05