Search (27 results, page 1 of 2)

  • × theme_ss:"Klassifikationssysteme im Online-Retrieval"
  • × year_i:[1990 TO 2000}
  1. Francu, V.: Building a multilingual thesaurus based on UDC (1996) 0.08
    0.07873234 = product of:
      0.11809851 = sum of:
        0.08966068 = weight(_text_:systematic in 7410) [ClassicSimilarity], result of:
          0.08966068 = score(doc=7410,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.31573826 = fieldWeight in 7410, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0390625 = fieldNorm(doc=7410)
        0.028437834 = product of:
          0.05687567 = sum of:
            0.05687567 = weight(_text_:indexing in 7410) [ClassicSimilarity], result of:
              0.05687567 = score(doc=7410,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29905218 = fieldWeight in 7410, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=7410)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Once the library has been through a process of transition from traditional library procedures to automated ones, natural language searching became a necessity for both indexers and searchers. Therefore, aside from the precoordinated classified catalogue we started to build a dictionary of terms in order to make postcoordinate search possible in keeping with the UDC notations assigned to each bibliographic record. After a while we came to the conclusion that the dictionary needed a control of its terms so that synonymous concepts and semantic ambuguities be avoided. The project presented in this paper shows how reality imposed the improvement of the quality of indexing and hence of the searching possibilities. Is also shows the reasons why we consider a multilingual thesaurus based on UDC an ideal indexing and searching device. The experiment applied on class 8 of UDC illustrates the way the UDC tables can be quite successfully used in building a thesaurus due to their qulities and how their limitations can be overcome by a thesaurus. An appendix to the paper contains a sample of the multilingual thesaurus given in both alphabetical and systematic layouts
  2. Pollitt, A.S.: ¬The application of Dewey Classification in a view-based searching OPAC (1998) 0.07
    0.07317951 = product of:
      0.10976927 = sum of:
        0.08966068 = weight(_text_:systematic in 73) [ClassicSimilarity], result of:
          0.08966068 = score(doc=73,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.31573826 = fieldWeight in 73, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0390625 = fieldNorm(doc=73)
        0.020108584 = product of:
          0.04021717 = sum of:
            0.04021717 = weight(_text_:indexing in 73) [ClassicSimilarity], result of:
              0.04021717 = score(doc=73,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.21146181 = fieldWeight in 73, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=73)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This paper examines issues relating to the use of the Dewey Decimal Classification (DDC) in a future development of view-based searching to Online Public Access Catalogues (OPAC). View-based searching systems, exercising the principles of fully faceted classification techniques for both bibliographic and corporate database retrieval applications, are now being applied to utilise Dewey concept hierarchies in a University OPAC. Issues of efficiency and effectiveness in the evolving organisation and classification of information within libraries are examined to explain why fully faceted classification schemes have yet to realise their full potential in libraries. The key to their application in OPACs lies in the use of faceted classification as pre-coordinated indexing and abandoning the single dimension relative ordering of books on shelves. The need to maintain a single relative physical position on a bookshelf is the major source of complexity in classification. Extensive latent benefits will be realised when systematic subject arrangements, providing alternative views onto OPACs, are coupled to view-based browser and search techniques. Time and effort will be saved, and effectiveness increased, as rapid access is provided to the most appropriate information to satisfy the needs of the user. A future for Dewey Classification divorced from its decimal notation is anticipated
  3. Slavic, A.; Turkulin, B.: Prevajanje klasifikacijskih oznak v naravni jezik (1998) 0.06
    0.059173033 = product of:
      0.1775191 = sum of:
        0.1775191 = weight(_text_:systematic in 3229) [ClassicSimilarity], result of:
          0.1775191 = score(doc=3229,freq=4.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.6251299 = fieldWeight in 3229, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3229)
      0.33333334 = coord(1/3)
    
    Abstract
    By providing the means for systematic arrangement, library classification has been used for centuries to organise document collections. The development of systems such as DDC, LCC a,d UDC, made classification sophisticated enough to be used both for the systematic arrangement of a collection and for information retrieval through the use of classified catalogues. Relatively short alphanumeric classification indexes are able to indicate quite sophisticated concepts, helping to overcome language barriers. With the advent of library automation, their role in the document retrieval process has become less important, in comparison with retrieval by natural language
  4. Classification research for knowledge representation and organization : Proc. of the 5th Int. Study Conf. on Classification Research, Toronto, Canada, 24.-28.6.1991 (1992) 0.05
    0.053849936 = product of:
      0.0807749 = sum of:
        0.053796407 = weight(_text_:systematic in 2072) [ClassicSimilarity], result of:
          0.053796407 = score(doc=2072,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.18944295 = fieldWeight in 2072, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0234375 = fieldNorm(doc=2072)
        0.026978498 = product of:
          0.053956997 = sum of:
            0.053956997 = weight(_text_:indexing in 2072) [ClassicSimilarity], result of:
              0.053956997 = score(doc=2072,freq=10.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.2837058 = fieldWeight in 2072, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=2072)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Content
    Enthält die Beiträge: SVENONIUS, E.: Classification: prospects, problems, and possibilities; BEALL, J.: Editing the Dewey Decimal Classification online: the evolution of the DDC database; BEGHTOL, C.: Toward a theory of fiction analysis for information storage and retrieval; CRAVEN, T.C.: Concept relation structures and their graphic display; FUGMANN, R.: Illusory goals in information science research; GILCHRIST, A.: UDC: the 1990's and beyond; GREEN, R.: The expression of syntagmatic relationships in indexing: are frame-based index languages the answer?; HUMPHREY, S.M.: Use and management of classification systems for knowledge-based indexing; MIKSA, F.L.: The concept of the universe of knowledge and the purpose of LIS classification; SCOTT, M. u. A.F. FONSECA: Methodology for functional appraisal of records and creation of a functional thesaurus; ALBRECHTSEN, H.: PRESS: a thesaurus-based information system for software reuse; AMAESHI, B.: A preliminary AAT compatible African art thesaurus; CHATTERJEE, A.: Structures of Indian classification systems of the pre-Ranganathan era and their impact on the Colon Classification; COCHRANE, P.A.: Indexing and searching thesauri, the Janus or Proteus of information retrieval; CRAVEN, T.C.: A general versus a special algorithm in the graphic display of thesauri; DAHLBERG, I.: The basis of a new universal classification system seen from a philosophy of science point of view: DRABENSTOTT, K.M., RIESTER, L.C. u. B.A.DEDE: Shelflisting using expert systems; FIDEL, R.: Thesaurus requirements for an intermediary expert system; GREEN, R.: Insights into classification from the cognitive sciences: ramifications for index languages; GROLIER, E. de: Towards a syndetic information retrieval system; GUENTHER, R.: The USMARC format for classification data: development and implementation; HOWARTH, L.C.: Factors influencing policies for the adoption and integration of revisions to classification schedules; HUDON, M.: Term definitions in subject thesauri: the Canadian literacy thesaurus experience; HUSAIN, S.: Notational techniques for the accomodation of subjects in Colon Classification 7th edition: theoretical possibility vis-à-vis practical need; KWASNIK, B.H. u. C. JORGERSEN: The exploration by means of repertory grids of semantic differences among names of official documents; MICCO, M.: Suggestions for automating the Library of Congress Classification schedules; PERREAULT, J.M.: An essay on the prehistory of general categories (II): G.W. Leibniz, Conrad Gesner; REES-POTTER, L.K.: How well do thesauri serve the social sciences?; REVIE, C.W. u. G. SMART: The construction and the use of faceted classification schema in technical domains; ROCKMORE, M.: Structuring a flexible faceted thsaurus record for corporate information retrieval; ROULIN, C.: Sub-thesauri as part of a metathesaurus; SMITH, L.C.: UNISIST revisited: compatibility in the context of collaboratories; STILES, W.G.: Notes concerning the use chain indexing as a possible means of simulating the inductive leap within artificial intelligence; SVENONIUS, E., LIU, S. u. B. SUBRAHMANYAM: Automation in chain indexing; TURNER, J.: Structure in data in the Stockshot database at the National Film Board of Canada; VIZINE-GOETZ, D.: The Dewey Decimal Classification as an online classification tool; WILLIAMSON, N.J.: Restructuring UDC: problems and possibilies; WILSON, A.: The hierarchy of belief: ideological tendentiousness in universal classification; WILSON, B.F.: An evaluation of the systematic botany schedule of the Universal Decimal Classification (English full edition, 1979); ZENG, L.: Research and development of classification and thesauri in China; CONFERENCE SUMMARY AND CONCLUSIONS
  5. Beall, J.: Editing the Dewey Decimal Classification online : the evolution of the DDC database (1992) 0.05
    0.04781903 = product of:
      0.14345708 = sum of:
        0.14345708 = weight(_text_:systematic in 2085) [ClassicSimilarity], result of:
          0.14345708 = score(doc=2085,freq=2.0), product of:
            0.28397155 = queryWeight, product of:
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.049684696 = queryNorm
            0.5051812 = fieldWeight in 2085, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.715473 = idf(docFreq=395, maxDocs=44218)
              0.0625 = fieldNorm(doc=2085)
      0.33333334 = coord(1/3)
    
    Abstract
    The database used to publish the 20th edition of the Dewey Decimal Classification (DDC) was produced on ESS (Editorial Support System). Systematic changes are being made in the database that will facilitate (1) converting ESS records to the USMARC format, (2) searching for component parts of and analyzing synthesized DDC numbers, and (3) tracking hierarchical relationships not expressed by the DDC notation.
  6. Ardo, A.; Lundberg, S.: ¬A regional distributed WWW search and indexing service : the DESIRE way (1998) 0.04
    0.041326456 = product of:
      0.12397936 = sum of:
        0.12397936 = sum of:
          0.083589815 = weight(_text_:indexing in 4190) [ClassicSimilarity], result of:
            0.083589815 = score(doc=4190,freq=6.0), product of:
              0.19018644 = queryWeight, product of:
                3.8278677 = idf(docFreq=2614, maxDocs=44218)
                0.049684696 = queryNorm
              0.4395151 = fieldWeight in 4190, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                3.8278677 = idf(docFreq=2614, maxDocs=44218)
                0.046875 = fieldNorm(doc=4190)
          0.04038954 = weight(_text_:22 in 4190) [ClassicSimilarity], result of:
            0.04038954 = score(doc=4190,freq=2.0), product of:
              0.17398734 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.049684696 = queryNorm
              0.23214069 = fieldWeight in 4190, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=4190)
      0.33333334 = coord(1/3)
    
    Abstract
    Creates an open, metadata aware system for distributed, collaborative WWW indexing. The system has 3 main components: a harvester (for collecting information), a database (for making the collection searchable), and a user interface (for making the information available). all components can be distributed across networked computers, thus supporting scalability. The system is metadata aware and thus allows searches on several fields including title, document author and URL. Nordic Web Index (NWI) is an application using this system to create a regional Nordic Web-indexing service. NWI is built using 5 collaborating service points within the Nordic countries. The NWI databases can be used to build additional services
    Date
    1. 8.1996 22:08:06
  7. Woods, E.W.; IFLA Section on classification and Indexing and Indexing and Information Technology; Joint Working Group on a Classification Format: Requirements for a format of classification data : Final report, July 1996 (1996) 0.02
    0.02275027 = product of:
      0.068250805 = sum of:
        0.068250805 = product of:
          0.13650161 = sum of:
            0.13650161 = weight(_text_:indexing in 3008) [ClassicSimilarity], result of:
              0.13650161 = score(doc=3008,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.7177252 = fieldWeight in 3008, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3008)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
  8. Riesthuis, G.J.A.: Zoeken met woorden : hergebruik van onderwerpsontsluiting (1998) 0.01
    0.0134057235 = product of:
      0.04021717 = sum of:
        0.04021717 = product of:
          0.08043434 = sum of:
            0.08043434 = weight(_text_:indexing in 3154) [ClassicSimilarity], result of:
              0.08043434 = score(doc=3154,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.42292362 = fieldWeight in 3154, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3154)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Footnote
    Übers. d. Titels: Searching with words: re-use of subject indexing
  9. Thielen, J. van: Online classification and indexation of documents at the European Patent Office (1998) 0.01
    0.013270989 = product of:
      0.039812967 = sum of:
        0.039812967 = product of:
          0.079625934 = sum of:
            0.079625934 = weight(_text_:indexing in 1607) [ClassicSimilarity], result of:
              0.079625934 = score(doc=1607,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.41867304 = fieldWeight in 1607, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1607)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The patent documentation at the European Patent Office (EPO) is organized according to the 120.000 subdivisions of the EPO's internal classification scheme (ECLA). In some fields, indexing schemes such as In Computer Only codes (ICO) provide additional ways of accessing the documentation. To meet the needs of patent examiners it was decided to develop a new DOCumentation TOOL (DOCTOOL). This tool would make it possible for examiners to enter their indexing codes directly online at their workplace and would, in addition to the existing ECLA and ICO schemes, also provide the possibility to assign keywords and free text to documents
  10. Svenonius, E.; Liu, S.; Subrahmanyam, B.: Automation of chain indexing (1992) 0.01
    0.011375135 = product of:
      0.034125403 = sum of:
        0.034125403 = product of:
          0.068250805 = sum of:
            0.068250805 = weight(_text_:indexing in 2114) [ClassicSimilarity], result of:
              0.068250805 = score(doc=2114,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.3588626 = fieldWeight in 2114, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2114)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Object
    Chain indexing
  11. Neelameghan, A.: S.R. Ranganathan's general theory of knowledge classification in designing, indexing and retrieving from specialised databases (1997) 0.01
    0.011375135 = product of:
      0.034125403 = sum of:
        0.034125403 = product of:
          0.068250805 = sum of:
            0.068250805 = weight(_text_:indexing in 3) [ClassicSimilarity], result of:
              0.068250805 = score(doc=3,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.3588626 = fieldWeight in 3, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Summarizes some experiences of the application of the priciples and postulates of S.R. Ranganathan's General Theory of Knowledge Classification, incorporating the freely faceted approach and analytico synthetic methods, to the design and development of specialized databases, including indexing, user interfaces and retrieval. Enumerates some of the earlier instances of the facet method in machine based systems, beginning with Hollerith's punched card system for the data processing of the US Census. Elaborates on Ranganathan's holistic approach to information systems and services provided by his normative principles. Notes similarities between the design of databases and faceted classification systems. Examples from working systems are given to demonstrate the usefulness of selected canons and principles of classification and the analytico synthetic methodology to database design. The examples are mostly operational database systems developed using Unesco's Micro CDS-ISIS software
  12. Pollitt, A.S.: ¬The key role of classification and indexing in view-based searching (1998) 0.01
    0.011375135 = product of:
      0.034125403 = sum of:
        0.034125403 = product of:
          0.068250805 = sum of:
            0.068250805 = weight(_text_:indexing in 4429) [ClassicSimilarity], result of:
              0.068250805 = score(doc=4429,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.3588626 = fieldWeight in 4429, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4429)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The application of classification schemes and thesauri to improve online information retrieval can be traced back to the beginnings of online searching itself, but the true potential for using knowledge structures in the user interface has yet to be realized. View-based searching seeks to exploit the classified arrangements in thesauri and existing classification schemes to improve the performance of such systems. HIBROWSE for EMBASE is a system which demonstrates the power of applying an approach to information retrieval which is strongly related to faceted classification. It does this by employing a point a click user interface with mutually constraining views utilising knowledge structure hierarchies for both query specification and the presentation of results. The relevance of this approach to library OPACs is discussed in the context of the digital library, concluding that out legacy of research in classification and indexing is more relevant than ever in the design of systems to cope with the problems of information access
  13. Ellis, D.; Vasconcelos, A.: Ranganathan and the Net : using facet analysis to search and organise the World Wide Web (1999) 0.01
    0.011375135 = product of:
      0.034125403 = sum of:
        0.034125403 = product of:
          0.068250805 = sum of:
            0.068250805 = weight(_text_:indexing in 726) [ClassicSimilarity], result of:
              0.068250805 = score(doc=726,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.3588626 = fieldWeight in 726, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=726)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This paper documents the continuing relevance of facet analysis as a technique for searching and organising WWW based materials. The 2 approaches underlying WWW searching and indexing - word and concept based indexing - are outlined. It is argued that facet analysis as an a posteriori approach to classification using words from the subject field as the concept terms in the classification derived represents an excellent approach to searching and organising the results of WWW searches using either search engines or search directories. Finally it is argued that the underlying philosophy of facet analysis is better suited to the disparate nature of WWW resources and searchers than the assumptions of contemporaray IR research.
  14. Comaromi, C.L.: Summation of classification as an enhancement of intellectual access to information in an online environment (1990) 0.01
    0.011219318 = product of:
      0.033657953 = sum of:
        0.033657953 = product of:
          0.06731591 = sum of:
            0.06731591 = weight(_text_:22 in 3576) [ClassicSimilarity], result of:
              0.06731591 = score(doc=3576,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.38690117 = fieldWeight in 3576, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3576)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    8. 1.2007 12:22:40
  15. Lincicum, S.: Critical appraisal of the use of classification in the future : non traditional uses of classification: report of a panel discussion (1995) 0.01
    0.009479279 = product of:
      0.028437834 = sum of:
        0.028437834 = product of:
          0.05687567 = sum of:
            0.05687567 = weight(_text_:indexing in 5570) [ClassicSimilarity], result of:
              0.05687567 = score(doc=5570,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29905218 = fieldWeight in 5570, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5570)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Dagobert Soergel opened the discussion by saying, "Users need our help to find and make sense of information." He said that he believes that classification can provide much of the support users need. He sees little distinction between the concept of a thesaurus and that of classification since both seek to provide structure for knowledge bases in order to facilitate information retrieval. Soergel's discussion of his concept of a multifunctional, multilingual thesaurus comprised the bulk of his presentation. This thesaurus would be a database of concepts, terms, and relationships which would include classification. In this context, classification has a much broader set of functions than it currently does in most American libraries where classification serves primarily as a method of shelf arrangement. The thesaurus Soergel envisions would lay out the semantic map of a field and could therefore be used as a learning tool or as a basis for research planning, or to assist users in clarifying terms and concepts. It could support indexing and searching and provide for the organization of knowledge for expert systems and other artificial intelligence applications. Among its other features, such a thesaurus could assist users in making sense of information by providing structured presentation of search results based on user needs and preferences, and it could enhance natural language processing capabilities such as automated indexing and abstracting and machine translation.
  16. Brandhorst, H.; Huisstede, P.V.: ICONCLASS in de computer : de classificatie van beeldmateriaal in een geautomatiseerde omgeving (1992) 0.01
    0.009384007 = product of:
      0.02815202 = sum of:
        0.02815202 = product of:
          0.05630404 = sum of:
            0.05630404 = weight(_text_:indexing in 6486) [ClassicSimilarity], result of:
              0.05630404 = score(doc=6486,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29604656 = fieldWeight in 6486, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=6486)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Contribution to thematic issue of art libraries in the Netherlands. Visual images are now widely recognised as an importatnt resource for the historical research. To improve access to collections of such material H. van de Wal developed ICONCLASS in the Netherlands. With some 150.000 controlled indexing terms the scheme is hierarchical and easy to use. In response to requests a version of ICONCLASS has been developed for use in automated systems. The scheme has been used by the Dutch Royal Library for its database of printers' devices, published in CD-ROM version in Dec. 91. Experience with the project shows the need for careful preparations and the use of trained staff
  17. Heiser, W.J.: Zoeken in de bibliografische ruimte : de bijzondere rol van de hierarchische classificaties bij geautomatiseerde onderwerpsontsluiting (1993) 0.01
    0.009384007 = product of:
      0.02815202 = sum of:
        0.02815202 = product of:
          0.05630404 = sum of:
            0.05630404 = weight(_text_:indexing in 3121) [ClassicSimilarity], result of:
              0.05630404 = score(doc=3121,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29604656 = fieldWeight in 3121, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3121)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Footnote
    Searching in the bibliographical space: the special function of hierarchical classifications in automated subject indexing
  18. Francu, V.: Construirea unui tezaur multilingv bazat pr CZU (1997) 0.01
    0.009384007 = product of:
      0.02815202 = sum of:
        0.02815202 = product of:
          0.05630404 = sum of:
            0.05630404 = weight(_text_:indexing in 3221) [ClassicSimilarity], result of:
              0.05630404 = score(doc=3221,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29604656 = fieldWeight in 3221, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3221)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The automation of Bucharest University Central Library involved the compilation of a dictionary of terms to facilitate postcoordinated searching in accordance with the UDC notation attached to every bibliographic record. Describes the project and demonstrates why a multilingual UDC based thesaurus is considered as ideal indexing and searching tool. The experiment, which applied to Class 8 of UDC (Linguisitcs and Literature), illustrates how all UDC tables can be successfully used to build a thesaurus and the ways in which their limitations can be overcome by a thesaurus
  19. Poynder, R.: Web research engines? (1996) 0.01
    0.0080434345 = product of:
      0.024130303 = sum of:
        0.024130303 = product of:
          0.048260607 = sum of:
            0.048260607 = weight(_text_:indexing in 5698) [ClassicSimilarity], result of:
              0.048260607 = score(doc=5698,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.2537542 = fieldWeight in 5698, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5698)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Describes the shortcomings of search engines for the WWW comparing their current capabilities to those of the first generation CD-ROM products. Some allow phrase searching and most are improving their Boolean searching. Few allow truncation, wild cards or nested logic. They are stateless, losing previous search criteria. Unlike the indexing and classification systems for today's CD-ROMs, those for Web pages are random, unstructured and of variable quality. Considers that at best Web search engines can only offer free text searching. Discusses whether automatic data classification systems such as Infoseek Ultra can overcome the haphazard nature of the Web with neural network technology, and whether Boolean search techniques may be redundant when replaced by technology such as the Euroferret search engine. However, artificial intelligence is rarely successful on huge, varied databases. Relevance ranking and automatic query expansion still use the same simple inverted indexes. Most Web search engines do nothing more than word counting. Further complications arise with foreign languages
  20. Liu, S.: Decomposing DDC synthesized numbers (1997) 0.01
    0.0080434345 = product of:
      0.024130303 = sum of:
        0.024130303 = product of:
          0.048260607 = sum of:
            0.048260607 = weight(_text_:indexing in 5968) [ClassicSimilarity], result of:
              0.048260607 = score(doc=5968,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.2537542 = fieldWeight in 5968, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5968)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Some empirical studies have explored the direct use of traditional classification schemes in the online environment; none has manipulated these manual classifications in such a way as to take full advantage of the power of both the classification and computer. It has been suggested that this power could be realized if the individual components of synthesized DDC numbers could be identified and indexed. Looks at the feasibility of automatically decomposing DDC synthesized numbers and the implications of such decompositions for informational retrieval. 1.701 sythesized numbers were decomposed by a computer system called DND (Dewey Number Decomposer). 600 were randomly selected for examination by 3 judges, each evaluating 200 numbers. The decomposition success rate was 100% and it was concluded that synthesized DDC numbers can be accurately decomposed automatically. The study has implications for information retrieval, expert systems for assigning DDC numbers, automatic indexing, switching language development and other important areas of cataloguing and classification