Search (68 results, page 1 of 4)

  • × theme_ss:"Klassifikationssysteme im Online-Retrieval"
  1. Vizine-Goetz, D.: OCLC investigates using classification tools to organize Internet data (1998) 0.05
    0.053239673 = product of:
      0.10647935 = sum of:
        0.10647935 = sum of:
          0.057025105 = weight(_text_:data in 2342) [ClassicSimilarity], result of:
            0.057025105 = score(doc=2342,freq=4.0), product of:
              0.16488427 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.052144732 = queryNorm
              0.34584928 = fieldWeight in 2342, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2342)
          0.049454242 = weight(_text_:22 in 2342) [ClassicSimilarity], result of:
            0.049454242 = score(doc=2342,freq=2.0), product of:
              0.18260197 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052144732 = queryNorm
              0.2708308 = fieldWeight in 2342, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2342)
      0.5 = coord(1/2)
    
    Date
    22. 9.1997 19:16:05
    Source
    Visualizing subject access for 21st century information resources: Papers presented at the 1997 Clinic on Library Applications of Data Processing, 2-4 Mar 1997, Graduate School of Library and Information Science, University of Illinois at Urbana-Champaign. Ed.: P.A. Cochrane et al
  2. Dack, D.: Australian attends conference on Dewey (1989) 0.04
    0.04488854 = product of:
      0.08977708 = sum of:
        0.08977708 = sum of:
          0.040322836 = weight(_text_:data in 2509) [ClassicSimilarity], result of:
            0.040322836 = score(doc=2509,freq=2.0), product of:
              0.16488427 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.052144732 = queryNorm
              0.24455236 = fieldWeight in 2509, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2509)
          0.049454242 = weight(_text_:22 in 2509) [ClassicSimilarity], result of:
            0.049454242 = score(doc=2509,freq=2.0), product of:
              0.18260197 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052144732 = queryNorm
              0.2708308 = fieldWeight in 2509, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2509)
      0.5 = coord(1/2)
    
    Abstract
    Edited version of a report to the Australian Library and Information Association on the Conference on classification theory in the computer age, Albany, New York, 18-19 Nov 88, and on the meeting of the Dewey Editorial Policy Committee which preceded it. The focus of the Editorial Policy Committee Meeting lay in the following areas: browsing; potential for improved subject access; system design; potential conflict between shelf location and information retrieval; and users. At the Conference on classification theory in the computer age the following papers were presented: Applications of artificial intelligence to bibliographic classification, by Irene Travis; Automation and classification, By Elaine Svenonious; Subject classification and language processing for retrieval in large data bases, by Diana Scott; Implications for information processing, by Carol Mandel; and implications for information science education, by Richard Halsey.
    Date
    8.11.1995 11:52:22
  3. Kent, R.E.: Organizing conceptual knowledge online : metadata interoperability and faceted classification (1998) 0.04
    0.04488854 = product of:
      0.08977708 = sum of:
        0.08977708 = sum of:
          0.040322836 = weight(_text_:data in 57) [ClassicSimilarity], result of:
            0.040322836 = score(doc=57,freq=2.0), product of:
              0.16488427 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.052144732 = queryNorm
              0.24455236 = fieldWeight in 57, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.0546875 = fieldNorm(doc=57)
          0.049454242 = weight(_text_:22 in 57) [ClassicSimilarity], result of:
            0.049454242 = score(doc=57,freq=2.0), product of:
              0.18260197 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052144732 = queryNorm
              0.2708308 = fieldWeight in 57, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=57)
      0.5 = coord(1/2)
    
    Abstract
    Conceptual Knowledge Markup Language (CKML), an application of XML, is a new standard being promoted for the specification of online conceptual knowledge (Kent and Shrivastava, 1998). CKML follows the philosophy of Conceptual Knowledge Processing (Wille, 1982), a principled approach to knowledge representation and data analysis, which advocates the development of methodologies and techniques to support people in their rational thinking, judgement and actions. CKML was developed and is being used in the WAVE networked information discovery and retrieval system (Kent and Neuss, 1994) as a standard for the specification of conceptual knowledge
    Date
    30.12.2001 16:22:41
  4. Chandler, A.; LeBlanc, J.: Exploring the potential of a virtual undergraduate library collection based on the hierarchical interface to LC Classification (2006) 0.04
    0.03847589 = product of:
      0.07695178 = sum of:
        0.07695178 = sum of:
          0.03456243 = weight(_text_:data in 769) [ClassicSimilarity], result of:
            0.03456243 = score(doc=769,freq=2.0), product of:
              0.16488427 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.052144732 = queryNorm
              0.2096163 = fieldWeight in 769, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.046875 = fieldNorm(doc=769)
          0.04238935 = weight(_text_:22 in 769) [ClassicSimilarity], result of:
            0.04238935 = score(doc=769,freq=2.0), product of:
              0.18260197 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052144732 = queryNorm
              0.23214069 = fieldWeight in 769, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=769)
      0.5 = coord(1/2)
    
    Abstract
    The Hierarchical Interface to Library of Congress Classification (HILCC) is a system developed by the Columbia University Library to leverage call number data from the MARC holdings records in Columbia's online catalog to create a structured, hierarchical menuing system that provides subject access to the library's electronic resources. In this paper, the authors describe a research initiative at the Cornell University Library to discover if the Columbia HILCC scheme can be used as developed or in modified form to create a virtual undergraduate print collection outside the context of the traditional online catalog. Their results indicate that, with certain adjustments, an HILCC model can indeed, be used to represent the holdings of a large research library's undergraduate collection of approximately 150,000 titles, but that such a model is not infinitely scalable and may require a new approach to browsing such a large information space.
    Date
    10. 9.2000 17:38:22
  5. Slavic, A.: On the nature and typology of documentary classifications and their use in a networked environment (2007) 0.04
    0.03847589 = product of:
      0.07695178 = sum of:
        0.07695178 = sum of:
          0.03456243 = weight(_text_:data in 780) [ClassicSimilarity], result of:
            0.03456243 = score(doc=780,freq=2.0), product of:
              0.16488427 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.052144732 = queryNorm
              0.2096163 = fieldWeight in 780, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.046875 = fieldNorm(doc=780)
          0.04238935 = weight(_text_:22 in 780) [ClassicSimilarity], result of:
            0.04238935 = score(doc=780,freq=2.0), product of:
              0.18260197 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052144732 = queryNorm
              0.23214069 = fieldWeight in 780, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=780)
      0.5 = coord(1/2)
    
    Abstract
    Networked orientated standards for vocabulary publishing and exchange and proposals for terminological services and terminology registries will improve sharing and use of all knowledge organization systems in the networked information environment. This means that documentary classifications may also become more applicable for use outside their original domain of application. The paper summarises some characteristics common to documentary classifications and explains some terminological, functional and implementation aspects. The original purpose behind each classification scheme determines the functions that the vocabulary is designed to facilitate. These functions influence the structure, semantics and syntax, scheme coverage and format in which classification data are published and made available. The author suggests that attention should be paid to the differences between documentary classifications as these may determine their suitability for a certain purpose and may impose different requirements with respect to their use online. As we speak, many classifications are being created for knowledge organization and it may be important to promote expertise from the bibliographic domain with respect to building and using classification systems.
    Date
    22.12.2007 17:22:31
  6. Frâncu, V.; Sabo, C.-N.: Implementation of a UDC-based multilingual thesaurus in a library catalogue : the case of BiblioPhil (2010) 0.04
    0.03847589 = product of:
      0.07695178 = sum of:
        0.07695178 = sum of:
          0.03456243 = weight(_text_:data in 3697) [ClassicSimilarity], result of:
            0.03456243 = score(doc=3697,freq=2.0), product of:
              0.16488427 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.052144732 = queryNorm
              0.2096163 = fieldWeight in 3697, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.046875 = fieldNorm(doc=3697)
          0.04238935 = weight(_text_:22 in 3697) [ClassicSimilarity], result of:
            0.04238935 = score(doc=3697,freq=2.0), product of:
              0.18260197 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052144732 = queryNorm
              0.23214069 = fieldWeight in 3697, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=3697)
      0.5 = coord(1/2)
    
    Abstract
    In order to enhance the use of Universal Decimal Classification (UDC) numbers in information retrieval, the authors have represented classification with multilingual thesaurus descriptors and implemented this solution in an automated way. The authors illustrate a solution implemented in a BiblioPhil library system. The standard formats used are UNIMARC for subject authority records (i.e. the UDC-based multilingual thesaurus) and MARC XML support for data transfer. The multilingual thesaurus was built according to existing standards, the constituent parts of the classification notations being used as the basis for search terms in the multilingual information retrieval. The verbal equivalents, descriptors and non-descriptors, are used to expand the number of concepts and are given in Romanian, English and French. This approach saves the time of the indexer and provides more user-friendly and easier access to the bibliographic information. The multilingual aspect of the thesaurus enhances information access for a greater number of online users
    Date
    22. 7.2010 20:40:56
  7. Hill, J.S.: Online classification number access : some practical considerations (1984) 0.03
    0.028259566 = product of:
      0.056519132 = sum of:
        0.056519132 = product of:
          0.113038264 = sum of:
            0.113038264 = weight(_text_:22 in 7684) [ClassicSimilarity], result of:
              0.113038264 = score(doc=7684,freq=2.0), product of:
                0.18260197 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052144732 = queryNorm
                0.61904186 = fieldWeight in 7684, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=7684)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Journal of academic librarianship. 10(1984), S.17-22
  8. Guenther, R.S.: Automating the Library of Congress Classification Scheme : implementation of the USMARC format for classification data (1996) 0.02
    0.024692593 = product of:
      0.049385186 = sum of:
        0.049385186 = product of:
          0.09877037 = sum of:
            0.09877037 = weight(_text_:data in 5578) [ClassicSimilarity], result of:
              0.09877037 = score(doc=5578,freq=12.0), product of:
                0.16488427 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.052144732 = queryNorm
                0.59902847 = fieldWeight in 5578, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5578)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Potential uses for classification data in machine readable form and reasons for the development of a standard, the USMARC Format for Classification Data, which allows for classification data to interact with other USMARC bibliographic and authority data are discussed. The development, structure, content, and use of the standard is reviewed with implementation decisions for the Library of Congress Classification scheme noted. The author examines the implementation of USMARC classification at LC, the conversion of the schedules, and the functionality of the software being used. Problems in the effort are explored, and enhancements desired for the online classification system are considered.
    Object
    USMARC for classification data
  9. Woods, E.W.; IFLA Section on classification and Indexing and Indexing and Information Technology; Joint Working Group on a Classification Format: Requirements for a format of classification data : Final report, July 1996 (1996) 0.02
    0.024439331 = product of:
      0.048878662 = sum of:
        0.048878662 = product of:
          0.097757325 = sum of:
            0.097757325 = weight(_text_:data in 3008) [ClassicSimilarity], result of:
              0.097757325 = score(doc=3008,freq=4.0), product of:
                0.16488427 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.052144732 = queryNorm
                0.5928845 = fieldWeight in 3008, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3008)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Object
    USMARC for classification data
  10. Concise UNIMARC Classification Format : Draft 5 (20000125) (2000) 0.02
    0.02304162 = product of:
      0.04608324 = sum of:
        0.04608324 = product of:
          0.09216648 = sum of:
            0.09216648 = weight(_text_:data in 4421) [ClassicSimilarity], result of:
              0.09216648 = score(doc=4421,freq=2.0), product of:
                0.16488427 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.052144732 = queryNorm
                0.5589768 = fieldWeight in 4421, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.125 = fieldNorm(doc=4421)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Object
    UNIMARC for classification data
  11. Lim, E.: Southeast Asian subject gateways : an examination of their classification practices (2000) 0.02
    0.021194674 = product of:
      0.04238935 = sum of:
        0.04238935 = product of:
          0.0847787 = sum of:
            0.0847787 = weight(_text_:22 in 6040) [ClassicSimilarity], result of:
              0.0847787 = score(doc=6040,freq=2.0), product of:
                0.18260197 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052144732 = queryNorm
                0.46428138 = fieldWeight in 6040, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6040)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 6.2002 19:42:47
  12. Aluri, R.D.; Kemp, A.; Boll, J.J.: Subject analysis in online catalogs (1991) 0.02
    0.020161418 = product of:
      0.040322836 = sum of:
        0.040322836 = product of:
          0.08064567 = sum of:
            0.08064567 = weight(_text_:data in 863) [ClassicSimilarity], result of:
              0.08064567 = score(doc=863,freq=8.0), product of:
                0.16488427 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.052144732 = queryNorm
                0.48910472 = fieldWeight in 863, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=863)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    LCSH
    Subject cataloguing / Data processing
    Machine / readable bibliographic data
    Subject
    Subject cataloguing / Data processing
    Machine / readable bibliographic data
  13. Guenther, R.S.: ¬The Library of Congress Classification in the USMARC format (1994) 0.02
    0.020161418 = product of:
      0.040322836 = sum of:
        0.040322836 = product of:
          0.08064567 = sum of:
            0.08064567 = weight(_text_:data in 8864) [ClassicSimilarity], result of:
              0.08064567 = score(doc=8864,freq=8.0), product of:
                0.16488427 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.052144732 = queryNorm
                0.48910472 = fieldWeight in 8864, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=8864)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The paper reviews the development of the USMARC Format for Classification Data, a standard for communication of classification data in machine-readable form. It considers the uses for online classification schedules, both for technical services and reference functions and gives an overview of the format specification details of data elements used and of the structure of the records. The paper describes an experiment conducted at the Library of Congress to test the format as well as the development of the classification database encompassing the LCC schedules. Features of the classification system are given. The LoC will complete its conversion of the LCC in mid-1995
    Object
    USMARC for classification data
  14. Guenther, R.S.: ¬The USMARC Format for Classification Data : development and implementation (1992) 0.02
    0.019954631 = product of:
      0.039909262 = sum of:
        0.039909262 = product of:
          0.079818524 = sum of:
            0.079818524 = weight(_text_:data in 2996) [ClassicSimilarity], result of:
              0.079818524 = score(doc=2996,freq=6.0), product of:
                0.16488427 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.052144732 = queryNorm
                0.48408815 = fieldWeight in 2996, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2996)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper discusses the newly developed USMARC Format for Classification Data. It reviews its potential uses within an online system and its development as one of the USMARC standards for representing bibliographic and related information in machine-readable form. It provides a summary of the fields in the format, and considers the prospects for its implementation.
    Object
    USMARC for classification data
  15. Guenther, R.S.: ¬The development and implementation of the USMARC format for classification data (1992) 0.02
    0.019954631 = product of:
      0.039909262 = sum of:
        0.039909262 = product of:
          0.079818524 = sum of:
            0.079818524 = weight(_text_:data in 8865) [ClassicSimilarity], result of:
              0.079818524 = score(doc=8865,freq=6.0), product of:
                0.16488427 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.052144732 = queryNorm
                0.48408815 = fieldWeight in 8865, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0625 = fieldNorm(doc=8865)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper discusses the newly developed USMARC Format for Classification Data. It reviews its potential uses within an online system and its development as one of the USMARC standards. It provides a summary of the fields in the format and considers the prospects for its implementation. The papaer describes an experiment currently being conducted at the Library of Congress to create USMARC classification records and use a classification database in classifying materials in the social sciences
    Object
    USMARC for classification data
  16. Comaromi, C.L.: Summation of classification as an enhancement of intellectual access to information in an online environment (1990) 0.02
    0.017662229 = product of:
      0.035324458 = sum of:
        0.035324458 = product of:
          0.070648916 = sum of:
            0.070648916 = weight(_text_:22 in 3576) [ClassicSimilarity], result of:
              0.070648916 = score(doc=3576,freq=2.0), product of:
                0.18260197 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052144732 = queryNorm
                0.38690117 = fieldWeight in 3576, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3576)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    8. 1.2007 12:22:40
  17. Reiner, U.: Automatische DDC-Klassifizierung von bibliografischen Titeldatensätzen (2009) 0.02
    0.017662229 = product of:
      0.035324458 = sum of:
        0.035324458 = product of:
          0.070648916 = sum of:
            0.070648916 = weight(_text_:22 in 611) [ClassicSimilarity], result of:
              0.070648916 = score(doc=611,freq=2.0), product of:
                0.18260197 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052144732 = queryNorm
                0.38690117 = fieldWeight in 611, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=611)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 8.2009 12:54:24
  18. Quick Guide to Publishing a Classification Scheme on the Semantic Web (2008) 0.02
    0.017460302 = product of:
      0.034920603 = sum of:
        0.034920603 = product of:
          0.069841206 = sum of:
            0.069841206 = weight(_text_:data in 3061) [ClassicSimilarity], result of:
              0.069841206 = score(doc=3061,freq=6.0), product of:
                0.16488427 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.052144732 = queryNorm
                0.42357713 = fieldWeight in 3061, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3061)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This document describes in brief how to express the content and structure of a classification scheme, and metadata about a classification scheme, in RDF using the SKOS vocabulary. RDF allows data to be linked to and/or merged with other RDF data by semantic web applications. The Semantic Web, which is based on the Resource Description Framework (RDF), provides a common framework that allows data to be shared and reused across application, enterprise, and community boundaries. Publishing classifications schemes in SKOS will unify the great many of existing classification efforts in the framework of the Semantic Web.
  19. Vizine-Goetz, D.: OCLC investigates using classification tools to organize Internet data (1997) 0.02
    0.017281216 = product of:
      0.03456243 = sum of:
        0.03456243 = product of:
          0.06912486 = sum of:
            0.06912486 = weight(_text_:data in 3410) [ClassicSimilarity], result of:
              0.06912486 = score(doc=3410,freq=2.0), product of:
                0.16488427 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.052144732 = queryNorm
                0.4192326 = fieldWeight in 3410, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3410)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  20. Slavic, A.; Cordeiro, M.I.: Core requirements for automation of analytico-synthetic classifications (2004) 0.02
    0.017281216 = product of:
      0.03456243 = sum of:
        0.03456243 = product of:
          0.06912486 = sum of:
            0.06912486 = weight(_text_:data in 2651) [ClassicSimilarity], result of:
              0.06912486 = score(doc=2651,freq=8.0), product of:
                0.16488427 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.052144732 = queryNorm
                0.4192326 = fieldWeight in 2651, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2651)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The paper analyses the importance of data presentation and modelling and its role in improving the management, use and exchange of analytico-synthetic classifications in automated systems. Inefficiencies, in this respect, hinder the automation of classification systems that offer the possibility of building compound index/search terms. The lack of machine readable data expressing the semantics and structure of a classification vocabulary has negative effects on information management and retrieval, thus restricting the potential of both automated systems and classifications themselves. The authors analysed the data representation structure of three general analytico-synthetic classification systems (BC2-Bliss Bibliographic Classification; BSO-Broad System of Ordering; UDC-Universal Decimal Classification) and put forward some core requirements for classification data representation

Years

Languages

  • e 57
  • d 8
  • ja 1
  • nl 1
  • More… Less…

Types

  • a 57
  • el 8
  • m 3
  • s 2
  • More… Less…