Search (35 results, page 1 of 2)

  • × theme_ss:"Metadaten"
  • × year_i:[2010 TO 2020}
  1. DeZelar-Tiedman, C.: Exploring user-contributed metadata's potential to enhance access to literary works (2011) 0.14
    0.1441605 = product of:
      0.21624073 = sum of:
        0.06374531 = weight(_text_:subject in 2595) [ClassicSimilarity], result of:
          0.06374531 = score(doc=2595,freq=6.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.41066417 = fieldWeight in 2595, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.046875 = fieldNorm(doc=2595)
        0.15249541 = sum of:
          0.11721466 = weight(_text_:headings in 2595) [ClassicSimilarity], result of:
            0.11721466 = score(doc=2595,freq=6.0), product of:
              0.21048847 = queryWeight, product of:
                4.849944 = idf(docFreq=940, maxDocs=44218)
                0.043400183 = queryNorm
              0.55686975 = fieldWeight in 2595, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                4.849944 = idf(docFreq=940, maxDocs=44218)
                0.046875 = fieldNorm(doc=2595)
          0.035280753 = weight(_text_:22 in 2595) [ClassicSimilarity], result of:
            0.035280753 = score(doc=2595,freq=2.0), product of:
              0.15198004 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043400183 = queryNorm
              0.23214069 = fieldWeight in 2595, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2595)
      0.6666667 = coord(2/3)
    
    Abstract
    Academic libraries have moved toward providing social networking features, such as tagging, in their library catalogs. To explore whether user tags can enhance access to individual literary works, the author obtained a sample of individual works of English and American literature from the twentieth and twenty-first centuries from a large academic library catalog and searched them in LibraryThing. The author compared match rates, the availability of subject headings and tags across various literary forms, and the terminology used in tags versus controlled-vocabulary headings on a subset of records. In addition, she evaluated the usefulness of available LibraryThing tags for the library catalog records that lacked subject headings. Options for utilizing the subject terms available in sources outside the local catalog also are discussed.
    Date
    10. 9.2000 17:38:22
  2. Sturmane, A.; Eglite, E.; Jankevica-Balode, M.: Subject metadata development for digital resources in Latvia (2014) 0.13
    0.12654677 = product of:
      0.18982016 = sum of:
        0.12144493 = weight(_text_:subject in 1963) [ClassicSimilarity], result of:
          0.12144493 = score(doc=1963,freq=16.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.7823804 = fieldWeight in 1963, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1963)
        0.06837522 = product of:
          0.13675044 = sum of:
            0.13675044 = weight(_text_:headings in 1963) [ClassicSimilarity], result of:
              0.13675044 = score(doc=1963,freq=6.0), product of:
                0.21048847 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.043400183 = queryNorm
                0.6496814 = fieldWeight in 1963, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1963)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The National Library of Latvia (NLL) made a decision to use the Library of Congress Subject Headings (LCSH) in 2000. At present the NLL Subject Headings Database in Latvian holds approximately 34,000 subject headings and is used for subject cataloging of textual resources, including articles from serials. For digital objects NLL uses a system like Faceted Application of Subject Terminology (FAST). We succesfully use it in the project "In Search of Lost Latvia," one of the milestones in the development of the subject cataloging of digital resources in Latvia.
    Footnote
    Contribution in a special issue "Beyond libraries: Subject metadata in the digital environment and Semantic Web" - Enthält Beiträge der gleichnamigen IFLA Satellite Post-Conference, 17-18 August 2012, Tallinn.
  3. Hook, P.A.; Gantchev, A.: Using combined metadata sources to visualize a small library (OBL's English Language Books) (2017) 0.08
    0.07827899 = product of:
      0.11741848 = sum of:
        0.06857903 = weight(_text_:subject in 3870) [ClassicSimilarity], result of:
          0.06857903 = score(doc=3870,freq=10.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.4418043 = fieldWeight in 3870, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3870)
        0.048839446 = product of:
          0.09767889 = sum of:
            0.09767889 = weight(_text_:headings in 3870) [ClassicSimilarity], result of:
              0.09767889 = score(doc=3870,freq=6.0), product of:
                0.21048847 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.043400183 = queryNorm
                0.46405816 = fieldWeight in 3870, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3870)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Data from multiple knowledge organization systems are combined to provide a global overview of the content holdings of a small personal library. Subject headings and classification data are used to effectively map the combined book and topic space of the library. While harvested and manipulated by hand, the work reveals issues and potential solutions when using automated techniques to produce topic maps of much larger libraries. The small library visualized consists of the thirty-nine, digital, English language books found in the Osama Bin Laden (OBL) compound in Abbottabad, Pakistan upon his death. As this list of books has garnered considerable media attention, it is worth providing a visual overview of the subject content of these books - some of which is not readily apparent from the titles. Metadata from subject headings and classification numbers was combined to create book-subject maps. Tree maps of the classification data were also produced. The books contain 328 subject headings. In order to enhance the base map with meaningful thematic overlay, library holding count data was also harvested (and aggregated from duplicates). This additional data revealed the relative scarcity or popularity of individual books.
  4. Leong, J.H.-t.: ¬The convergence of metadata and bibliographic control? : trends and patterns in addressing the current issues and challenges of providing subject access (2010) 0.07
    0.07162914 = product of:
      0.107443705 = sum of:
        0.073606744 = weight(_text_:subject in 3355) [ClassicSimilarity], result of:
          0.073606744 = score(doc=3355,freq=8.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.4741941 = fieldWeight in 3355, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.046875 = fieldNorm(doc=3355)
        0.03383696 = product of:
          0.06767392 = sum of:
            0.06767392 = weight(_text_:headings in 3355) [ClassicSimilarity], result of:
              0.06767392 = score(doc=3355,freq=2.0), product of:
                0.21048847 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.043400183 = queryNorm
                0.3215089 = fieldWeight in 3355, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3355)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Resource description and discovery have been facilitated generally in two approaches, namely bibliographic control and metadata, which now may converge in response to current issues and challenges of providing subject access. Four categories of major issues and challenges in the provision of subject access to digital and non-digital resources are: 1) the advancement of new knowledge; 2) the fall of controlled vocabulary and the rise of natural language; 3) digitizing and networking the traditional catalogue systems; and 4) electronic publishing and the Internet. The creation of new knowledge and the debate about the use of natural language and controlled vocabulary as subject headings becomes even more intense in the digital and online environment. The third and fourth categories are conceived after the emergence of networked environments and the rapid expansion of electronic resources. Recognizing the convergence of metadata schemas and bibliographic control calls for adapting to the new environment by developing tools that exploit the strengths of both.
  5. Dunsire, G.; Willer, M.: Initiatives to make standard library metadata models and structures available to the Semantic Web (2010) 0.04
    0.04440023 = product of:
      0.066600345 = sum of:
        0.034698553 = weight(_text_:subject in 3965) [ClassicSimilarity], result of:
          0.034698553 = score(doc=3965,freq=4.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.22353725 = fieldWeight in 3965, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.03125 = fieldNorm(doc=3965)
        0.03190179 = product of:
          0.06380358 = sum of:
            0.06380358 = weight(_text_:headings in 3965) [ClassicSimilarity], result of:
              0.06380358 = score(doc=3965,freq=4.0), product of:
                0.21048847 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.043400183 = queryNorm
                0.3031215 = fieldWeight in 3965, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3965)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The paper discusses the importance of these initiatives in releasing as linked data the very large quantities of rich, professionally-generated metadata stored in formats based on these standards, such as UNIMARC and MARC21, addressing such issues as critical mass for semantic and statistical inferencing, integration with user- and machine-generated metadata, and authenticity, veracity and trust. The paper also discusses related initiatives to release controlled vocabularies, including the Dewey Decimal Classification (DDC), ISBD, Library of Congress Name Authority File (LCNAF), Library of Congress Subject Headings (LCSH), Rameau (French subject headings), Universal Decimal Classification (UDC), and the Virtual International Authority File (VIAF) as linked data. Finally, the paper discusses the potential collective impact of these initiatives on metadata workflows and management systems.
  6. Strobel, S.; Marín-Arraiza, P.: Metadata for scientific audiovisual media : current practices and perspectives of the TIB / AV-portal (2015) 0.04
    0.03924463 = product of:
      0.058866944 = sum of:
        0.030669477 = weight(_text_:subject in 3667) [ClassicSimilarity], result of:
          0.030669477 = score(doc=3667,freq=2.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.19758089 = fieldWeight in 3667, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3667)
        0.028197467 = product of:
          0.056394935 = sum of:
            0.056394935 = weight(_text_:headings in 3667) [ClassicSimilarity], result of:
              0.056394935 = score(doc=3667,freq=2.0), product of:
                0.21048847 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.043400183 = queryNorm
                0.2679241 = fieldWeight in 3667, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3667)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Descriptive metadata play a key role in finding relevant search results in large amounts of unstructured data. However, current scientific audiovisual media are provided with little metadata, which makes them hard to find, let alone individual sequences. In this paper, the TIB / AV-Portal is presented as a use case where methods concerning the automatic generation of metadata, a semantic search and cross-lingual retrieval (German/English) have already been applied. These methods result in a better discoverability of the scientific audiovisual media hosted in the portal. Text, speech, and image content of the video are automatically indexed by specialised GND (Gemeinsame Normdatei) subject headings. A semantic search is established based on properties of the GND ontology. The cross-lingual retrieval uses English 'translations' that were derived by an ontology mapping (DBpedia i. a.). Further ways of increasing the discoverability and reuse of the metadata are publishing them as Linked Open Data and interlinking them with other data sets.
  7. Zavalina, O.L.: Complementarity in subject metadata in large-scale digital libraries : a comparative analysis (2014) 0.04
    0.03879416 = product of:
      0.11638248 = sum of:
        0.11638248 = weight(_text_:subject in 1972) [ClassicSimilarity], result of:
          0.11638248 = score(doc=1972,freq=20.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.7497667 = fieldWeight in 1972, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.046875 = fieldNorm(doc=1972)
      0.33333334 = coord(1/3)
    
    Abstract
    Provision of high-quality subject metadata is crucial for organizing adequate subject access to rich content aggregated by digital libraries. A number of large-scale digital libraries worldwide are now generating subject metadata to describe not only individual objects but entire digital collections as an integral whole. However, little research to date has been conducted to empirically evaluate the quality of this collection-level subject metadata. The study presented in this article compares free-text and controlled-vocabulary collection-level subject metadata in three large-scale cultural heritage digital libraries in the United States and the European Union. As revealed by this study, the emerging best practices for creating rich collection-level subject metadata includes describing a collection's subject matter with mutually complementary data values in controlled-vocabulary and free-text subject metadata elements. Three kinds of complementarity were observed in this study: one-way complementarity, two-way complementarity, and multiple complementarity.
    Footnote
    Contribution in a special issue "Beyond libraries: Subject metadata in the digital environment and Semantic Web" - Enthält Beiträge der gleichnamigen IFLA Satellite Post-Conference, 17-18 August 2012, Tallinn.
  8. Baker, T.: Dublin Core Application Profiles : current approaches (2010) 0.04
    0.03629583 = product of:
      0.054443747 = sum of:
        0.036803372 = weight(_text_:subject in 3737) [ClassicSimilarity], result of:
          0.036803372 = score(doc=3737,freq=2.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.23709705 = fieldWeight in 3737, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.046875 = fieldNorm(doc=3737)
        0.017640376 = product of:
          0.035280753 = sum of:
            0.035280753 = weight(_text_:22 in 3737) [ClassicSimilarity], result of:
              0.035280753 = score(doc=3737,freq=2.0), product of:
                0.15198004 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043400183 = queryNorm
                0.23214069 = fieldWeight in 3737, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3737)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The Dublin Core Metadata Initiative currently defines a Dublin Core Application Profile as a set of specifications about the metadata design of a particular application or for a particular domain or community of users. The current approach to application profiles is summarized in the Singapore Framework for Application Profiles [SINGAPORE-FRAMEWORK] (see Figure 1). While the approach originally developed as a means of specifying customized applications based on the fifteen elements of the Dublin Core Element Set (e.g., Title, Date, Subject), it has evolved into a generic approach to creating metadata that meets specific local requirements while integrating coherently with other RDF-based metadata.
    Source
    Wissensspeicher in digitalen Räumen: Nachhaltigkeit - Verfügbarkeit - semantische Interoperabilität. Proceedings der 11. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation, Konstanz, 20. bis 22. Februar 2008. Hrsg.: J. Sieglerschmidt u. H.P.Ohly
  9. Ashton, J.; Kent, C.: New approaches to subject indexing at the British Library (2017) 0.03
    0.028624846 = product of:
      0.085874535 = sum of:
        0.085874535 = weight(_text_:subject in 5158) [ClassicSimilarity], result of:
          0.085874535 = score(doc=5158,freq=8.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.5532265 = fieldWeight in 5158, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5158)
      0.33333334 = coord(1/3)
    
    Abstract
    The constantly changing metadata landscape means that libraries need to re-think their approach to standards and subject analysis, to enable the discovery of vast areas of both print and digital content. This article presents a case study from the British Library that assesses the feasibility of adopting FAST (Faceted Application of Subject Terminology) to selectively extend the scope of subject indexing of current and legacy content, or implement FAST as a replacement for all LCSH in current cataloging workflows.
  10. Bundza, M.: ¬The choice is yours! : researchers assign subject metadata to their own materials in institutional repositories (2014) 0.02
    0.024789846 = product of:
      0.074369535 = sum of:
        0.074369535 = weight(_text_:subject in 1968) [ClassicSimilarity], result of:
          0.074369535 = score(doc=1968,freq=6.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.4791082 = fieldWeight in 1968, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1968)
      0.33333334 = coord(1/3)
    
    Abstract
    The Digital Commons platform for institutional repositories provides a three-tiered taxonomy of academic disciplines for each item submitted to the repository. Since faculty and departmental administrators across campuses are encouraged to submit materials to the institutional repository themselves, they must also assign disciplines or subject categories for their own work. The expandable drop-down menu of about 1,000 categories is easy to use, and facilitates the growth of the institutional repository and access to the materials through the Internet.
    Footnote
    Contribution in a special issue "Beyond libraries: Subject metadata in the digital environment and Semantic Web" - Enthält Beiträge der gleichnamigen IFLA Satellite Post-Conference, 17-18 August 2012, Tallinn.
  11. Carlson, S.; Seely, A.: Using OpenRefine's reconciliation to validate local authority headings (2017) 0.02
    0.021267861 = product of:
      0.06380358 = sum of:
        0.06380358 = product of:
          0.12760717 = sum of:
            0.12760717 = weight(_text_:headings in 5142) [ClassicSimilarity], result of:
              0.12760717 = score(doc=5142,freq=4.0), product of:
                0.21048847 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.043400183 = queryNorm
                0.606243 = fieldWeight in 5142, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5142)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    In 2015, the Cataloging and Metadata Services department of Rice University's Fondren Library developed a process to reconcile four years of authority headings against an internally developed thesaurus. With a goal of immediate cleanup as well as an ongoing maintenance procedure, staff developed a "hack" of OpenRefine's normal Reconciliation function that ultimately yielded 99.6% authority reconciliation and a stable process for monthly data verification.
  12. Raja, N.A.: Digitized content and index pages as alternative subject access fields (2012) 0.02
    0.021248437 = product of:
      0.06374531 = sum of:
        0.06374531 = weight(_text_:subject in 870) [ClassicSimilarity], result of:
          0.06374531 = score(doc=870,freq=6.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.41066417 = fieldWeight in 870, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.046875 = fieldNorm(doc=870)
      0.33333334 = coord(1/3)
    
    Abstract
    This article describes a pilot study undertaken to test the benefits of the digitized Content and Index pages of books and content pages of journal Issues in providing subject access to documents in a collection. A partial digitization strategy is used to fossick specific information using the alternative subject access fields in bibliographic records. A pilot study was carried out to search for books and journal articles containing information on "Leadership., "Women Entrepreneurs., "Disinvestment. and "Digital preservation. through normal procedu re and based on information stored in MARC 21 fields 653, 505 and 520 of the bibliographic records in the University of Mumbai Library. The results are compared to draw the conclusions.
  13. Wolfe, EW.: a case study in automated metadata enhancement : Natural Language Processing in the humanities (2019) 0.01
    0.014312423 = product of:
      0.042937268 = sum of:
        0.042937268 = weight(_text_:subject in 5236) [ClassicSimilarity], result of:
          0.042937268 = score(doc=5236,freq=2.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.27661324 = fieldWeight in 5236, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5236)
      0.33333334 = coord(1/3)
    
    Abstract
    The Black Book Interactive Project at the University of Kansas (KU) is developing an expanded corpus of novels by African American authors, with an emphasis on lesser known writers and a goal of expanding research in this field. Using a custom metadata schema with an emphasis on race-related elements, each novel is analyzed for a variety of elements such as literary style, targeted content analysis, historical context, and other areas. Librarians at KU have worked to develop a variety of computational text analysis processes designed to assist with specific aspects of this metadata collection, including text mining and natural language processing, automated subject extraction based on word sense disambiguation, harvesting data from Wikidata, and other actions.
  14. Bartczak, J.; Glendon, I.: Python, Google Sheets, and the Thesaurus for Graphic Materials for efficient metadata project workflows (2017) 0.01
    0.012267791 = product of:
      0.036803372 = sum of:
        0.036803372 = weight(_text_:subject in 3893) [ClassicSimilarity], result of:
          0.036803372 = score(doc=3893,freq=2.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.23709705 = fieldWeight in 3893, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.046875 = fieldNorm(doc=3893)
      0.33333334 = coord(1/3)
    
    Abstract
    In 2017, the University of Virginia (U.Va.) will launch a two year initiative to celebrate the bicentennial anniversary of the University's founding in 1819. The U.Va. Library is participating in this event by digitizing some 20,000 photographs and negatives that document student life on the U.Va. grounds in the 1960s and 1970s. Metadata librarians and archivists are well-versed in the challenges associated with generating digital content and accompanying description within the context of limited resources. This paper describes how technology and new approaches to metadata design have enabled the University of Virginia's Metadata Analysis and Design Department to rapidly and successfully generate accurate description for these digital objects. Python's pandas module improves efficiency by cleaning and repurposing data recorded at digitization, while the lxml module builds MODS XML programmatically from CSV tables. A simplified technique for subject heading selection and assignment in Google Sheets provides a collaborative environment for streamlined metadata creation and data quality control.
  15. Khoo, M.J.; Ahn, J.-w.; Binding, C.; Jones, H.J.; Lin, X.; Massam, D.; Tudhope, D.: Augmenting Dublin Core digital library metadata with Dewey Decimal Classification (2015) 0.01
    0.011566184 = product of:
      0.034698553 = sum of:
        0.034698553 = weight(_text_:subject in 2320) [ClassicSimilarity], result of:
          0.034698553 = score(doc=2320,freq=4.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.22353725 = fieldWeight in 2320, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.03125 = fieldNorm(doc=2320)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The purpose of this paper is to describe a new approach to a well-known problem for digital libraries, how to search across multiple unrelated libraries with a single query. Design/methodology/approach - The approach involves creating new Dewey Decimal Classification terms and numbers from existing Dublin Core records. In total, 263,550 records were harvested from three digital libraries. Weighted key terms were extracted from the title, description and subject fields of each record. Ranked DDC classes were automatically generated from these key terms by considering DDC hierarchies via a series of filtering and aggregation stages. A mean reciprocal ranking evaluation compared a sample of 49 generated classes against DDC classes created by a trained librarian for the same records. Findings - The best results combined weighted key terms from the title, description and subject fields. Performance declines with increased specificity of DDC level. The results compare favorably with similar studies. Research limitations/implications - The metadata harvest required manual intervention and the evaluation was resource intensive. Future research will look at evaluation methodologies that take account of issues of consistency and ecological validity. Practical implications - The method does not require training data and is easily scalable. The pipeline can be customized for individual use cases, for example, recall or precision enhancing. Social implications - The approach can provide centralized access to information from multiple domains currently provided by individual digital libraries. Originality/value - The approach addresses metadata normalization in the context of web resources. The automatic classification approach accounts for matches within hierarchies, aggregating lower level matches to broader parents and thus approximates the practices of a human cataloger.
  16. Wisser, K.: ¬The errors of our ways : using metadata quality research to understand common error patterns in the application of name headings (2014) 0.01
    0.011278987 = product of:
      0.03383696 = sum of:
        0.03383696 = product of:
          0.06767392 = sum of:
            0.06767392 = weight(_text_:headings in 1574) [ClassicSimilarity], result of:
              0.06767392 = score(doc=1574,freq=2.0), product of:
                0.21048847 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.043400183 = queryNorm
                0.3215089 = fieldWeight in 1574, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1574)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
  17. Tosaka, Y.; Park, J.-r.: RDA: Resource description & access : a survey of the current state of the art (2013) 0.01
    0.01022316 = product of:
      0.030669477 = sum of:
        0.030669477 = weight(_text_:subject in 677) [ClassicSimilarity], result of:
          0.030669477 = score(doc=677,freq=2.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.19758089 = fieldWeight in 677, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0390625 = fieldNorm(doc=677)
      0.33333334 = coord(1/3)
    
    Abstract
    Resource Description & Access (RDA) is intended to provide a flexible and extensible framework that can accommodate all types of content and media within rapidly evolving digital environments while also maintaining compatibility with the Anglo-American Cataloguing Rules, 2nd edition (AACR2). The cataloging community is grappling with practical issues in navigating the transition from AACR2 to RDA; there is a definite need to evaluate major subject areas and broader themes in information organization under the new RDA paradigm. This article aims to accomplish this task through a thorough and critical review of the emerging RDA literature published from 2005 to 2011. The review mostly concerns key areas of difference between RDA and AACR2, the relationship of the new cataloging code to metadata standards, the impact on encoding standards such as Machine-Readable Cataloging (MARC), end user considerations, and practitioners' views on RDA implementation and training. Future research will require more in-depth studies of RDA's expected benefits and the manner in which the new cataloging code will improve resource retrieval and bibliographic control for users and catalogers alike over AACR2. The question as to how the cataloging community can best move forward to the post-AACR2/MARC environment must be addressed carefully so as to chart the future of bibliographic control in the evolving environment of information production, management, and use.
  18. Syn, S.Y.; Spring, M.B.: Finding subject terms for classificatory metadata from user-generated social tags (2013) 0.01
    0.01022316 = product of:
      0.030669477 = sum of:
        0.030669477 = weight(_text_:subject in 745) [ClassicSimilarity], result of:
          0.030669477 = score(doc=745,freq=2.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.19758089 = fieldWeight in 745, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0390625 = fieldNorm(doc=745)
      0.33333334 = coord(1/3)
    
  19. Hider, P.: ¬A survey of the coverage and methodologies of schemas and vocabularies used to describe information resources (2015) 0.01
    0.01022316 = product of:
      0.030669477 = sum of:
        0.030669477 = weight(_text_:subject in 2195) [ClassicSimilarity], result of:
          0.030669477 = score(doc=2195,freq=2.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.19758089 = fieldWeight in 2195, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2195)
      0.33333334 = coord(1/3)
    
    Abstract
    Riley's survey (2010) of metadata standards for cultural heritage collections represents a rare attempt to classify such standards, in this case according to their domain, community, function and purpose. This paper reports on a survey of metadata standards with particular functions, i.e. those of schemas and vocabularies, but that have been published online for any domain or community (and not just those of the cultural heritage sector). In total, 53 schemas and 328 vocabularies were identified as within scope, and were classified according to their subject coverage and the type of warrant used in their reported development, i.e. resource, expert or user warrant, or a combination of these types. There was found to be a general correlation between the coverage of the schemas and vocabularies. Areas of underrepresentation would appear to be the humanities and the fine arts, and, in the case of schemas, also law, engineering, manufacturing and sport. Schemas would appear to be constructed more by consulting experts and considering endusers' search behaviour; vocabularies, on the other hand, are developed more by considering the information resources themselves, or by combining a range of methods.
  20. Rousidis, D.; Garoufallou, E.; Balatsoukas, P.; Sicilia, M.-A.: Evaluation of metadata in research data repositories : the case of the DC.Subject Element (2015) 0.01
    0.01022316 = product of:
      0.030669477 = sum of:
        0.030669477 = weight(_text_:subject in 2392) [ClassicSimilarity], result of:
          0.030669477 = score(doc=2392,freq=2.0), product of:
            0.15522492 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.043400183 = queryNorm
            0.19758089 = fieldWeight in 2392, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2392)
      0.33333334 = coord(1/3)
    
    Abstract
    Research Data repositories are growing in terms of volume rapidly and exponentially. Their main goal is to provide scientists the essential mechanism to store, share, and re-use datasets generated at various stages of the research process. Despite the fact that metadata play an important role for research data management in the context of these repositories, several factors - such as the big volume of data and its complex lifecycles, as well as operational constraints related to financial resources and human factors - may impede the effectiveness of several metadata elements. The aim of the research reported in this paper was to perform a descriptive analysis of the DC.Subject metadata element and to identify its data quality problems in the context of the Dryad research data repository. In order to address this aim a total of 4.557 packages and 13.638 data files were analysed following a data-preprocessing method. The findings showed emerging trends about the subject coverage of the repository (e.g. the most popular subjects and the authors that contributed the most for these subjects). Also, quality problems related to the lack of controlled vocabulary and standardisation were very common. This study has implications for the evaluation of metadata and the improvement of the quality of the research data annotation process.

Languages

  • e 34
  • d 1
  • More… Less…

Types

  • a 32
  • el 5
  • m 2
  • s 1
  • More… Less…