Search (134 results, page 2 of 7)

  • × theme_ss:"Metadaten"
  • × year_i:[2010 TO 2020}
  1. Taniguchi, S.: Understanding RDA as a DC application profile (2013) 0.00
    0.0022155463 = product of:
      0.0044310926 = sum of:
        0.0044310926 = product of:
          0.008862185 = sum of:
            0.008862185 = weight(_text_:a in 1906) [ClassicSimilarity], result of:
              0.008862185 = score(doc=1906,freq=8.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.20383182 = fieldWeight in 1906, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1906)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The applicability of Dublin Core Application Profiles (DCAP) and the Singapore Framework for DCAPs to Resource Description and Access (RDA) were assessed. First, a draft RDA application profile is outlined, which reveals their applicability to RDA as a whole. Then, the current situation and issues involved in defining and specifying the RDA vocabularies, description structures, and syntaxes, all of which form the RDA application profile, are reviewed, for four levels of the RDA description structure; that is, the levels of aggregates and components of statements.
    Type
    a
  2. Long, K.; Thompson, S.; Potvin, S.; Rivero, M.: ¬The "wicked problem" of neutral description : toward a documentation approach to metadata standards (2017) 0.00
    0.0022155463 = product of:
      0.0044310926 = sum of:
        0.0044310926 = product of:
          0.008862185 = sum of:
            0.008862185 = weight(_text_:a in 5146) [ClassicSimilarity], result of:
              0.008862185 = score(doc=5146,freq=8.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.20383182 = fieldWeight in 5146, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5146)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Increasingly, metadata standards have been recognized as constructed rather than neutral. In this article, we argue for the importance of a documentation approach to metadata standards creation as a codification of this growing recognition. By making design decisions explicit, the documentation approach dispels presumptions of neutrality and, drawing on the "wicked problems" theoretical framework, acknowledges the constructed nature of standards as "clumsy solutions."
    Type
    a
  3. Park, H.; Smiraglia, R.P.: Enhancing data curation of cultural heritage for information sharing : a case study using open Government data (2014) 0.00
    0.0021981692 = product of:
      0.0043963385 = sum of:
        0.0043963385 = product of:
          0.008792677 = sum of:
            0.008792677 = weight(_text_:a in 1575) [ClassicSimilarity], result of:
              0.008792677 = score(doc=1575,freq=14.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.20223314 = fieldWeight in 1575, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1575)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The purpose of this paper is to enhance cultural heritage data curation. A core research question of this study is how to share cultural heritage data by using ontologies. A case study was conducted using open government data mapped with the CIDOC-CRM (Conceptual Reference Model). Twelve library-related files in unstructured data format were collected from an open government website, Seoul Metropolitan Government of Korea (http://data.seoul.go.kr). By using the ontologies of the CIDOC CRM 5.1.2, we conducted a mapping process as a way of enhancing cultural heritage information to share information as a data component. We graphed each file then mapped each file in tables. Implications of this study are both the enhanced discoverability of unstructured data and the reusability of mapped information. Issues emerging from this study involve verification of detail for complete compatibility without further input from domain experts.
    Type
    a
  4. Welhouse, Z.; Lee, J.H.; Bancroft, J.: "What am I fighting for?" : creating a controlled vocabulary for video game plot metadata (2015) 0.00
    0.0021981692 = product of:
      0.0043963385 = sum of:
        0.0043963385 = product of:
          0.008792677 = sum of:
            0.008792677 = weight(_text_:a in 2015) [ClassicSimilarity], result of:
              0.008792677 = score(doc=2015,freq=14.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.20223314 = fieldWeight in 2015, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2015)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    A video game's plot is one of its defining features, and prior research confirms the importance of plot metadata to users through persona analysis, interviews, and surveys. However, existing organizational systems, including library catalogs, game-related websites, and traditional plot classification systems, do not adequately describe the plot information of video games, in other words, what the game is really about. We attempt to address the issue by creating a controlled vocabulary based on a domain analysis involving a review of relevant literature and existing data structures. The controlled vocabulary is constructed in a pair structure for maximizing flexibility and extensibility. Adopting this controlled vocabulary for describing plot information of games will allow for useful search and collocation of video games.
    Type
    a
  5. Edmunds, J.: Roadmap to nowhere : BIBFLOW, BIBFRAME, and linked data for libraries (2017) 0.00
    0.0021981692 = product of:
      0.0043963385 = sum of:
        0.0043963385 = product of:
          0.008792677 = sum of:
            0.008792677 = weight(_text_:a in 3523) [ClassicSimilarity], result of:
              0.008792677 = score(doc=3523,freq=14.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.20223314 = fieldWeight in 3523, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3523)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    On December 12, 2016, Carl Stahmer and MacKenzie Smith presented at the CNI Members Fall Meeting about the BIBFLOW project, self-described on Twitter as "a two-year project of the UC Davis University Library and Zepheira investigating the future of library technical services." In her opening remarks, Ms. Smith, University Librarian at UC Davis, stated that one of the goals of the project was to devise a roadmap "to get from where we are today, which is kind of the 1970s with a little lipstick on it, to 2020, which is where we're going to be very soon." The notion that where libraries are today is somehow behind the times is one of the commonly heard rationales behind a move to linked data. Stated more precisely: - Libraries devote considerable time and resources to producing high-quality bibliographic metadata - This metadata is stored in unconnected silos - This metadata is in a format (MARC) that is incompatible with technologies of the emerging Semantic Web - The visibility of library metadata is diminished as a result of the two points above Are these assertions true? If yes, is linked data the solution?
    Type
    a
  6. Bellotto, A.; Bekesi, J.: Enriching metadata for a university repository by modelling and infrastructure : a new vocabulary server for Phaidra (2019) 0.00
    0.0021981692 = product of:
      0.0043963385 = sum of:
        0.0043963385 = product of:
          0.008792677 = sum of:
            0.008792677 = weight(_text_:a in 5693) [ClassicSimilarity], result of:
              0.008792677 = score(doc=5693,freq=14.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.20223314 = fieldWeight in 5693, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5693)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper illustrates an initial step towards the 'semantic enrichment' of University of Vienna's Phaidra repository as one of the valuable and up-to-date strategies able to enhance its role and usage. Firstly, a technical report points out the choice made in a local context, i.e. the deployment of the vocabulary server iQvoc instead of the formerly used SKOSMOS, explaining design decisions behind the current tool and additional features that the implementation required. Afterwards, some modelling characteristics of the local LOD controlled vocabulary are described according to SKOS documentation and best practices, highlighting which approaches can be pursued for rendering a LOD KOS available in the Web as well as issues that can be possibly encountered.
    Type
    a
  7. Peters, I.; Stock, W.G.: Power tags in information retrieval (2010) 0.00
    0.002189429 = product of:
      0.004378858 = sum of:
        0.004378858 = product of:
          0.008757716 = sum of:
            0.008757716 = weight(_text_:a in 865) [ClassicSimilarity], result of:
              0.008757716 = score(doc=865,freq=20.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.20142901 = fieldWeight in 865, product of:
                  4.472136 = tf(freq=20.0), with freq of:
                    20.0 = termFreq=20.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=865)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Purpose - Many Web 2.0 services (including Library 2.0 catalogs) make use of folksonomies. The purpose of this paper is to cut off all tags in the long tail of a document-specific tag distribution. The remaining tags at the beginning of a tag distribution are considered power tags and form a new, additional search option in information retrieval systems. Design/methodology/approach - In a theoretical approach the paper discusses document-specific tag distributions (power law and inverse-logistic shape), the development of such distributions (Yule-Simon process and shuffling theory) and introduces search tags (besides the well-known index tags) as a possibility for generating tag distributions. Findings - Search tags are compatible with broad and narrow folksonomies and with all knowledge organization systems (e.g. classification systems and thesauri), while index tags are only applicable in broad folksonomies. Based on these findings, the paper presents a sketch of an algorithm for mining and processing power tags in information retrieval systems. Research limitations/implications - This conceptual approach is in need of empirical evaluation in a concrete retrieval system. Practical implications - Power tags are a new search option for retrieval systems to limit the amount of hits. Originality/value - The paper introduces power tags as a means for enhancing the precision of search results in information retrieval systems that apply folksonomies, e.g. catalogs in Library 2.0environments.
    Type
    a
  8. Sturmane, A.; Eglite, E.; Jankevica-Balode, M.: Subject metadata development for digital resources in Latvia (2014) 0.00
    0.0021674242 = product of:
      0.0043348484 = sum of:
        0.0043348484 = product of:
          0.008669697 = sum of:
            0.008669697 = weight(_text_:a in 1963) [ClassicSimilarity], result of:
              0.008669697 = score(doc=1963,freq=10.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.19940455 = fieldWeight in 1963, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1963)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The National Library of Latvia (NLL) made a decision to use the Library of Congress Subject Headings (LCSH) in 2000. At present the NLL Subject Headings Database in Latvian holds approximately 34,000 subject headings and is used for subject cataloging of textual resources, including articles from serials. For digital objects NLL uses a system like Faceted Application of Subject Terminology (FAST). We succesfully use it in the project "In Search of Lost Latvia," one of the milestones in the development of the subject cataloging of digital resources in Latvia.
    Footnote
    Contribution in a special issue "Beyond libraries: Subject metadata in the digital environment and Semantic Web" - Enthält Beiträge der gleichnamigen IFLA Satellite Post-Conference, 17-18 August 2012, Tallinn.
    Type
    a
  9. Kleeck, D. Van; Langford, G.; Lundgren, J.; Nakano, H.; O'Dell, A.J.; Shelton, T.: Managing bibliographic data quality in a consortial academic library : a case study (2016) 0.00
    0.0021674242 = product of:
      0.0043348484 = sum of:
        0.0043348484 = product of:
          0.008669697 = sum of:
            0.008669697 = weight(_text_:a in 5133) [ClassicSimilarity], result of:
              0.008669697 = score(doc=5133,freq=10.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.19940455 = fieldWeight in 5133, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5133)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This article presents a case study of quality management for print and electronic resource metadata, summarizing problems and solutions encountered by the Cataloging and Discovery Services Department in the George A. Smathers Libraries at the University of Florida. The authors discuss national, state, and local standards for cataloging, automated and manual record enhancements for data, user feedback, and statewide consortial factors. Findings show that adherence to standards, proactive cleanup of data via manual processes and automated tools, collaboration with vendors and stakeholders, and continual assessment of workflows are key to the management of biblio-graphic data quality in consortial academic libraries.
    Type
    a
  10. Gursoy, A.; Wickett, K.; Feinberg, M.: Understanding tag functions in a moderated, user-generated metadata ecosystem (2018) 0.00
    0.0020770747 = product of:
      0.0041541494 = sum of:
        0.0041541494 = product of:
          0.008308299 = sum of:
            0.008308299 = weight(_text_:a in 3946) [ClassicSimilarity], result of:
              0.008308299 = score(doc=3946,freq=18.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.19109234 = fieldWeight in 3946, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3946)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Purpose The purpose of this paper is to investigate tag use in a metadata ecosystem that supports a fan work repository to identify functions of tags and explore the system as a co-constructed communicative context. Design/methodology/approach Using modified techniques from grounded theory (Charmaz, 2007), this paper integrates humanistic and social science methods to identify kinds of tag use in a rich setting. Findings Three primary roles of tags emerge out of detailed study of the metadata ecosystem: tags can identify elements in the fan work, tags can reflect on how those elements are used or adapted in the fan work, and finally, tags can express the fan author's sense of her role in the discursive context of the fan work repository. Attending to each of the tag roles shifts focus away from just what tags say to include how they say it. Practical implications Instead of building metadata systems designed solely for retrieval or description, this research suggests that it may be fruitful to build systems that recognize various metadata functions and allow for expressivity. This research also suggests that attending to metadata previously considered unusable in systems may reflect the participants' sense of the system and their role within it. Originality/value In addition to accommodating a wider range of tag functions, this research implies consideration of metadata ecosystems, where different kinds of tags do different things and work together to create a multifaceted artifact.
    Type
    a
  11. Bogaard, T.; Hollink, L.; Wielemaker, J.; Ossenbruggen, J. van; Hardman, L.: Metadata categorization for identifying search patterns in a digital library (2019) 0.00
    0.0020770747 = product of:
      0.0041541494 = sum of:
        0.0041541494 = product of:
          0.008308299 = sum of:
            0.008308299 = weight(_text_:a in 5281) [ClassicSimilarity], result of:
              0.008308299 = score(doc=5281,freq=18.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.19109234 = fieldWeight in 5281, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5281)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Purpose For digital libraries, it is useful to understand how users search in a collection. Investigating search patterns can help them to improve the user interface, collection management and search algorithms. However, search patterns may vary widely in different parts of a collection. The purpose of this paper is to demonstrate how to identify these search patterns within a well-curated historical newspaper collection using the existing metadata. Design/methodology/approach The authors analyzed search logs combined with metadata records describing the content of the collection, using this metadata to create subsets in the logs corresponding to different parts of the collection. Findings The study shows that faceted search is more prevalent than non-faceted search in terms of number of unique queries, time spent, clicks and downloads. Distinct search patterns are observed in different parts of the collection, corresponding to historical periods, geographical regions or subject matter. Originality/value First, this study provides deeper insights into search behavior at a fine granularity in a historical newspaper collection, by the inclusion of the metadata in the analysis. Second, it demonstrates how to use metadata categorization as a way to analyze distinct search patterns in a collection.
    Type
    a
  12. Farney, T.: using Google Tag Manager to share code : Designing shareable tags (2019) 0.00
    0.0020770747 = product of:
      0.0041541494 = sum of:
        0.0041541494 = product of:
          0.008308299 = sum of:
            0.008308299 = weight(_text_:a in 5443) [ClassicSimilarity], result of:
              0.008308299 = score(doc=5443,freq=18.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.19109234 = fieldWeight in 5443, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5443)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Sharing code between libraries is not a new phenomenon and neither is Google Tag Manager (GTM). GTM launched in 2012 as a JavaScript and HTML manager with the intent of easing the implementation of different analytics trackers and marketing scripts on a website. However, it can be used to load other code using its tag system onto a website. It's a simple process to export and import tags facilitating the code sharing process without requiring a high degree of coding experience. The entire process involves creating the script tag in GTM, exporting the GTM content into a sharable export file for someone else to import into their library's GTM container, and finally publishing that imported file to push the code to the website it was designed for. This case study provides an example of designing and sharing a GTM container loaded with advanced Google Analytics configurations such as event tracking and custom dimensions for other libraries using the Summon discovery service. It also discusses processes for designing GTM tags for export, best practices on importing and testing GTM content created by other libraries and concludes with evaluating the pros and cons of encouraging GTM use.
    Type
    a
  13. Raja, N.A.: Digitized content and index pages as alternative subject access fields (2012) 0.00
    0.002035109 = product of:
      0.004070218 = sum of:
        0.004070218 = product of:
          0.008140436 = sum of:
            0.008140436 = weight(_text_:a in 870) [ClassicSimilarity], result of:
              0.008140436 = score(doc=870,freq=12.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.18723148 = fieldWeight in 870, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=870)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This article describes a pilot study undertaken to test the benefits of the digitized Content and Index pages of books and content pages of journal Issues in providing subject access to documents in a collection. A partial digitization strategy is used to fossick specific information using the alternative subject access fields in bibliographic records. A pilot study was carried out to search for books and journal articles containing information on "Leadership., "Women Entrepreneurs., "Disinvestment. and "Digital preservation. through normal procedu re and based on information stored in MARC 21 fields 653, 505 and 520 of the bibliographic records in the University of Mumbai Library. The results are compared to draw the conclusions.
    Source
    Categories, contexts and relations in knowledge organization: Proceedings of the Twelfth International ISKO Conference 6-9 August 2012, Mysore, India. Eds.: Neelameghan, A. u. K.S. Raghavan
    Type
    a
  14. Haynes, D.: Metadata for information management and retrieval : understanding metadata and its use (2018) 0.00
    0.002035109 = product of:
      0.004070218 = sum of:
        0.004070218 = product of:
          0.008140436 = sum of:
            0.008140436 = weight(_text_:a in 4096) [ClassicSimilarity], result of:
              0.008140436 = score(doc=4096,freq=12.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.18723148 = fieldWeight in 4096, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4096)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This new and updated second edition of a classic text provides a thought-provoking introduction to metadata for all library and information students and professionals. Metadata for Information Management and Retrieval has been fully revised by David Haynes to bring it up to date with new technology and standards. The new edition, containing new chapters on Metadata Standards and Encoding Schemes, assesses the current theory and practice of metadata and examines key developments in terms of both policy and technology. Coverage includes: an introduction to the concept of metadata a description of the main components of metadata systems and standards an overview of the scope of metadata and its applications a description of typical information retrieval issues in corporate and research environments a demonstration of ways in which metadata is used to improve retrieval a look at ways in which metadata is used to manage information consideration of the role of metadata in information governance.
  15. Mi, X.M.; Pollock, B.M.: Metadata schema to facilitate linked data for 3D digital models of cultural heritage collections : a University of South Florida Libraries case study (2018) 0.00
    0.002035109 = product of:
      0.004070218 = sum of:
        0.004070218 = product of:
          0.008140436 = sum of:
            0.008140436 = weight(_text_:a in 5171) [ClassicSimilarity], result of:
              0.008140436 = score(doc=5171,freq=12.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.18723148 = fieldWeight in 5171, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5171)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The University of South Florida Libraries house and provide access to a collection of cultural heritage and 3D digital models. In an effort to provide greater access to these collections, a linked data project has been implemented. A metadata schema for the 3D cultural heritage objects which uses linked data is an excellent way to share these collections with other repositories, thus gaining global exposure and access to these valuable resources. This article will share the process of building the 3D cultural heritage metadata model as well as an assessment of the model and recommendations for future linked data projects.
    Footnote
    Beitrag in einem Heft: 'Setting standards to work and live by: A memorial Festschrift for Valerie Bross'.
    Type
    a
  16. Suranofsky, M.; McColl, L.: a Google sheets add-on that uses the WorldCat search API : MatchMarc (2019) 0.00
    0.002035109 = product of:
      0.004070218 = sum of:
        0.004070218 = product of:
          0.008140436 = sum of:
            0.008140436 = weight(_text_:a in 5442) [ClassicSimilarity], result of:
              0.008140436 = score(doc=5442,freq=12.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.18723148 = fieldWeight in 5442, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5442)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Lehigh University Libraries has developed a new tool for querying WorldCat using the WorldCat Search API. The tool is a Google Sheet Add-on and is available now via the Google Sheets Add-ons menu under the name "MatchMarc." The add-on is easily customizable, with no knowledge of coding needed. The tool will return a single "best" OCLC record number, and its bibliographic information for a given ISBN or LCCN, allowing the user to set up and define "best." Because all of the information, the input, the criteria, and the results exist in the Google Sheets environment, efficient workflows can be developed from this flexible starting point. This article will discuss the development of the add-on, how it works, and future plans for development.
    Type
    a
  17. Neumann, M.; Steinberg, J.; Schaer, P.: Web-ccraping for non-programmers : introducing OXPath for digital library metadata harvesting (2017) 0.00
    0.0019582848 = product of:
      0.0039165695 = sum of:
        0.0039165695 = product of:
          0.007833139 = sum of:
            0.007833139 = weight(_text_:a in 3895) [ClassicSimilarity], result of:
              0.007833139 = score(doc=3895,freq=16.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.18016359 = fieldWeight in 3895, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3895)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Building up new collections for digital libraries is a demanding task. Available data sets have to be extracted which is usually done with the help of software developers as it involves custom data handlers or conversion scripts. In cases where the desired data is only available on the data provider's website custom web scrapers are needed. This may be the case for small to medium-size publishers, research institutes or funding agencies. As data curation is a typical task that is done by people with a library and information science background, these people are usually proficient with XML technologies but are not full-stack programmers. Therefore we would like to present a web scraping tool that does not demand the digital library curators to program custom web scrapers from scratch. We present the open-source tool OXPath, an extension of XPath, that allows the user to define data to be extracted from websites in a declarative way. By taking one of our own use cases as an example, we guide you in more detail through the process of creating an OXPath wrapper for metadata harvesting. We also point out some practical things to consider when creating a web scraper (with OXPath). On top of that, we also present a syntax highlighting plugin for the popular text editor Atom that we developed to further support OXPath users and to simplify the authoring process.
    Type
    a
  18. Stevens, G.: New metadata recipes for old cookbooks : creating and analyzing a digital collection using the HathiTrust Research Center Portal (2017) 0.00
    0.0019582848 = product of:
      0.0039165695 = sum of:
        0.0039165695 = product of:
          0.007833139 = sum of:
            0.007833139 = weight(_text_:a in 3897) [ClassicSimilarity], result of:
              0.007833139 = score(doc=3897,freq=16.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.18016359 = fieldWeight in 3897, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3897)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The Early American Cookbooks digital project is a case study in analyzing collections as data using HathiTrust and the HathiTrust Research Center (HTRC) Portal. The purposes of the project are to create a freely available, searchable collection of full-text early American cookbooks within the HathiTrust Digital Library, to offer an overview of the scope and contents of the collection, and to analyze trends and patterns in the metadata and the full text of the collection. The digital project has two basic components: a collection of 1450 full-text cookbooks published in the United States between 1800 and 1920 and a website to present a guide to the collection and the results of the analysis. This article will focus on the workflow for analyzing the metadata and the full-text of the collection. The workflow will cover: 1) creating a searchable public collection of full-text titles within the HathiTrust Digital Library and uploading it to the HTRC Portal, 2) analyzing and visualizing legacy MARC data for the collection using MarcEdit, OpenRefine and Tableau, and 3) using the text analysis tools in the HTRC Portal to look for trends and patterns in the full text of the collection.
    Type
    a
  19. Stiller, J.; Olensky, M.; Petras, V.: ¬A framework for the evaluation of automatic metadata enrichments (2014) 0.00
    0.001938603 = product of:
      0.003877206 = sum of:
        0.003877206 = product of:
          0.007754412 = sum of:
            0.007754412 = weight(_text_:a in 1587) [ClassicSimilarity], result of:
              0.007754412 = score(doc=1587,freq=8.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.17835285 = fieldWeight in 1587, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1587)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Automatic enrichment of collections connects data to vocabularies, which supports the contextualization of content and adds searchable text to metadata. The paper introduces a framework of four dimensions (frequency, coverage, relevance and error rate) that measure both the suitability of the enrichment for the object and the enrichments' contribution to search success. To verify the framework, it is applied to the evaluation of automatic enrichments in the digital library Europeana. The analysis of 100 result sets and their corresponding queries (1,121 documents total) shows the framework is a valuable tool for guiding enrichments and determining the value of enrichment efforts.
    Type
    a
  20. Radio, E.: Semiotic principles for metadata auditing and evaluation (2016) 0.00
    0.001938603 = product of:
      0.003877206 = sum of:
        0.003877206 = product of:
          0.007754412 = sum of:
            0.007754412 = weight(_text_:a in 4078) [ClassicSimilarity], result of:
              0.007754412 = score(doc=4078,freq=8.0), product of:
                0.043477926 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.037706986 = queryNorm
                0.17835285 = fieldWeight in 4078, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4078)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Metadata auditing is a common practice used to assess quality by drawing on a selection of documents to create a representative sample of the corpus. This article examines common techniques involved in auditing and argues that underlying semiotic principles should be considered before remediation and transformation work are undertaken. This article provides an inquiry into the structural nature of metadata, records, and corpora to analyze the semiotic processes and boundaries that can affect audits. Suggestions for effective auditing practices are offered as well as what can be obfuscated when transformations are undertaken without concern for the sign-functions of metadata.
    Type
    a

Authors

Languages

  • e 116
  • d 17
  • pt 1
  • More… Less…

Types

  • a 117
  • el 25
  • m 11
  • s 5
  • n 1
  • x 1
  • More… Less…

Subjects