Search (320 results, page 16 of 16)

  • × theme_ss:"Metadaten"
  1. Rousidis, D.; Garoufallou, E.; Balatsoukas, P.; Sicilia, M.-A.: Evaluation of metadata in research data repositories : the case of the DC.Subject Element (2015) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 2392) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=2392,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 2392, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2392)
      0.25 = coord(1/4)
    
    Series
    Communications in computer and information science; 544
  2. Bohne-Lang, A.: Semantische Metadaten für den Webauftritt einer Bibliothek (2016) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 3337) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=3337,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 3337, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3337)
      0.25 = coord(1/4)
    
    Source
    GMS Medizin - Bibliothek - Information. 16(2016) Nr.3, 11 S. [http://www.egms.de/static/pdf/journals/mbi/2017-16/mbi000372.pdf]
  3. Strobel, S.; Marín-Arraiza, P.: Metadata for scientific audiovisual media : current practices and perspectives of the TIB / AV-portal (2015) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 3667) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=3667,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 3667, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3667)
      0.25 = coord(1/4)
    
    Series
    Communications in computer and information science; 544
  4. Neumann, M.; Steinberg, J.; Schaer, P.: Web-ccraping for non-programmers : introducing OXPath for digital library metadata harvesting (2017) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 3895) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=3895,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 3895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3895)
      0.25 = coord(1/4)
    
    Abstract
    Building up new collections for digital libraries is a demanding task. Available data sets have to be extracted which is usually done with the help of software developers as it involves custom data handlers or conversion scripts. In cases where the desired data is only available on the data provider's website custom web scrapers are needed. This may be the case for small to medium-size publishers, research institutes or funding agencies. As data curation is a typical task that is done by people with a library and information science background, these people are usually proficient with XML technologies but are not full-stack programmers. Therefore we would like to present a web scraping tool that does not demand the digital library curators to program custom web scrapers from scratch. We present the open-source tool OXPath, an extension of XPath, that allows the user to define data to be extracted from websites in a declarative way. By taking one of our own use cases as an example, we guide you in more detail through the process of creating an OXPath wrapper for metadata harvesting. We also point out some practical things to consider when creating a web scraper (with OXPath). On top of that, we also present a syntax highlighting plugin for the popular text editor Atom that we developed to further support OXPath users and to simplify the authoring process.
  5. Çelebi, A.; Özgür, A.: Segmenting hashtags and analyzing their grammatical structure (2018) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 4221) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=4221,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 4221, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4221)
      0.25 = coord(1/4)
    
    Source
    Journal of the Association for Information Science and Technology. 69(2018) no.5, S.675-686
  6. Hansson, K.; Dahlgren, A.: Open research data repositories : practices, norms, and metadata for sharing images (2022) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 472) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=472,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 472, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=472)
      0.25 = coord(1/4)
    
    Source
    Journal of the Association for Information Science and Technology. 73(2022) no.2, S.303-316
  7. Lee, S.: Pidgin metadata framework as a mediator for metadata interoperability (2021) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 654) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=654,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 654, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=654)
      0.25 = coord(1/4)
    
    Abstract
    A pidgin metadata framework based on the concept of pidgin metadata is proposed to complement the limitations of existing approaches to metadata interoperability and to achieve more reliable metadata interoperability. The framework consists of three layers, with a hierarchical structure, and reflects the semantic and structural characteristics of various metadata. Layer 1 performs both an external function, serving as an anchor for semantic association between metadata elements, and an internal function, providing semantic categories that can encompass detailed elements. Layer 2 is an arbitrary layer composed of substantial elements from existing metadata and performs a function in which different metadata elements describing the same or similar aspects of information resources are associated with the semantic categories of Layer 1. Layer 3 implements the semantic relationships between Layer 1 and Layer 2 through the Resource Description Framework syntax. With this structure, the pidgin metadata framework can establish the criteria for semantic connection between different elements and fully reflect the complexity and heterogeneity among various metadata. Additionally, it is expected to provide a bibliographic environment that can achieve more reliable metadata interoperability than existing approaches by securing the communication between metadata.
  8. Qin, C.; Liu, Y.; Ma, X.; Chen, J.; Liang, H.: Designing for serendipity in online knowledge communities : an investigation of tag presentation formats and openness to experience (2022) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 664) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=664,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 664, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=664)
      0.25 = coord(1/4)
    
    Source
    Journal of the Association for Information Science and Technology. 73(2022) no.10, S.1401-1417
  9. Laparra, E.; Binford-Walsh, A.; Emerson, K.; Miller, M.L.; López-Hoffman, L.; Currim, F.; Bethard, S.: Addressing structural hurdles for metadata extraction from environmental impact statements (2023) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 1042) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=1042,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 1042, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1042)
      0.25 = coord(1/4)
    
    Source
    Journal of the Association for Information Science and Technology. 74(2023) no.9, S.1124-1139
  10. Heery, R.; Wagner, H.: ¬A metadata registry for the Semantic Web (2002) 0.00
    0.0014723204 = product of:
      0.0058892816 = sum of:
        0.0058892816 = weight(_text_:information in 1210) [ClassicSimilarity], result of:
          0.0058892816 = score(doc=1210,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.0960027 = fieldWeight in 1210, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1210)
      0.25 = coord(1/4)
    
    Abstract
    * Agencies maintaining directories of data elements in a domain area in accordance with ISO/IEC 11179 (This standard specifies good practice for data element definition as well as the registration process. Example implementations are the National Health Information Knowledgebase hosted by the Australian Institute of Health and Welfare and the Environmental Data Registry hosted by the US Environmental Protection Agency.); * The xml.org directory of the Extended Markup Language (XML) document specifications facilitating re-use of Document Type Definition (DTD), hosted by the Organization for the Advancement of Structured Information Standards (OASIS); * The MetaForm database of Dublin Core usage and mappings maintained at the State and University Library in Goettingen; * The Semantic Web Agreement Group Dictionary, a database of terms for the Semantic Web that can be referred to by humans and software agents; * LEXML, a multi-lingual and multi-jurisdictional RDF Dictionary for the legal world; * The SCHEMAS registry maintained by the European Commission funded SCHEMAS project, which indexes several metadata element sets as well as a large number of activity reports describing metadata related activities and initiatives. Metadata registries essentially provide an index of terms. Given the distributed nature of the Web, there are a number of ways this can be accomplished. For example, the registry could link to terms and definitions in schemas published by implementers and stored locally by the schema maintainer. Alternatively, the registry might harvest various metadata schemas from their maintainers. Registries provide 'added value' to users by indexing schemas relevant to a particular 'domain' or 'community of use' and by simplifying the navigation of terms by enabling multiple schemas to be accessed from one view. An important benefit of this approach is an increase in the reuse of existing terms, rather than users having to reinvent them. Merging schemas to one view leads to harmonization between applications and helps avoid duplication of effort. Additionally, the establishment of registries to index terms actively being used in local implementations facilitates the metadata standards activity by providing implementation experience transferable to the standards-making process.
  11. Gömpel, R.; Altenhöner, R.; Kunz, M.; Oehlschläger, S.; Werner, C.: Weltkongress Bibliothek und Information, 70. IFLA-Generalkonferenz in Buenos Aires : Aus den Veranstaltungen der Division IV Bibliographic Control, der Core Activities ICABS und UNIMARC sowie der Information Technology Section (2004) 0.00
    0.0014572192 = product of:
      0.005828877 = sum of:
        0.005828877 = weight(_text_:information in 2874) [ClassicSimilarity], result of:
          0.005828877 = score(doc=2874,freq=12.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09501803 = fieldWeight in 2874, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.015625 = fieldNorm(doc=2874)
      0.25 = coord(1/4)
    
    Abstract
    "Libraries: Tools for Education and Development" war das Motto der 70. IFLA-Generalkonferenz, dem Weltkongress Bibliothek und Information, der vom 22.-27. August 2004 in Buenos Aires, Argentinien, und damit erstmals in Lateinamerika stattfand. Rund 3.000 Teilnehmerinnen und Teilnehmer, davon ein Drittel aus spanischsprachigen Ländern, allein 600 aus Argentinien, besuchten die von der IFLA und dem nationalen Organisationskomitee gut organisierte Tagung mit mehr als 200 Sitzungen und Veranstaltungen. Aus Deutschland waren laut Teilnehmerverzeichnis leider nur 45 Kolleginnen und Kollegen angereist, womit ihre Zahl wieder auf das Niveau von Boston gesunken ist. Erfreulicherweise gab es nunmehr bereits im dritten Jahr eine deutschsprachige Ausgabe des IFLA-Express. Auch in diesem Jahr soll hier über die Veranstaltungen der Division IV Bibliographic Control berichtet werden. Die Arbeit der Division mit ihren Sektionen Bibliography, Cataloguing, Classification and Indexing sowie der neuen Sektion Knowledge Management bildet einen der Schwerpunkte der IFLA-Arbeit, die dabei erzielten konkreten Ergebnisse und Empfehlungen haben maßgeblichen Einfluss auf die tägliche Arbeit der Bibliothekarinnen und Bibliothekare. Erstmals wird auch ausführlich über die Arbeit der Core Activities ICABS und UNIMARC und der Information Technology Section berichtet.
    Content
    Cataloguing Section (Sektion Katalogisierung) Der Schwerpunkt der Arbeit dieser Sektion liegt auf der Erarbeitung bzw. internationalen Harmonisierung von Strukturen, Regeln und Arbeitsverfahren mit dem Ziel, die internationale Kooperation im Katalogisierungsbereich zu verbessern. In Anbetracht des laufenden Evaluierungsprozesses wurde der Strategieplan der Sektion zum jetzigen Zeitpunkt nur dort aktualisiert, wo es unbedingt erforderlich war. Neue Ziele wurden nicht aufgenommen. Oberste Priorität bei den strategischen Zielen behielt die Entwicklung internationaler Katalogisierungsstandards für die bibliographische Beschreibung und den Zugriff. In ihrer zentralen Bedeutung bestätigt wurden auch die "Functional Requirements for Bibliographic Records" (FRBR). Darüber hinaus gehört auch in Zukunft die Weiterentwicklung und Revision der ISBDs zu den zentralen Anliegen der Arbeit der Sektion Katalogisierung. Ein weiteres vorrangiges Ziel bleibt die Erarbeitung von Standards, Regeln und Informationslisten, um den Zugang zu bibliographischen Daten in allen Sprachen zu ermöglichen. Hierzu zählen u. a.: - die vollständige Veröffentlichung der Anonymous Classics: Der Teil für europäische Literatur ist inzwischen veröffentlicht'. Für die Erarbeitung weiterer Teile (Lateinamerika, Afrika und Asien) soll das Verfahren gestrafft und ein Zeitplan erstellt werden. - die Beobachtung der Aktivitäten zu Unicode und die Information der Sektionsmitglieder darüber zur Förderung des mehrsprachigen Zugangs zu bibliographischer Information - die Entwicklung eines web-basierten multilingualen Wörterbuchs für Katalogisierungsbegriffe - die Entwicklung und der Test von (Daten-)Modellen für eine virtuelle internationale Normdatei - die Überarbeitung der "IFLA Names of persons". Das Open Programme der Sektion stand in diesem Jahr unter dem Motto "Developments in Cataloguing Guidelines" und wurde von Barbara Tillett, Lynne Howarth und Carol van Nuys bestritten. Lynne Howarth ging in ihrem Vortrag "Enabling metadata: creating a core record for resource discovery" auf die Reaktionen im weltweiten Stellungnahmeverfahren auf die Veröffentlichung des Papiers "Guidance an the Structure, Content and Application of Metadata Records for digital resources and collections" der Working Group an the Use of Metadata Schemes ein. Carol van Nuys stellte das norwegische "Paradigma Project and its quest for metadata solutions and services" vor.
  12. Bazillion, R.J.; Caplan, P.: Metadata fundamentals for all librarians (2003) 0.00
    0.0011898145 = product of:
      0.004759258 = sum of:
        0.004759258 = weight(_text_:information in 3197) [ClassicSimilarity], result of:
          0.004759258 = score(doc=3197,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.0775819 = fieldWeight in 3197, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=3197)
      0.25 = coord(1/4)
    
    Footnote
    Rez.: JASIST 56(2005) no.13, S.1264 (W. Koehler: "Priscilla Caplan provides us with a sweeping but very welcome survey of the various approaches to metadata in practice or proposed in libraries and archives today. One of the key strengths of the book and paradoxically one of its key weaknesses is that the work is descriptive in nature. While relationships between one system and another may be noted, no general conclusions of a practical or theoretical nature are drawn of the relative merits of one metadata or metametadata scheure as against another. That said, let us remember that this is an American Library Association publication, published as a descriptive resource. Caplan does very well what she sets out to do. The work is divided into two parts: "Principles and Practice" and "Metadata Schemes," and is further subdivided into eighteen chapters. The book begins with short yet more than adequate chapters defining terms, vocabularies, and concepts. It discusses interoperability and the various levels of quality among systems. Perhaps Chapter 5, "Metadata and the Web" is the weakest chapter of the work. There is a brief discussion of how search engines work and some of the more recent initiatives (e.g., the Semantic Web) to develop better retrieval agents. The chapter is weck not in its description but in what it fails to discuss. The second section, "Metadata Schemes," which encompasses chapters six through eighteen, is particularly rich. Thirteen different metadata or metametadata schema are described to provide the interested librarian with a better than adequate introduction to the purpose, application, and operability of each metadata scheme. These are: library cataloging (chiefly MARC), TEI, Dublin Core, Archival Description and EAD, Art and Architecture, GILS, Education, ONIX, Geospatial, Data Documentation Initiative, Administrative Metadata, Structural Metadata, and Rights Metadata. The last three chapters introduce concepts heretofore "foreign" to the realm of the catalog or metadata. Descriptive metadata was . . . intended to help in finding, discovering, and identifying an information resource." (p. 151) Administrative metadata is an aid to ". . . the owners or caretakers of the resource." Structural metadata describe the relationships of data elements. Rights metadata describe (or as Caplan points out, may describe, as definition is still as yet ambiguous) end user rights to use and reproduce material in digital format. Keeping in mind that the work is intended for the general practitioner librarian, the book has a particularly useful glossary and index. Caplan also provides useful suggestions for additional reading at the end of each chapter. 1 intend to adopt Metadata Fundamentals for All Librarians when next I teach a digital cataloging course. Caplan's book provides an excellent introduction to the basic concepts. It is, however, not a "cookbook" nor a guidebook into the complexities of the application of any metadata scheme."
  13. Cwiok, J.: ¬The defining element : a discussion of the creator element within metadata schemas (2005) 0.00
    0.0011898145 = product of:
      0.004759258 = sum of:
        0.004759258 = weight(_text_:information in 5732) [ClassicSimilarity], result of:
          0.004759258 = score(doc=5732,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.0775819 = fieldWeight in 5732, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=5732)
      0.25 = coord(1/4)
    
    Abstract
    The speed with change takes place is startling and has left the information community with little time to consider how the development of electronic resources and the metadata schemas created to describe them effect how we view a work and its components. In terms of the attribution of authorship in the context of electronic works, this is a salient point. How does one determine authorship of a complex electronic resource, which is the culmination of the work of a myriad of entities? How does one determine the authorship when the content of the electronic resource may change at any moment without warning? What is the semantic content of the element that denotes authorship or responsibility for an electronic resource and how does the term used determine the element's meaning? The conceptual difficulty in the definition of the Creator element is deciphering what exactly the metadata schema should be describing. We also need to establish what purpose the element is intended to serve. In essence, we are at a crossroads. It is clear that once a work is digitized it exists in a significantly different medium, but how do we provide access to it? It is necessary to critically assess the accuracy of digital surrogates and to note that webmasters have a significant amount of intellectual responsibility invested in the sites they create. The solution to the problem in the Creator element may lie in moving from the concept of "authorship" and "origination" to a concept of intellectual responsibility. Perhaps the problematic nature of the Creator element allows us to move forward in our assessment and treatment of knowledge. One solution may be to standardize the definitions within various element sets. As the semantic web continues to grow and librarians strive to catalog electronic resources, the establishment of standard definitions for elements is becoming more relevant and important.
  14. Miksa, S.D.; Moen, WE.; Snyder, G.; Polyakov, S.; Eklund, A.: Metadata assistance of the Functional Requirements for Bibliographic Record's four user tasks : a report on the MARC content designation utilization (MCDU) project (2006) 0.00
    0.0011898145 = product of:
      0.004759258 = sum of:
        0.004759258 = weight(_text_:information in 125) [ClassicSimilarity], result of:
          0.004759258 = score(doc=125,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.0775819 = fieldWeight in 125, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=125)
      0.25 = coord(1/4)
    
    Abstract
    This paper describes the work of the MARC Content Designation Utilization (MCDU) Project, funded by a National Leadership Grant from the U.S. Federal Institute of Museum and Library Services (IMLS). The MCDU Project is analyzing approximately 56 million MARC 21 Format for Bibliographic Data records from OCLC's WorldCat database to identify actual use of the content designation available in the MARC bibliographic record. We consider bibliographic records as artifacts resulting from the overall cataloging enterprise, of which the encoding of the bibliographic data into MARC is only one part. Concepts from the Functional Requirements for Bibliographic Records (FRBR) can be used to examine and critically assess the availability of bibliographic data in these records, data meant to assist end users in finding, identifying, selecting, and obtaining relevant information resources. Overall, the MCDU Project will provide empirical data reflecting the actual use of MARC content designation structures in this set of records. Specifically, the data can be used to demonstrate how catalogers' coding of bibliographic data may or may not assist end users in these four tasks. The project is using the mapping by Delsey of MARC data elements to FRBR user tasks in this analysis. These data are crucial for making decisions about the future of MARC and may inform current work on bibliographic rules reflected in the development of the next version of cataloging rules (i.e., Resource Description and Access) by the Joint Steering Committee for the Revision of the Anglo-American Cataloguing Rules.
  15. Baker, T.; Dekkers, M.: Identifying metadata elements with URIs : The CORES resolution (2003) 0.00
    0.0011898145 = product of:
      0.004759258 = sum of:
        0.004759258 = weight(_text_:information in 1199) [ClassicSimilarity], result of:
          0.004759258 = score(doc=1199,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.0775819 = fieldWeight in 1199, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=1199)
      0.25 = coord(1/4)
    
    Abstract
    On 18 November 2002, at a meeting organised by the CORES Project (Information Society Technologies Programme, European Union), several organisations regarded as maintenance authorities for metadata elements achieved consensus on a resolution to assign Uniform Resource Identifiers (URIs) to metadata elements as a useful first step towards the development of mapping infrastructures and interoperability services. The signatories of the CORES Resolution agreed to promote this consensus in their communities and beyond and to implement an action plan in the following six months. Six months having passed, the maintainers of GILS, ONIX, MARC 21, CERIF, DOI, IEEE/LOM, and Dublin Core report on their implementations of the resolution and highlight issues of relevance to establishing good-practice conventions for declaring, identifying, and maintaining metadata elements more generally. In June 2003, the resolution was also endorsed by the maintainers of UNIMARC. The "Resolution on Metadata Element Identifiers", or CORES Resolution, is an agreement among the maintenance organisations for several major metadata standards - GILS, ONIX, MARC 21, UNIMARC, CERIF, DOI®, IEEE/LOM, and Dublin Core - to identify their metadata elements using Uniform Resource Identifiers (URIs). The Uniform Resource Identifier, defined in the IETF RFC 2396 as "a compact string of characters for identifying an abstract or physical resource", has been promoted for use as a universal form of identification by the World Wide Web Consortium. The CORES Resolution, formulated at a meeting organised by the European project CORES in November 2002, included a commitment to publicise the consensus statement to a wider audience of metadata standards initiatives and to implement key points of the agreement within the following six months - specifically, to define URI assignment mechanisms, assign URIs to elements, and formulate policies for the persistence of those URIs. This article marks the passage of six months by reporting on progress made in implementing this common action plan. After presenting the text of the CORES Resolution and its three "clarifications", the article summarises the position of each signatory organisation towards assigning URIs to its metadata elements, noting any practical or strategic problems that may have emerged. These progress reports were based on input from Thomas Baker, José Borbinha, Eliot Christian, Erik Duval, Keith Jeffery, Rebecca Guenther, and Norman Paskin. The article closes with a few general observations about these first steps towards the clarification of shared conventions for the identification of metadata elements and perhaps, one can hope, towards the ultimate goal of improving interoperability among a diversity of metadata communities.
  16. Khoo, M.J.; Ahn, J.-w.; Binding, C.; Jones, H.J.; Lin, X.; Massam, D.; Tudhope, D.: Augmenting Dublin Core digital library metadata with Dewey Decimal Classification (2015) 0.00
    0.0011898145 = product of:
      0.004759258 = sum of:
        0.004759258 = weight(_text_:information in 2320) [ClassicSimilarity], result of:
          0.004759258 = score(doc=2320,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.0775819 = fieldWeight in 2320, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=2320)
      0.25 = coord(1/4)
    
    Abstract
    Purpose - The purpose of this paper is to describe a new approach to a well-known problem for digital libraries, how to search across multiple unrelated libraries with a single query. Design/methodology/approach - The approach involves creating new Dewey Decimal Classification terms and numbers from existing Dublin Core records. In total, 263,550 records were harvested from three digital libraries. Weighted key terms were extracted from the title, description and subject fields of each record. Ranked DDC classes were automatically generated from these key terms by considering DDC hierarchies via a series of filtering and aggregation stages. A mean reciprocal ranking evaluation compared a sample of 49 generated classes against DDC classes created by a trained librarian for the same records. Findings - The best results combined weighted key terms from the title, description and subject fields. Performance declines with increased specificity of DDC level. The results compare favorably with similar studies. Research limitations/implications - The metadata harvest required manual intervention and the evaluation was resource intensive. Future research will look at evaluation methodologies that take account of issues of consistency and ecological validity. Practical implications - The method does not require training data and is easily scalable. The pipeline can be customized for individual use cases, for example, recall or precision enhancing. Social implications - The approach can provide centralized access to information from multiple domains currently provided by individual digital libraries. Originality/value - The approach addresses metadata normalization in the context of web resources. The automatic classification approach accounts for matches within hierarchies, aggregating lower level matches to broader parents and thus approximates the practices of a human cataloger.
  17. Managing metadata in web-scale discovery systems (2016) 0.00
    0.0011898145 = product of:
      0.004759258 = sum of:
        0.004759258 = weight(_text_:information in 3336) [ClassicSimilarity], result of:
          0.004759258 = score(doc=3336,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.0775819 = fieldWeight in 3336, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=3336)
      0.25 = coord(1/4)
    
    Abstract
    This book shows you how to harness the power of linked data and web-scale discovery systems to manage and link widely varied content across your library collection. Libraries are increasingly using web-scale discovery systems to help clients find a wide assortment of library materials, including books, journal articles, special collections, archival collections, videos, music and open access collections. Depending on the library material catalogued, the discovery system might need to negotiate different metadata standards, such as AACR, RDA, RAD, FOAF, VRA Core, METS, MODS, RDF and more. In Managing Metadata in Web-Scale Discovery Systems, editor Louise Spiteri and a range of international experts show you how to: * maximize the effectiveness of web-scale discovery systems * provide a smooth and seamless discovery experience to your users * help users conduct searches that yield relevant results * manage the sheer volume of items to which you can provide access, so your users can actually find what they need * maintain shared records that reflect the needs, languages, and identities of culturally and ethnically varied communities * manage metadata both within, across, and outside, library discovery tools by converting your library metadata to linked open data that all systems can access * manage user generated metadata from external services such as Goodreads and LibraryThing * mine user generated metadata to better serve your users in areas such as collection development or readers' advisory. The book will be essential reading for cataloguers, technical services and systems librarians and library and information science students studying modules on metadata, cataloguing, systems design, data management, and digital libraries. The book will also be of interest to those managing metadata in archives, museums and other cultural heritage institutions.
  18. Chen, J.; Wang, D.; Xie, I.; Lu, Q.: Image annotation tactics : transitions, strategies and efficiency (2018) 0.00
    0.0011898145 = product of:
      0.004759258 = sum of:
        0.004759258 = weight(_text_:information in 5046) [ClassicSimilarity], result of:
          0.004759258 = score(doc=5046,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.0775819 = fieldWeight in 5046, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=5046)
      0.25 = coord(1/4)
    
    Source
    Information processing and management. 54(2018) no.6, S.985-1001
  19. Haslhofer, B.: ¬A Web-based mapping technique for establishing metadata interoperability (2008) 0.00
    0.0010516574 = product of:
      0.0042066295 = sum of:
        0.0042066295 = weight(_text_:information in 3173) [ClassicSimilarity], result of:
          0.0042066295 = score(doc=3173,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.068573356 = fieldWeight in 3173, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.01953125 = fieldNorm(doc=3173)
      0.25 = coord(1/4)
    
    Abstract
    The integration of metadata from distinct, heterogeneous data sources requires metadata interoperability, which is a qualitative property of metadata information objects that is not given by default. The technique of metadata mapping allows domain experts to establish metadata interoperability in a certain integration scenario. Mapping solutions, as a technical manifestation of this technique, are already available for the intensively studied domain of database system interoperability, but they rarely exist for the Web. If we consider the amount of steadily increasing structured metadata and corresponding metadata schemes on theWeb, we can observe a clear need for a mapping solution that can operate in aWeb-based environment. To achieve that, we first need to build its technical core, which is a mapping model that provides the language primitives to define mapping relationships. Existing SemanticWeb languages such as RDFS and OWL define some basic mapping elements (e.g., owl:equivalentProperty, owl:sameAs), but do not address the full spectrum of semantic and structural heterogeneities that can occur among distinct, incompatible metadata information objects. Furthermore, it is still unclear how to process defined mapping relationships during run-time in order to deliver metadata to the client in a uniform way. As the main contribution of this thesis, we present an abstract mapping model, which reflects the mapping problem on a generic level and provides the means for reconciling incompatible metadata. Instance transformation functions and URIs take a central role in that model. The former cover a broad spectrum of possible structural and semantic heterogeneities, while the latter bind the complete mapping model to the architecture of the Word Wide Web. On the concrete, language-specific level we present a binding of the abstract mapping model for the RDF Vocabulary Description Language (RDFS), which allows us to create mapping specifications among incompatible metadata schemes expressed in RDFS. The mapping model is embedded in a cyclic process that categorises the requirements a mapping solution should fulfil into four subsequent phases: mapping discovery, mapping representation, mapping execution, and mapping maintenance. In this thesis, we mainly focus on mapping representation and on the transformation of mapping specifications into executable SPARQL queries. For mapping discovery support, the model provides an interface for plugging-in schema and ontology matching algorithms. For mapping maintenance we introduce the concept of a simple, but effective mapping registry. Based on the mapping model, we propose aWeb-based mediator wrapper-architecture that allows domain experts to set up mediation endpoints that provide a uniform SPARQL query interface to a set of distributed metadata sources. The involved data sources are encapsulated by wrapper components that expose the contained metadata and the schema definitions on the Web and provide a SPARQL query interface to these metadata. In this thesis, we present the OAI2LOD Server, a wrapper component for integrating metadata that are accessible via the Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH). In a case study, we demonstrate how mappings can be created in aWeb environment and how our mediator wrapper architecture can easily be configured in order to integrate metadata from various heterogeneous data sources without the need to install any mapping solution or metadata integration solution in a local system environment.
  20. Baker, T.: Languages for Dublin Core (1998) 0.00
    0.0010410878 = product of:
      0.004164351 = sum of:
        0.004164351 = weight(_text_:information in 1257) [ClassicSimilarity], result of:
          0.004164351 = score(doc=1257,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.06788416 = fieldWeight in 1257, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1257)
      0.25 = coord(1/4)
    
    Abstract
    Over the past three years, the Dublin Core Metadata Initiative has achieved a broad international consensus on the semantics of a simple element set for describing electronic resources. Since the first workshop in March 1995, which was reported in the very first issue of D-Lib Magazine, Dublin Core has been the topic of perhaps a dozen articles here. Originally intended to be simple and intuitive enough for authors to tag Web pages without special training, Dublin Core is being adapted now for more specialized uses, from government information and legal deposit to museum informatics and electronic commerce. To meet such specialized requirements, Dublin Core can be customized with additional elements or qualifiers. However, these refinements can compromise interoperability across applications. There are tradeoffs between using specific terms that precisely meet local needs versus general terms that are understood more widely. We can better understand this inevitable tension between simplicity and complexity if we recognize that metadata is a form of human language. With Dublin Core, as with a natural language, people are inclined to stretch definitions, make general terms more specific, specific terms more general, misunderstand intended meanings, and coin new terms. One goal of this paper, therefore, will be to examine the experience of some related ways to seek semantic interoperability through simplicity: planned languages, interlingua constructs, and pidgins. The problem of semantic interoperability is compounded when we consider Dublin Core in translation. All of the workshops, documents, mailing lists, user guides, and working group outputs of the Dublin Core Initiative have been in English. But in many countries and for many applications, people need a metadata standard in their own language. In principle, the broad elements of Dublin Core can be defined equally well in Bulgarian or Hindi. Since Dublin Core is a controlled standard, however, any parallel definitions need to be kept in sync as the standard evolves. Another goal of the paper, then, will be to define the conceptual and organizational problem of maintaining a metadata standard in multiple languages. In addition to a name and definition, which are meant for human consumption, each Dublin Core element has a label, or indexing token, meant for harvesting by search engines. For practical reasons, these machine-readable tokens are English-looking strings such as Creator and Subject (just as HTML tags are called HEAD, BODY, or TITLE). These tokens, which are shared by Dublin Cores in every language, ensure that metadata fields created in any particular language are indexed together across repositories. As symbols of underlying universal semantics, these tokens form the basis of semantic interoperability among the multiple Dublin Cores. As long as we limit ourselves to sharing these indexing tokens among exact translations of a simple set of fifteen broad elements, the definitions of which fit easily onto two pages, the problem of Dublin Core in multiple languages is straightforward. But nothing having to do with human language is ever so simple. Just as speakers of various languages must learn the language of Dublin Core in their own tongues, we must find the right words to talk about a metadata language that is expressable in many discipline-specific jargons and natural languages and that inevitably will evolve and change over time.

Years

Languages

Types

  • a 279
  • el 33
  • m 24
  • s 13
  • n 4
  • b 2
  • x 2
  • r 1
  • More… Less…

Subjects