Search (266 results, page 1 of 14)

  • × theme_ss:"Metadaten"
  1. Greenberg, J.: Metadata and the World Wide Web (2002) 0.03
    0.02607361 = product of:
      0.12167685 = sum of:
        0.035931468 = weight(_text_:open in 4264) [ClassicSimilarity], result of:
          0.035931468 = score(doc=4264,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.24876907 = fieldWeight in 4264, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4264)
        0.04354783 = weight(_text_:source in 4264) [ClassicSimilarity], result of:
          0.04354783 = score(doc=4264,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.27386856 = fieldWeight in 4264, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4264)
        0.042197548 = weight(_text_:web in 4264) [ClassicSimilarity], result of:
          0.042197548 = score(doc=4264,freq=10.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.40312994 = fieldWeight in 4264, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4264)
      0.21428572 = coord(3/14)
    
    Abstract
    Metadata is of paramount importance for persons, organizations, and endeavors of every dimension that are increasingly turning to the World Wide Web (hereafter referred to as the Web) as a chief conduit for accessing and disseminating information. This is evidenced by the development and implementation of metadata schemas supporting projects ranging from restricted corporate intranets, data warehouses, and consumer-oriented electronic commerce enterprises to freely accessible digital libraries, educational initiatives, virtual museums, and other public Web sites. Today's metadata activities are unprecedented because they extend beyond the traditional library environment in an effort to deal with the Web's exponential growth. This article considers metadata in today's Web environment. The article defines metadata, examines the relationship between metadata and cataloging, provides definitions for key metadata vocabulary terms, and explores the topic of metadata generation. Metadata is an extensive and expanding subject that is prevalent in many environments. For practical reasons, this article has elected to concentrate an the information resource domain, which is defined by electronic textual documents, graphical images, archival materials, museum artifacts, and other objects found in both digital and physical information centers (e.g., libraries, museums, record centers, and archives). To show the extent and larger application of metadata, several examples are also drawn from the data warehouse, electronic commerce, open source, and medical communities.
  2. Neumann, M.; Steinberg, J.; Schaer, P.: Web-ccraping for non-programmers : introducing OXPath for digital library metadata harvesting (2017) 0.03
    0.02607361 = product of:
      0.12167685 = sum of:
        0.035931468 = weight(_text_:open in 3895) [ClassicSimilarity], result of:
          0.035931468 = score(doc=3895,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.24876907 = fieldWeight in 3895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3895)
        0.04354783 = weight(_text_:source in 3895) [ClassicSimilarity], result of:
          0.04354783 = score(doc=3895,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.27386856 = fieldWeight in 3895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3895)
        0.042197548 = weight(_text_:web in 3895) [ClassicSimilarity], result of:
          0.042197548 = score(doc=3895,freq=10.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.40312994 = fieldWeight in 3895, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3895)
      0.21428572 = coord(3/14)
    
    Abstract
    Building up new collections for digital libraries is a demanding task. Available data sets have to be extracted which is usually done with the help of software developers as it involves custom data handlers or conversion scripts. In cases where the desired data is only available on the data provider's website custom web scrapers are needed. This may be the case for small to medium-size publishers, research institutes or funding agencies. As data curation is a typical task that is done by people with a library and information science background, these people are usually proficient with XML technologies but are not full-stack programmers. Therefore we would like to present a web scraping tool that does not demand the digital library curators to program custom web scrapers from scratch. We present the open-source tool OXPath, an extension of XPath, that allows the user to define data to be extracted from websites in a declarative way. By taking one of our own use cases as an example, we guide you in more detail through the process of creating an OXPath wrapper for metadata harvesting. We also point out some practical things to consider when creating a web scraper (with OXPath). On top of that, we also present a syntax highlighting plugin for the popular text editor Atom that we developed to further support OXPath users and to simplify the authoring process.
  3. Han, M.-J.K.; Ream-Sotomayor, N.E.; Lampron, P.; Kudeki, D.: "Making Metadata Maker" (2016) 0.03
    0.025290156 = product of:
      0.11802073 = sum of:
        0.043117758 = weight(_text_:open in 2883) [ClassicSimilarity], result of:
          0.043117758 = score(doc=2883,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.2985229 = fieldWeight in 2883, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.046875 = fieldNorm(doc=2883)
        0.052257393 = weight(_text_:source in 2883) [ClassicSimilarity], result of:
          0.052257393 = score(doc=2883,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.32864225 = fieldWeight in 2883, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.046875 = fieldNorm(doc=2883)
        0.02264558 = weight(_text_:web in 2883) [ClassicSimilarity], result of:
          0.02264558 = score(doc=2883,freq=2.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.21634221 = fieldWeight in 2883, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=2883)
      0.21428572 = coord(3/14)
    
    Abstract
    Cataloging and metadata operations in academic libraries are focusing on original cataloging of their unique and hidden collections that have not been available to users because of a lack of metadata. However, creating MARC format metadata is an expensive process; libraries need professional catalogers with appropriate experience and knowledge or must train staff to do the work. To improve the cataloging and metadata creation workflow, the University of Illinois at Urbana- Champaign Library developed a web application, Metadata Maker, which allows anyone to create metadata in four different formats, including MARC21 for an online public access catalog, regardless of their familiarity with metadata standards or systems that utilize the metadata. Released as an open source application, Metadata Maker supports diacritics and Unicode non-Roman language encoding, and creates metadata records that ensure discovery and access of unique library collections.
  4. Hardesty, J.L.; Young, J.B.: ¬The semantics of metadata : Avalon Media System and the move to RDF (2017) 0.03
    0.025290156 = product of:
      0.11802073 = sum of:
        0.043117758 = weight(_text_:open in 3896) [ClassicSimilarity], result of:
          0.043117758 = score(doc=3896,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.2985229 = fieldWeight in 3896, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.046875 = fieldNorm(doc=3896)
        0.052257393 = weight(_text_:source in 3896) [ClassicSimilarity], result of:
          0.052257393 = score(doc=3896,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.32864225 = fieldWeight in 3896, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.046875 = fieldNorm(doc=3896)
        0.02264558 = weight(_text_:web in 3896) [ClassicSimilarity], result of:
          0.02264558 = score(doc=3896,freq=2.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.21634221 = fieldWeight in 3896, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=3896)
      0.21428572 = coord(3/14)
    
    Abstract
    The Avalon Media System (Avalon) provides access and management for digital audio and video collections in libraries and archives. The open source project is led by the libraries of Indiana University Bloomington and Northwestern University and is funded in part by grants from The Andrew W. Mellon Foundation and Institute of Museum and Library Services. Avalon is based on the Samvera Community (formerly Hydra Project) software stack and uses Fedora as the digital repository back end. The Avalon project team is in the process of migrating digital repositories from Fedora 3 to Fedora 4 and incorporating metadata statements using the Resource Description Framework (RDF) instead of XML files accompanying the digital objects in the repository. The Avalon team has worked on the migration path for technical metadata and is now working on the migration paths for structural metadata (PCDM) and descriptive metadata (from MODS XML to RDF). This paper covers the decisions made to begin using RDF for software development and offers a window into how Semantic Web technology functions in the real world.
  5. Hooland, S. van; Bontemps, Y.; Kaufman, S.: Answering the call for more accountability : applying data profiling to museum metadata (2008) 0.02
    0.023231149 = product of:
      0.10841203 = sum of:
        0.043117758 = weight(_text_:open in 2644) [ClassicSimilarity], result of:
          0.043117758 = score(doc=2644,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.2985229 = fieldWeight in 2644, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.046875 = fieldNorm(doc=2644)
        0.052257393 = weight(_text_:source in 2644) [ClassicSimilarity], result of:
          0.052257393 = score(doc=2644,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.32864225 = fieldWeight in 2644, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.046875 = fieldNorm(doc=2644)
        0.013036874 = product of:
          0.026073748 = sum of:
            0.026073748 = weight(_text_:22 in 2644) [ClassicSimilarity], result of:
              0.026073748 = score(doc=2644,freq=2.0), product of:
                0.11231873 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0320743 = queryNorm
                0.23214069 = fieldWeight in 2644, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2644)
          0.5 = coord(1/2)
      0.21428572 = coord(3/14)
    
    Abstract
    Although the issue of metadata quality is recognized as an important topic within the metadata research community, the cultural heritage sector has been slow to develop methodologies, guidelines and tools for addressing this topic in practice. This paper concentrates on metadata quality specifically within the museum sector and describes the potential of data-profiling techniques for metadata quality evaluation. A case study illustrates the application of a generalpurpose data-profiling tool on a large collection of metadata records from an ethnographic collection. After an analysis of the results of the case-study the paper reviews further steps in our research and presents the implementation of a metadata quality tool within an open-source collection management software.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  6. Metadata and semantics research : 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings (2016) 0.02
    0.019700054 = product of:
      0.091933586 = sum of:
        0.05030405 = weight(_text_:open in 3283) [ClassicSimilarity], result of:
          0.05030405 = score(doc=3283,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.3482767 = fieldWeight in 3283, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3283)
        0.026419844 = weight(_text_:web in 3283) [ClassicSimilarity], result of:
          0.026419844 = score(doc=3283,freq=2.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.25239927 = fieldWeight in 3283, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3283)
        0.015209687 = product of:
          0.030419374 = sum of:
            0.030419374 = weight(_text_:22 in 3283) [ClassicSimilarity], result of:
              0.030419374 = score(doc=3283,freq=2.0), product of:
                0.11231873 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0320743 = queryNorm
                0.2708308 = fieldWeight in 3283, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3283)
          0.5 = coord(1/2)
      0.21428572 = coord(3/14)
    
    Abstract
    This book constitutes the refereed proceedings of the 10th Metadata and Semantics Research Conference, MTSR 2016, held in Göttingen, Germany, in November 2016. The 26 full papers and 6 short papers presented were carefully reviewed and selected from 67 submissions. The papers are organized in several sessions and tracks: Digital Libraries, Information Retrieval, Linked and Social Data, Metadata and Semantics for Open Repositories, Research Information Systems and Data Infrastructures, Metadata and Semantics for Agriculture, Food and Environment, Metadata and Semantics for Cultural Collections and Applications, European and National Projects.
    Theme
    Semantic Web
  7. Panskus, E.J.: Metadaten zur Identifizierung von Falschmeldungen im digitalen Raum : eine praktische Annäherung (2019) 0.02
    0.017184598 = product of:
      0.12029219 = sum of:
        0.062801845 = weight(_text_:medien in 5452) [ClassicSimilarity], result of:
          0.062801845 = score(doc=5452,freq=2.0), product of:
            0.15096188 = queryWeight, product of:
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.0320743 = queryNorm
            0.4160113 = fieldWeight in 5452, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.0625 = fieldNorm(doc=5452)
        0.057490345 = weight(_text_:open in 5452) [ClassicSimilarity], result of:
          0.057490345 = score(doc=5452,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.39803052 = fieldWeight in 5452, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0625 = fieldNorm(doc=5452)
      0.14285715 = coord(2/14)
    
    Abstract
    In vielen Ländern erstarken populistische und rassistische Kräfte. Mit Polen und Ungarn schwächen selbst Mitglieder der Europäischen Union rechtsstaatliche Institutionen.[1] Die Türkei wendet sich immer stärker von der EU ab und driftet an den Rand einer Diktatur. In Österreich konnte ein Rechtspopulist nur knapp als Bundespräsident verhindert werden. All diese Ereignisse finden oder fanden auch wegen Missmut und Misstrauen gegenüber staatlichen und etablierten Institutionen wie klassischen Medien, Regierungen und der Wirtschaft statt.
    Source
    Open Password. 2019, Nr.666 vom 22.11.2019, [https://www.password-online.de/?wysija-page=1&controller=email&action=view&email_id=823&wysijap=subscriptions&user_id=1045]
  8. Kopácsi, S.; Hudak, R.; Ganguly, R.: Implementation of a classification server to support metadata organization for long term preservation systems (2017) 0.02
    0.015895858 = product of:
      0.11127101 = sum of:
        0.05030405 = weight(_text_:open in 3915) [ClassicSimilarity], result of:
          0.05030405 = score(doc=3915,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.3482767 = fieldWeight in 3915, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3915)
        0.060966957 = weight(_text_:source in 3915) [ClassicSimilarity], result of:
          0.060966957 = score(doc=3915,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.38341597 = fieldWeight in 3915, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3915)
      0.14285715 = coord(2/14)
    
    Abstract
    In diesem Artikel beschreiben wir die Implementierung eines Klassifikationsservers für Metadatenorganisation in einem Langzeitarchivierungssystem für digitale Objekte. Nach einer kurzen Einführung in Klassifikationen und Wissensorganisationen stellen wir die Anforderungen an das zu implementierende System vor. Wir beschreiben sämtliche Simple Knowledge Organization System (SKOS) Management Tools, die wir untersucht haben, darunter auch Skosmos, die Lösung, die wir für die Implementierung gewählt haben. Skosmos ist ein open source, webbasierter SKOS Browser, basierend auf dem Jena Fuseki SPARQL Server. Wir diskutieren einige entscheidende Schritte während der Installation der ausgewählten Tools und präsentieren sowohl die potentiell auftretenden Probleme mit den verwendeten Klassifikationen als auch mögliche Lösungen.
  9. Belém, F.M.; Almeida, J.M.; Gonçalves, M.A.: ¬A survey on tag recommendation methods : a review (2017) 0.02
    0.015746485 = product of:
      0.0734836 = sum of:
        0.035931468 = weight(_text_:open in 3524) [ClassicSimilarity], result of:
          0.035931468 = score(doc=3524,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.24876907 = fieldWeight in 3524, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3524)
        0.026688073 = weight(_text_:web in 3524) [ClassicSimilarity], result of:
          0.026688073 = score(doc=3524,freq=4.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.25496176 = fieldWeight in 3524, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3524)
        0.010864062 = product of:
          0.021728124 = sum of:
            0.021728124 = weight(_text_:22 in 3524) [ClassicSimilarity], result of:
              0.021728124 = score(doc=3524,freq=2.0), product of:
                0.11231873 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0320743 = queryNorm
                0.19345059 = fieldWeight in 3524, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3524)
          0.5 = coord(1/2)
      0.21428572 = coord(3/14)
    
    Abstract
    Tags (keywords freely assigned by users to describe web content) have become highly popular on Web 2.0 applications, because of the strong stimuli and easiness for users to create and describe their own content. This increase in tag popularity has led to a vast literature on tag recommendation methods. These methods aim at assisting users in the tagging process, possibly increasing the quality of the generated tags and, consequently, improving the quality of the information retrieval (IR) services that rely on tags as data sources. Regardless of the numerous and diversified previous studies on tag recommendation, to our knowledge, no previous work has summarized and organized them into a single survey article. In this article, we propose a taxonomy for tag recommendation methods, classifying them according to the target of the recommendations, their objectives, exploited data sources, and underlying techniques. Moreover, we provide a critical overview of these methods, pointing out their advantages and disadvantages. Finally, we describe the main open challenges related to the field, such as tag ambiguity, cold start, and evaluation issues.
    Date
    16.11.2017 13:30:22
  10. Mittler, E.: Dublin Core und deutsche Bibliotheken (2000) 0.02
    0.015675597 = product of:
      0.10972917 = sum of:
        0.06661142 = weight(_text_:medien in 4455) [ClassicSimilarity], result of:
          0.06661142 = score(doc=4455,freq=4.0), product of:
            0.15096188 = queryWeight, product of:
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.0320743 = queryNorm
            0.44124663 = fieldWeight in 4455, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.046875 = fieldNorm(doc=4455)
        0.043117758 = weight(_text_:open in 4455) [ClassicSimilarity], result of:
          0.043117758 = score(doc=4455,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.2985229 = fieldWeight in 4455, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.046875 = fieldNorm(doc=4455)
      0.14285715 = coord(2/14)
    
    Abstract
    Ende Oktober 1999 hat das 7. Dublin core Meeting in der Deutschen Bibliothek in Frankfurt/M. in beeindruckender Weise die Internationalität und Interdisziplinarität dieses Ansatzes zur Beschreibung elektronischer Medien verdeutlicht. Es war zugleich eine Anerkennung für die vielfältigen deutschen Aktivitäten. Anhand von Beispielen (Math-Net und Math-Bibl-Net; SSG-FI der SUB Göttingen; Südwestverbund und HBZ) werden die Zielvorstellungen von Dublin Core herausgearbeitet. Um die Auffindbarkeit der elektronischen Dokumente durch Standardisierung der Metadaten zu sichern setzt DC schon beim Autor an. Es ist u.a. gelungen, DC im Rahmen von neuen Standards wie open-e-book zu implementieren. Damit wird die Interoperabilität von Metadaten wesentlich verbessert. Die deutschen Bibliotheken haben sich erfolgreich in die internationale Entwicklung integriert. Auch die Teilnahme am OCLC-Projekt CORC trägt dazu bei. Das kann Rückwirkungen auch auf die Regeln für die Katalogisierung gedruckter Medien haben
  11. Al-Eryani, S.; Bucher, G.; Rühle, S: ¬Ein Metadatenmodell für gemischte Sammlungen (2018) 0.02
    0.015657939 = product of:
      0.109605566 = sum of:
        0.071862936 = weight(_text_:open in 5110) [ClassicSimilarity], result of:
          0.071862936 = score(doc=5110,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.49753815 = fieldWeight in 5110, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.078125 = fieldNorm(doc=5110)
        0.037742633 = weight(_text_:web in 5110) [ClassicSimilarity], result of:
          0.037742633 = score(doc=5110,freq=2.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.36057037 = fieldWeight in 5110, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.078125 = fieldNorm(doc=5110)
      0.14285715 = coord(2/14)
    
    Abstract
    Im Rahmen des DFG-geförderten Projekts "Entwicklung von interoperablen Standards für die Kontextualisierung heterogener Objekte am Beispiel der Provenienz Asch" wurde ein Semantic Web und Linked Open Data fähiges Metadatenmodell entwickelt, das es ermöglicht, institutionsübergreifend Kulturerbe und dessen Provenienz zu kontextualisieren.
  12. Greenberg, J.: Metadata generation : processes, people and tools (2003) 0.01
    0.014313038 = product of:
      0.100191265 = sum of:
        0.057490345 = weight(_text_:open in 1251) [ClassicSimilarity], result of:
          0.057490345 = score(doc=1251,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.39803052 = fieldWeight in 1251, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0625 = fieldNorm(doc=1251)
        0.042700917 = weight(_text_:web in 1251) [ClassicSimilarity], result of:
          0.042700917 = score(doc=1251,freq=4.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.4079388 = fieldWeight in 1251, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0625 = fieldNorm(doc=1251)
      0.14285715 = coord(2/14)
    
    Abstract
    Metadata generation is the act of creating or producing metadata. Generating good quality metadata in an efficient manner is essential for organizing and making accessible the growing number of rich resources available an the Web. The success of digital libraries, the sustenance of interoperability - as promoted by the Open Archives Initiative - and the evolution of Semantic Web all rely an efficient metadata generation. This article sketches a metadata generation framework that involves processes, people and tools. It also presents selected research initiatives and highlights the goals of the Metadata Generation Research Project.
  13. Steele, C.: Organization of knowledge on the Internet (1996) 0.01
    0.014267234 = product of:
      0.09987064 = sum of:
        0.069676526 = weight(_text_:source in 6437) [ClassicSimilarity], result of:
          0.069676526 = score(doc=6437,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.4381897 = fieldWeight in 6437, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0625 = fieldNorm(doc=6437)
        0.030194107 = weight(_text_:web in 6437) [ClassicSimilarity], result of:
          0.030194107 = score(doc=6437,freq=2.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.2884563 = fieldWeight in 6437, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0625 = fieldNorm(doc=6437)
      0.14285715 = coord(2/14)
    
    Abstract
    Discusses how information will be found in digital libraries of the future. Suggests that a metadata approach to cataloguing will prevail, whereby the content of the data will be analysed at source, rather than on receipt by an intermediary, altering the role of the cataloguer, who will be required to evaluate as well as classify and index. Web site administrators may have to provide site metadata for use by global indexers; provides some examples of projects aimed at developing standards and assisting information providers in creating descriptive metadata
  14. Li, C.; Sugimoto, S.: Provenance description of metadata application profiles for long-term maintenance of metadata schemas : Luciano Floridi's philosophy of information as the foundation for library and information science (2018) 0.01
    0.014208509 = product of:
      0.09945956 = sum of:
        0.026688073 = weight(_text_:web in 4048) [ClassicSimilarity], result of:
          0.026688073 = score(doc=4048,freq=4.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.25496176 = fieldWeight in 4048, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4048)
        0.07277149 = weight(_text_:log in 4048) [ClassicSimilarity], result of:
          0.07277149 = score(doc=4048,freq=2.0), product of:
            0.205552 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0320743 = queryNorm
            0.3540296 = fieldWeight in 4048, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4048)
      0.14285715 = coord(2/14)
    
    Abstract
    Purpose Provenance information is crucial for consistent maintenance of metadata schemas over time. The purpose of this paper is to propose a provenance model named DSP-PROV to keep track of structural changes of metadata schemas. Design/methodology/approach The DSP-PROV model is developed through applying the general provenance description standard PROV of the World Wide Web Consortium to the Dublin Core Application Profile. Metadata Application Profile of Digital Public Library of America is selected as a case study to apply the DSP-PROV model. Finally, this paper evaluates the proposed model by comparison between formal provenance description in DSP-PROV and semi-formal change log description in English. Findings Formal provenance description in the DSP-PROV model has advantages over semi-formal provenance description in English to keep metadata schemas consistent over time. Research limitations/implications The DSP-PROV model is applicable to keep track of the structural changes of metadata schema over time. Provenance description of other features of metadata schema such as vocabulary and encoding syntax are not covered. Originality/value This study proposes a simple model for provenance description of structural features of metadata schemas based on a few standards widely accepted on the Web and shows the advantage of the proposed model to conventional semi-formal provenance description.
  15. Heidorn, P.B.; Wei, Q.: Automatic metadata extraction from museum specimen labels (2008) 0.01
    0.014071467 = product of:
      0.06566685 = sum of:
        0.035931468 = weight(_text_:open in 2624) [ClassicSimilarity], result of:
          0.035931468 = score(doc=2624,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.24876907 = fieldWeight in 2624, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2624)
        0.018871317 = weight(_text_:web in 2624) [ClassicSimilarity], result of:
          0.018871317 = score(doc=2624,freq=2.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.18028519 = fieldWeight in 2624, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2624)
        0.010864062 = product of:
          0.021728124 = sum of:
            0.021728124 = weight(_text_:22 in 2624) [ClassicSimilarity], result of:
              0.021728124 = score(doc=2624,freq=2.0), product of:
                0.11231873 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0320743 = queryNorm
                0.19345059 = fieldWeight in 2624, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2624)
          0.5 = coord(1/2)
      0.21428572 = coord(3/14)
    
    Abstract
    This paper describes the information properties of museum specimen labels and machine learning tools to automatically extract Darwin Core (DwC) and other metadata from these labels processed through Optical Character Recognition (OCR). The DwC is a metadata profile describing the core set of access points for search and retrieval of natural history collections and observation databases. Using the HERBIS Learning System (HLS) we extract 74 independent elements from these labels. The automated text extraction tools are provided as a web service so that users can reference digital images of specimens and receive back an extended Darwin Core XML representation of the content of the label. This automated extraction task is made more difficult by the high variability of museum label formats, OCR errors and the open class nature of some elements. In this paper we introduce our overall system architecture, and variability robust solutions including, the application of Hidden Markov and Naïve Bayes machine learning models, data cleaning, use of field element identifiers, and specialist learning models. The techniques developed here could be adapted to any metadata extraction situation with noisy text and weakly ordered elements.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  16. Mora-Mcginity, M. et al.: MusicWeb: music discovery with open linked semantic metadata (2016) 0.01
    0.013370153 = product of:
      0.093591064 = sum of:
        0.071862936 = weight(_text_:open in 3282) [ClassicSimilarity], result of:
          0.071862936 = score(doc=3282,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.49753815 = fieldWeight in 3282, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.078125 = fieldNorm(doc=3282)
        0.021728124 = product of:
          0.04345625 = sum of:
            0.04345625 = weight(_text_:22 in 3282) [ClassicSimilarity], result of:
              0.04345625 = score(doc=3282,freq=2.0), product of:
                0.11231873 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0320743 = queryNorm
                0.38690117 = fieldWeight in 3282, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3282)
          0.5 = coord(1/2)
      0.14285715 = coord(2/14)
    
    Source
    Metadata and semantics research: 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings. Eds.: E. Garoufallou
  17. Godby, C.J.; Young, J.A.; Childress, E.: ¬A repository of metadata crosswalks (2004) 0.01
    0.01248383 = product of:
      0.0873868 = sum of:
        0.060966957 = weight(_text_:source in 1155) [ClassicSimilarity], result of:
          0.060966957 = score(doc=1155,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.38341597 = fieldWeight in 1155, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1155)
        0.026419844 = weight(_text_:web in 1155) [ClassicSimilarity], result of:
          0.026419844 = score(doc=1155,freq=2.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.25239927 = fieldWeight in 1155, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1155)
      0.14285715 = coord(2/14)
    
    Abstract
    This paper proposes a model for metadata crosswalks that associates three pieces of information: the crosswalk, the source metadata standard, and the target metadata standard, each of which may have a machine-readable encoding and human-readable description. The crosswalks are encoded as METS records that are made available to a repository for processing by search engines, OAI harvesters, and custom-designed Web services. The METS object brings together all of the information required to access and interpret crosswalks and represents a significant improvement over previously available formats. But it raises questions about how best to describe these complex objects and exposes gaps that must eventually be filled in by the digital library community.
  18. Managing metadata in web-scale discovery systems (2016) 0.01
    0.012395455 = product of:
      0.08676818 = sum of:
        0.049788103 = weight(_text_:open in 3336) [ClassicSimilarity], result of:
          0.049788103 = score(doc=3336,freq=6.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.34470457 = fieldWeight in 3336, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.03125 = fieldNorm(doc=3336)
        0.036980078 = weight(_text_:web in 3336) [ClassicSimilarity], result of:
          0.036980078 = score(doc=3336,freq=12.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.35328537 = fieldWeight in 3336, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=3336)
      0.14285715 = coord(2/14)
    
    Abstract
    This book shows you how to harness the power of linked data and web-scale discovery systems to manage and link widely varied content across your library collection. Libraries are increasingly using web-scale discovery systems to help clients find a wide assortment of library materials, including books, journal articles, special collections, archival collections, videos, music and open access collections. Depending on the library material catalogued, the discovery system might need to negotiate different metadata standards, such as AACR, RDA, RAD, FOAF, VRA Core, METS, MODS, RDF and more. In Managing Metadata in Web-Scale Discovery Systems, editor Louise Spiteri and a range of international experts show you how to: * maximize the effectiveness of web-scale discovery systems * provide a smooth and seamless discovery experience to your users * help users conduct searches that yield relevant results * manage the sheer volume of items to which you can provide access, so your users can actually find what they need * maintain shared records that reflect the needs, languages, and identities of culturally and ethnically varied communities * manage metadata both within, across, and outside, library discovery tools by converting your library metadata to linked open data that all systems can access * manage user generated metadata from external services such as Goodreads and LibraryThing * mine user generated metadata to better serve your users in areas such as collection development or readers' advisory. The book will be essential reading for cataloguers, technical services and systems librarians and library and information science students studying modules on metadata, cataloguing, systems design, data management, and digital libraries. The book will also be of interest to those managing metadata in archives, museums and other cultural heritage institutions.
    Content
    1. Introduction: the landscape of web-scale discovery - Louise Spiteri 2. Sharing metadata across discovery systems - Marshall Breeding, Angela Kroeger and Heather Moulaison Sandy 3. Managing linked open data across discovery systems - Ali Shiri and Danoosh Davoodi 4. Redefining library resources in discovery systems - Christine DeZelar-Tiedman 5. Managing volume in discovery systems - Aaron Tay 6. Managing outsourced metadata in discovery systems - Laurel Tarulli 7. Managing user-generated metadata in discovery systems - Louise Spiteri
  19. Warner, S.: E-prints and the Open Archives Initiative (2003) 0.01
    0.012335767 = product of:
      0.08635037 = sum of:
        0.07114068 = weight(_text_:open in 4772) [ClassicSimilarity], result of:
          0.07114068 = score(doc=4772,freq=4.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.49253768 = fieldWeight in 4772, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4772)
        0.015209687 = product of:
          0.030419374 = sum of:
            0.030419374 = weight(_text_:22 in 4772) [ClassicSimilarity], result of:
              0.030419374 = score(doc=4772,freq=2.0), product of:
                0.11231873 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0320743 = queryNorm
                0.2708308 = fieldWeight in 4772, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4772)
          0.5 = coord(1/2)
      0.14285715 = coord(2/14)
    
    Abstract
    The Open Archives Initiative (OAI) was created as a practical way to promote interoperability between e-print repositories. Although the scope of the OAI has been broadened, e-print repositories still represent a significant fraction of OAI data providers. This article presents a brief survey of OAI e-print repositories, and of services using metadata harvested from e-print repositories using the OAI protocol for metadata harvesting (OAI-PMH). It then discusses several situations where metadata harvesting may be used to further improve the utility of e-print archives as a component of the scholarly communication infrastructure.
    Date
    18.12.2005 13:18:22
  20. Koho, M.; Burrows, T.; Hyvönen, E.; Ikkala, E.; Page, K.; Ransom, L.; Tuominen, J.; Emery, D.; Fraas, M.; Heller, B.; Lewis, D.; Morrison, A.; Porte, G.; Thomson, E.; Velios, A.; Wijsman, H.: Harmonizing and publishing heterogeneous premodern manuscript metadata as Linked Open Data (2022) 0.01
    0.011586635 = product of:
      0.08110645 = sum of:
        0.06223513 = weight(_text_:open in 466) [ClassicSimilarity], result of:
          0.06223513 = score(doc=466,freq=6.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.43088073 = fieldWeight in 466, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=466)
        0.018871317 = weight(_text_:web in 466) [ClassicSimilarity], result of:
          0.018871317 = score(doc=466,freq=2.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.18028519 = fieldWeight in 466, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=466)
      0.14285715 = coord(2/14)
    
    Abstract
    Manuscripts are a crucial form of evidence for research into all aspects of premodern European history and culture, and there are numerous databases devoted to describing them in detail. This descriptive information, however, is typically available only in separate data silos based on incompatible data models and user interfaces. As a result, it has been difficult to study manuscripts comprehensively across these various platforms. To address this challenge, a team of manuscript scholars and computer scientists worked to create "Mapping Manuscript Migrations" (MMM), a semantic portal, and a Linked Open Data service. MMM stands as a successful proof of concept for integrating distinct manuscript datasets into a shared platform for research and discovery with the potential for future expansion. This paper will discuss the major products of the MMM project: a unified data model, a repeatable data transformation pipeline, a Linked Open Data knowledge graph, and a Semantic Web portal. It will also examine the crucial importance of an iterative process of multidisciplinary collaboration embedded throughout the project, enabling humanities researchers to shape the development of a digital platform and tools, while also enabling the same researchers to ask more sophisticated and comprehensive research questions of the aggregated data.

Years

Languages

Types

  • a 224
  • el 39
  • m 18
  • s 14
  • x 3
  • b 2
  • More… Less…

Subjects