Search (10 results, page 1 of 1)

  • × theme_ss:"Formalerschließung"
  • × type_ss:"el"
  • × year_i:[2010 TO 2020}
  1. Delsey, T.: ¬The Making of RDA (2016) 0.02
    0.02397574 = product of:
      0.059939347 = sum of:
        0.04841807 = weight(_text_:context in 2946) [ClassicSimilarity], result of:
          0.04841807 = score(doc=2946,freq=2.0), product of:
            0.17622331 = queryWeight, product of:
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.04251826 = queryNorm
            0.27475408 = fieldWeight in 2946, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.046875 = fieldNorm(doc=2946)
        0.011521274 = product of:
          0.03456382 = sum of:
            0.03456382 = weight(_text_:22 in 2946) [ClassicSimilarity], result of:
              0.03456382 = score(doc=2946,freq=2.0), product of:
                0.1488917 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04251826 = queryNorm
                0.23214069 = fieldWeight in 2946, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2946)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Abstract
    The author revisits the development of RDA from its inception in 2005 through to its initial release in 2010. The development effort is set in the context of an evolving digital environment that was transforming both the production and dissemination of information resources and the technologies used to create, store, and access data describing those resources. The author examines the interplay between strategic commitments to align RDA with new conceptual models, emerging database structures, and metadata developments in allied communities, on the one hand, and compatibility with AACR2 legacy databases on the other. Aspects of the development effort examined include the structuring of RDA as a resource description language, organizing the new standard as a working tool, and refining guidelines and instructions for recording RDA data.
    Date
    17. 5.2016 19:22:40
  2. Teal, W.: Alma enumerator : automating repetitive cataloging tasks with Python (2018) 0.02
    0.018473173 = product of:
      0.04618293 = sum of:
        0.03261943 = weight(_text_:system in 5348) [ClassicSimilarity], result of:
          0.03261943 = score(doc=5348,freq=2.0), product of:
            0.13391352 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.04251826 = queryNorm
            0.2435858 = fieldWeight in 5348, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5348)
        0.013563501 = product of:
          0.0406905 = sum of:
            0.0406905 = weight(_text_:29 in 5348) [ClassicSimilarity], result of:
              0.0406905 = score(doc=5348,freq=2.0), product of:
                0.14956595 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04251826 = queryNorm
                0.27205724 = fieldWeight in 5348, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5348)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Abstract
    In June 2016, the Warburg College library migrated to a new integrated library system, Alma. In the process, we lost the enumeration and chronology data for roughly 79,000 print serial item records. Re-entering all this data by hand seemed an unthinkable task. Fortunately, the information was recorded as free text in each item's description field. By using Python, Alma's API and much trial and error, the Wartburg College library was able to parse the serial item descriptions into enumeration and chronology data that was uploaded back into Alma. This paper discusses the design and feasibility considerations addressed in trying to solve this problem, the complications encountered during development, and the highlights and shortcomings of the collection of Python scripts that became Alma Enumerator.
    Date
    10.11.2018 16:29:37
  3. Mayo, D.; Bowers, K.: ¬The devil's shoehorn : a case study of EAD to ArchivesSpace migration at a large university (2017) 0.02
    0.017055526 = product of:
      0.042638816 = sum of:
        0.032950602 = weight(_text_:system in 3373) [ClassicSimilarity], result of:
          0.032950602 = score(doc=3373,freq=4.0), product of:
            0.13391352 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.04251826 = queryNorm
            0.24605882 = fieldWeight in 3373, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3373)
        0.009688215 = product of:
          0.029064644 = sum of:
            0.029064644 = weight(_text_:29 in 3373) [ClassicSimilarity], result of:
              0.029064644 = score(doc=3373,freq=2.0), product of:
                0.14956595 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04251826 = queryNorm
                0.19432661 = fieldWeight in 3373, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3373)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Abstract
    A band of archivists and IT professionals at Harvard took on a project to convert nearly two million descriptions of archival collection components from marked-up text into the ArchivesSpace archival metadata management system. Starting in the mid-1990s, Harvard was an alpha implementer of EAD, an SGML (later XML) text markup language for electronic inventories, indexes, and finding aids that archivists use to wend their way through the sometimes quirky filing systems that bureaucracies establish for their records or the utter chaos in which some individuals keep their personal archives. These pathfinder documents, designed to cope with messy reality, can themselves be difficult to classify. Portions of them are rigorously structured, while other parts are narrative. Early documents predate the establishment of the standard; many feature idiosyncratic encoding that had been through several machine conversions, while others were freshly encoded and fairly consistent. In this paper, we will cover the practical and technical challenges involved in preparing a large (900MiB) corpus of XML for ingest into an open-source archival information system (ArchivesSpace). This case study will give an overview of the project, discuss problem discovery and problem solving, and address the technical challenges, analysis, solutions, and decisions and provide information on the tools produced and lessons learned. The authors of this piece are Kate Bowers, Collections Services Archivist for Metadata, Systems, and Standards at the Harvard University Archive, and Dave Mayo, a Digital Library Software Engineer for Harvard's Library and Technology Services. Kate was heavily involved in both metadata analysis and later problem solving, while Dave was the sole full-time developer assigned to the migration project.
    Date
    31. 1.2017 13:29:56
  4. Buttò, S.: RDA: analyses, considerations and activities by the Central Institute for the Union Catalogue of Italian Libraries and Bibliographic Information (ICCU) (2016) 0.01
    0.011412249 = product of:
      0.057061244 = sum of:
        0.057061244 = weight(_text_:context in 2958) [ClassicSimilarity], result of:
          0.057061244 = score(doc=2958,freq=4.0), product of:
            0.17622331 = queryWeight, product of:
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.04251826 = queryNorm
            0.32380077 = fieldWeight in 2958, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2958)
      0.2 = coord(1/5)
    
    Abstract
    The report aims to analyze the applicability of the Resource Description and Access (RDA) within the Italian public libraries, and also in the archives and museums in order to contribute to the discussion at international level. The Central Institute for the Union Catalogue of Italian libraries (ICCU) manages the online catalogue of the Italian libraries and the network of bibliographic services. ICCU has the institutional task of coordinating the cataloging and the documentation activities for the Italian libraries. On March 31 st 2014, the Institute signed the Agreement with the American Library Association,Publishing ALA, for the Italian translation rights of RDA, now available and published inRDAToolkit. The Italian translation has been carried out and realized by the Technical Working Group, made up of the main national and academic libraries, cultural Institutions and bibliographic agencies. The Group started working from the need of studying the new code in its textual detail, to better understand the principles, purposes, and applicability and finally its sustainability within the national context in relation to the area of the bibliographic control. At international level, starting from the publication of the Italian version of RDA and through the research carried out by ICCU and by the national Working Groups, the purpose is a more direct comparison with the experiences of the other European countries, also within EURIG international context, for an exchange of experiences aimed at strengthening the informational content of the data cataloging, with respect to history, cultural traditions and national identities of the different countries.
  5. Bianchini, C.; Guerrini, M.: RDA: a content standard to ensure the quality of data : history of a relationship (2016) 0.01
    0.009683615 = product of:
      0.04841807 = sum of:
        0.04841807 = weight(_text_:context in 2948) [ClassicSimilarity], result of:
          0.04841807 = score(doc=2948,freq=2.0), product of:
            0.17622331 = queryWeight, product of:
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.04251826 = queryNorm
            0.27475408 = fieldWeight in 2948, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.046875 = fieldNorm(doc=2948)
      0.2 = coord(1/5)
    
    Abstract
    RDA Resource Description and Access are guidelines for description and access to resources designed for digital environment and released, in its first version, in 2010. RDA is based on FRBR and its derived models, that focus on users' needs and on resources of any kind of content, medium and carrier. The paper discusses relevance of main features of RDA for the future role of libraries in the context of semantic web and metadata creation and exchange. The paper aims to highlight many consequences deriving from RDA being a content standard, and in particular the change from record management to data management, differences among the two functions realized by RDA (to identify and to relate entities) and functions realized by other standard such as MARC21 (to archive data) and ISB (to visualize data) and show how, as all these functions are necessary for the catalog, RDA needs to be integrated by other rules and standard and that these tools allow the fulfilment of the variation principle defined by S.R. Ranganathan.
  6. Schürmann, H.: Sacherschliessung nach RDA (2015) 0.01
    0.008970084 = product of:
      0.044850416 = sum of:
        0.044850416 = weight(_text_:index in 1831) [ClassicSimilarity], result of:
          0.044850416 = score(doc=1831,freq=2.0), product of:
            0.18579477 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.04251826 = queryNorm
            0.24139762 = fieldWeight in 1831, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1831)
      0.2 = coord(1/5)
    
    Abstract
    "Fazit: Die Sacherschliessung im Bibliothekskontext richtet sich nach dem Angebot und dem Zugang, den die Bibliothek bietet. Ein Regelwerk muss so gestaltet sein, dass die Bibliothek die Erschliessungstiefe selbst bestimmen kann. Im Datenaustausch macht dann die Übernahme von Fremddaten nur unter ähnlichen Bibliotheken Sinn. Metakataloge können keine sinnvollen Facetten anbieten, hier muss ein Relevanz-Ranking genügen. Dasselbe gilt für die Discovery Tools, in denen Quellen mit verschiedenen Erschliessungssystemen unter einer Oberfläche suchbar gemacht werden. In Kombination mit den Daten der Formalerschliessung sollen hingegen auch bei den Discovery Tools Filter so gestaltet sein, dass in spezifischen Beständen, die intellektuell sachlich erschlossen sind, ein Index der Schlagwörter als Themenfacetten angezeigt und genutzt werden kann. Die RDA wird dafür den Rahmen geben müssen."
  7. Harlow, C.: Data munging tools in Preparation for RDF : Catmandu and LODRefine (2015) 0.01
    0.008069678 = product of:
      0.040348392 = sum of:
        0.040348392 = weight(_text_:context in 2277) [ClassicSimilarity], result of:
          0.040348392 = score(doc=2277,freq=2.0), product of:
            0.17622331 = queryWeight, product of:
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.04251826 = queryNorm
            0.22896172 = fieldWeight in 2277, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2277)
      0.2 = coord(1/5)
    
    Abstract
    Data munging, or the work of remediating, enhancing and transforming library datasets for new or improved uses, has become more important and staff-inclusive in many library technology discussions and projects. Many times we know how we want our data to look, as well as how we want our data to act in discovery interfaces or when exposed, but we are uncertain how to make the data we have into the data we want. This article introduces and compares two library data munging tools that can help: LODRefine (OpenRefine with the DERI RDF Extension) and Catmandu. The strengths and best practices of each tool are discussed in the context of metadata munging use cases for an institution's metadata migration workflow. There is a focus on Linked Open Data modeling and transformation applications of each tool, in particular how metadataists, catalogers, and programmers can create metadata quality reports, enhance existing data with LOD sets, and transform that data to a RDF model. Integration of these tools with other systems and projects, the use of domain specific transformation languages, and the expansion of vocabulary reconciliation services are mentioned.
  8. Eversberg, B.: Zum Thema "Migration" - Beispiel USA (2018) 0.01
    0.0065901205 = product of:
      0.032950602 = sum of:
        0.032950602 = weight(_text_:system in 4386) [ClassicSimilarity], result of:
          0.032950602 = score(doc=4386,freq=4.0), product of:
            0.13391352 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.04251826 = queryNorm
            0.24605882 = fieldWeight in 4386, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4386)
      0.2 = coord(1/5)
    
    Abstract
    Zu den Systemen KOHA und FOLIO gibt es folgende aktuelle Demos, die man mit allen Funktionen ausprobieren kann: KOHA Komplette Demo-Anwendung von Bywater Solutions: https://bywatersolutions.com/koha-demo user = bywater / password = bywater Empfohlen: Cataloguing, mit den MARC-Formularen und Direkt-Datenabruf per Z39 FOLIO (GBV: "The Next-Generation Library System") Demo: https://folio-demo.gbv.de/ user = diku_admin / password = admin Empfohlen: "Inventory" und dann Button "New" zum Katalogisieren Dann "Title Data" für neuen Datensatz. Das ist wohl aber noch in einem Beta-Zustand. Ferner: FOLIO-Präsentation Göttingen April 2018: https://www.zbw-mediatalk.eu/de/2018/05/folio-info-day-a-look-at-the-next-generation-library-system/
  9. Edmunds, J.: Zombrary apocalypse!? : RDA, LRM, and the death of cataloging (2017) 0.01
    0.0064557428 = product of:
      0.032278713 = sum of:
        0.032278713 = weight(_text_:context in 3818) [ClassicSimilarity], result of:
          0.032278713 = score(doc=3818,freq=2.0), product of:
            0.17622331 = queryWeight, product of:
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.04251826 = queryNorm
            0.18316938 = fieldWeight in 3818, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.03125 = fieldNorm(doc=3818)
      0.2 = coord(1/5)
    
    Abstract
    A brochure on RDA issued in 2010 includes the statements that "RDA goes beyond earlier cataloguing codes in that it provides guidelines on cataloguing digital resources and a stronger emphasis on helping users find, identify, select, and obtain the information they want. RDA also supports clustering of bibliographic records to show relationships between works and their creators. This important new feature makes users more aware of a work's different editions, translations, or physical formats - an exciting development." Setting aside the fact that the author(s) of these statements and I differ on the definition of exciting, their claims are, at best, dubious. There is no evidence-empirical or anecdotal-that bibliographic records created using RDA are any better than records created using AACR2 (or AACR, for that matter) in "helping users find, identify, select, and obtain the information they want." The claim is especially unfounded in the context of the current discovery ecosystem, in which users are perfectly capable of finding, identifying, selecting, and obtaining information with absolutely no assistance from libraries or the bibliographic data libraries create.
  10. Riva, P.; Boeuf, P. le; Zumer, M.: IFLA Library Reference Model : a conceptual model for bibliographic information (2017) 0.00
    0.0027127003 = product of:
      0.013563501 = sum of:
        0.013563501 = product of:
          0.0406905 = sum of:
            0.0406905 = weight(_text_:29 in 5179) [ClassicSimilarity], result of:
              0.0406905 = score(doc=5179,freq=2.0), product of:
                0.14956595 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04251826 = queryNorm
                0.27205724 = fieldWeight in 5179, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5179)
          0.33333334 = coord(1/3)
      0.2 = coord(1/5)
    
    Date
    19. 3.2019 12:26:29