Search (29 results, page 1 of 2)

  • × type_ss:"el"
  • × theme_ss:"Formalerschließung"
  1. Mayo, D.; Bowers, K.: ¬The devil's shoehorn : a case study of EAD to ArchivesSpace migration at a large university (2017) 0.00
    0.0035600248 = product of:
      0.024920173 = sum of:
        0.019013375 = weight(_text_:system in 3373) [ClassicSimilarity], result of:
          0.019013375 = score(doc=3373,freq=4.0), product of:
            0.07727166 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02453417 = queryNorm
            0.24605882 = fieldWeight in 3373, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3373)
        0.005906798 = weight(_text_:information in 3373) [ClassicSimilarity], result of:
          0.005906798 = score(doc=3373,freq=4.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.13714671 = fieldWeight in 3373, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3373)
      0.14285715 = coord(2/14)
    
    Abstract
    A band of archivists and IT professionals at Harvard took on a project to convert nearly two million descriptions of archival collection components from marked-up text into the ArchivesSpace archival metadata management system. Starting in the mid-1990s, Harvard was an alpha implementer of EAD, an SGML (later XML) text markup language for electronic inventories, indexes, and finding aids that archivists use to wend their way through the sometimes quirky filing systems that bureaucracies establish for their records or the utter chaos in which some individuals keep their personal archives. These pathfinder documents, designed to cope with messy reality, can themselves be difficult to classify. Portions of them are rigorously structured, while other parts are narrative. Early documents predate the establishment of the standard; many feature idiosyncratic encoding that had been through several machine conversions, while others were freshly encoded and fairly consistent. In this paper, we will cover the practical and technical challenges involved in preparing a large (900MiB) corpus of XML for ingest into an open-source archival information system (ArchivesSpace). This case study will give an overview of the project, discuss problem discovery and problem solving, and address the technical challenges, analysis, solutions, and decisions and provide information on the tools produced and lessons learned. The authors of this piece are Kate Bowers, Collections Services Archivist for Metadata, Systems, and Standards at the Harvard University Archive, and Dave Mayo, a Digital Library Software Engineer for Harvard's Library and Technology Services. Kate was heavily involved in both metadata analysis and later problem solving, while Dave was the sole full-time developer assigned to the migration project.
  2. Teal, W.: Alma enumerator : automating repetitive cataloging tasks with Python (2018) 0.00
    0.0035242445 = product of:
      0.02466971 = sum of:
        0.018822279 = weight(_text_:system in 5348) [ClassicSimilarity], result of:
          0.018822279 = score(doc=5348,freq=2.0), product of:
            0.07727166 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02453417 = queryNorm
            0.2435858 = fieldWeight in 5348, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5348)
        0.0058474317 = weight(_text_:information in 5348) [ClassicSimilarity], result of:
          0.0058474317 = score(doc=5348,freq=2.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.13576832 = fieldWeight in 5348, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5348)
      0.14285715 = coord(2/14)
    
    Abstract
    In June 2016, the Warburg College library migrated to a new integrated library system, Alma. In the process, we lost the enumeration and chronology data for roughly 79,000 print serial item records. Re-entering all this data by hand seemed an unthinkable task. Fortunately, the information was recorded as free text in each item's description field. By using Python, Alma's API and much trial and error, the Wartburg College library was able to parse the serial item descriptions into enumeration and chronology data that was uploaded back into Alma. This paper discusses the design and feasibility considerations addressed in trying to solve this problem, the complications encountered during development, and the highlights and shortcomings of the collection of Python scripts that became Alma Enumerator.
  3. Eversberg, B.: Zum Thema "Migration" - Beispiel USA (2018) 0.00
    0.0033128732 = product of:
      0.02319011 = sum of:
        0.019013375 = weight(_text_:system in 4386) [ClassicSimilarity], result of:
          0.019013375 = score(doc=4386,freq=4.0), product of:
            0.07727166 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02453417 = queryNorm
            0.24605882 = fieldWeight in 4386, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4386)
        0.004176737 = weight(_text_:information in 4386) [ClassicSimilarity], result of:
          0.004176737 = score(doc=4386,freq=2.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.09697737 = fieldWeight in 4386, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4386)
      0.14285715 = coord(2/14)
    
    Abstract
    Wir leiden an einer bedauerlichen Knappheit von Information über die Marktlage und Erfahrungen bei den bibliothekarischen DV-Systemen. Das ist jenseits des "Teiches" anders: Jedes Jahr veröffentlicht der Konsultant Marshall Breeding eine umfangreiche Marktübersicht und hält auf seiner Website eine Menge andere Berichte und interessante Übersichten bereit, z.B. "Migration Reports": https://librarytechnology.org/Library Technology Guides. Da sieht man in Tabellen, z.B. auch für Koha, wieviele Anwender von welchen Systemen insgesamt schon zu Koha gewandert sind, die Gesamtzahl ist 3.547! Zu Alma dagegen nur 1151, zu WMS 464, zu Aleph 1036. ("allegro-C" ist nicht dabei, aber es gab ja auch nie irgendwelche Anwender in USA, außer Goethe-Institute, aber die sind wohl nicht erfaßt.) Breedings neueste Marktübersicht für 2018 ist im Journal "American Libraries" veröffentlicht: https://americanlibrariesmagazine.org/2018/05/01/library-systems-report-2018/
    Zu den Systemen KOHA und FOLIO gibt es folgende aktuelle Demos, die man mit allen Funktionen ausprobieren kann: KOHA Komplette Demo-Anwendung von Bywater Solutions: https://bywatersolutions.com/koha-demo user = bywater / password = bywater Empfohlen: Cataloguing, mit den MARC-Formularen und Direkt-Datenabruf per Z39 FOLIO (GBV: "The Next-Generation Library System") Demo: https://folio-demo.gbv.de/ user = diku_admin / password = admin Empfohlen: "Inventory" und dann Button "New" zum Katalogisieren Dann "Title Data" für neuen Datensatz. Das ist wohl aber noch in einem Beta-Zustand. Ferner: FOLIO-Präsentation Göttingen April 2018: https://www.zbw-mediatalk.eu/de/2018/05/folio-info-day-a-look-at-the-next-generation-library-system/
  4. Gatenby, J.; Thornburg, G.; Weitz, J.: Collected work clustering in WorldCat : three techniques for maintaining records (2015) 0.00
    0.0028419765 = product of:
      0.019893834 = sum of:
        0.0050120843 = weight(_text_:information in 2276) [ClassicSimilarity], result of:
          0.0050120843 = score(doc=2276,freq=2.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.116372846 = fieldWeight in 2276, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2276)
        0.014881751 = weight(_text_:retrieval in 2276) [ClassicSimilarity], result of:
          0.014881751 = score(doc=2276,freq=2.0), product of:
            0.07421378 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.02453417 = queryNorm
            0.20052543 = fieldWeight in 2276, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=2276)
      0.14285715 = coord(2/14)
    
    Abstract
    WorldCat records are clustered into works, and within works, into content and manifestation clusters. A recent project revisited the clustering of collected works that had been previously sidelined because of the challenges posed by their complexity. Attention was given to both the identification of collected works and to the determination of the component works within them. By extensively analysing cast-list information, performance notes, contents notes, titles, uniform titles and added entries, the contents of collected works could be identified and differentiated so that correct clustering was achieved. Further work is envisaged in the form of refining the tests and weights and also in the creation and use of name/title authority records and other knowledge cards in clustering. There is a requirement to link collected works with their component works for use in search and retrieval.
  5. Mimno, D.; Crane, G.; Jones, A.: Hierarchical catalog records : implementing a FRBR catalog (2005) 0.00
    0.0024911955 = product of:
      0.017438367 = sum of:
        0.010755588 = weight(_text_:system in 1183) [ClassicSimilarity], result of:
          0.010755588 = score(doc=1183,freq=2.0), product of:
            0.07727166 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02453417 = queryNorm
            0.13919188 = fieldWeight in 1183, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=1183)
        0.006682779 = weight(_text_:information in 1183) [ClassicSimilarity], result of:
          0.006682779 = score(doc=1183,freq=8.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.1551638 = fieldWeight in 1183, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=1183)
      0.14285715 = coord(2/14)
    
    Abstract
    IFLA's Functional Requirements for Bibliographic Records (FRBR) lay the foundation for a new generation of cataloging systems that recognize the difference between a particular work (e.g., Moby Dick), diverse expressions of that work (e.g., translations into German, Japanese and other languages), different versions of the same basic text (e.g., the Modern Library Classics vs. Penguin editions), and particular items (a copy of Moby Dick on the shelf). Much work has gone into finding ways to infer FRBR relationships between existing catalog records and modifying catalog interfaces to display those relationships. Relatively little work, however, has gone into exploring the creation of catalog records that are inherently based on the FRBR hierarchy of works, expressions, manifestations, and items. The Perseus Digital Library has created a new catalog that implements such a system for a small collection that includes many works with multiple versions. We have used this catalog to explore some of the implications of hierarchical catalog records for searching and browsing. Current online library catalog interfaces present many problems for searching. One commonly cited failure is the inability to find and collocate all versions of a distinct intellectual work that exist in a collection and the inability to take into account known variations in titles and personal names (Yee 2005). The IFLA Functional Requirements for Bibliographic Records (FRBR) attempts to address some of these failings by introducing the concept of multiple interrelated bibliographic entities (IFLA 1998). In particular, relationships between abstract intellectual works and the various published instances of those works are divided into a four-level hierarchy of works (such as the Aeneid), expressions (Robert Fitzgerald's translation of the Aeneid), manifestations (a particular paperback edition of Robert Fitzgerald's translation of the Aeneid), and items (my copy of a particular paperback edition of Robert Fitzgerald's translation of the Aeneid). In this formulation, each level in the hierarchy "inherits" information from the preceding level. Much of the work on FRBRized catalogs so far has focused on organizing existing records that describe individual physical books. Relatively little work has gone into rethinking what information should be in catalog records, or how the records should relate to each other. It is clear, however, that a more "native" FRBR catalog would include separate records for works, expressions, manifestations, and items. In this way, all information about a work would be centralized in one record. Records for subsequent expressions of that work would add only the information specific to each expression: Samuel Butler's translation of the Iliad does not need to repeat the fact that the work was written by Homer. This approach has certain inherent advantages for collections with many versions of the same works: new publications can be cataloged more quickly, and records can be stored and updated more efficiently.
  6. Delsey, T.: ¬The Making of RDA (2016) 0.00
    0.0021406014 = product of:
      0.014984209 = sum of:
        0.0050120843 = weight(_text_:information in 2946) [ClassicSimilarity], result of:
          0.0050120843 = score(doc=2946,freq=2.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.116372846 = fieldWeight in 2946, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2946)
        0.009972124 = product of:
          0.019944249 = sum of:
            0.019944249 = weight(_text_:22 in 2946) [ClassicSimilarity], result of:
              0.019944249 = score(doc=2946,freq=2.0), product of:
                0.085914485 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02453417 = queryNorm
                0.23214069 = fieldWeight in 2946, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2946)
          0.5 = coord(1/2)
      0.14285715 = coord(2/14)
    
    Abstract
    The author revisits the development of RDA from its inception in 2005 through to its initial release in 2010. The development effort is set in the context of an evolving digital environment that was transforming both the production and dissemination of information resources and the technologies used to create, store, and access data describing those resources. The author examines the interplay between strategic commitments to align RDA with new conceptual models, emerging database structures, and metadata developments in allied communities, on the one hand, and compatibility with AACR2 legacy databases on the other. Aspects of the development effort examined include the structuring of RDA as a resource description language, organizing the new standard as a working tool, and refining guidelines and instructions for recording RDA data.
    Date
    17. 5.2016 19:22:40
  7. Report on the future of bibliographic control : draft for public comment (2007) 0.00
    0.0020293165 = product of:
      0.0142052155 = sum of:
        0.008066691 = weight(_text_:system in 1271) [ClassicSimilarity], result of:
          0.008066691 = score(doc=1271,freq=2.0), product of:
            0.07727166 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02453417 = queryNorm
            0.104393914 = fieldWeight in 1271, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1271)
        0.0061385245 = weight(_text_:information in 1271) [ClassicSimilarity], result of:
          0.0061385245 = score(doc=1271,freq=12.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.14252704 = fieldWeight in 1271, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1271)
      0.14285715 = coord(2/14)
    
    Abstract
    The future of bibliographic control will be collaborative, decentralized, international in scope, and Web-based. Its realization will occur in cooperation with the private sector, and with the active collaboration of library users. Data will be gathered from multiple sources; change will happen quickly; and bibliographic control will be dynamic, not static. The underlying technology that makes this future possible and necessary-the World Wide Web-is now almost two decades old. Libraries must continue the transition to this future without delay in order to retain their relevance as information providers. The Working Group on the Future of Bibliographic Control encourages the library community to take a thoughtful and coordinated approach to effecting significant changes in bibliographic control. Such an approach will call for leadership that is neither unitary nor centralized. Nor will the responsibility to provide such leadership fall solely to the Library of Congress (LC). That said, the Working Group recognizes that LC plays a unique role in the library community of the United States, and the directions that LC takes have great impact on all libraries. We also recognize that there are many other institutions and organizations that have the expertise and the capacity to play significant roles in the bibliographic future. Wherever possible, those institutions must step forward and take responsibility for assisting with navigating the transition and for playing appropriate ongoing roles after that transition is complete. To achieve the goals set out in this document, we must look beyond individual libraries to a system wide deployment of resources. We must realize efficiencies in order to be able to reallocate resources from certain lower-value components of the bibliographic control ecosystem into other higher-value components of that same ecosystem. The recommendations in this report are directed at a number of parties, indicated either by their common initialism (e.g., "LC" for Library of Congress, "PCC" for Program for Cooperative Cataloging) or by their general category (e.g., "Publishers," "National Libraries"). When the recommendation is addressed to "All," it is intended for the library community as a whole and its close collaborators.
    The Library of Congress must begin by prioritizing the recommendations that are directed in whole or in part at LC. Some define tasks that can be achieved immediately and with moderate effort; others will require analysis and planning that will have to be coordinated broadly and carefully. The Working Group has consciously not associated time frames with any of its recommendations. The recommendations fall into five general areas: 1. Increase the efficiency of bibliographic production for all libraries through increased cooperation and increased sharing of bibliographic records, and by maximizing the use of data produced throughout the entire "supply chain" for information resources. 2. Transfer effort into higher-value activity. In particular, expand the possibilities for knowledge creation by "exposing" rare and unique materials held by libraries that are currently hidden from view and, thus, underused. 3. Position our technology for the future by recognizing that the World Wide Web is both our technology platform and the appropriate platform for the delivery of our standards. Recognize that people are not the only users of the data we produce in the name of bibliographic control, but so too are machine applications that interact with those data in a variety of ways. 4. Position our community for the future by facilitating the incorporation of evaluative and other user-supplied information into our resource descriptions. Work to realize the potential of the FRBR framework for revealing and capitalizing on the various relationships that exist among information resources. 5. Strengthen the library profession through education and the development of metrics that will inform decision-making now and in the future. The Working Group intends what follows to serve as a broad blueprint for the Library of Congress and its colleagues in the library and information technology communities for extending and promoting access to information resources.
  8. Gonzalez, L.: What is FRBR? (2005) 0.00
    0.0017751834 = product of:
      0.012426283 = sum of:
        0.010755588 = weight(_text_:system in 3401) [ClassicSimilarity], result of:
          0.010755588 = score(doc=3401,freq=8.0), product of:
            0.07727166 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02453417 = queryNorm
            0.13919188 = fieldWeight in 3401, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.015625 = fieldNorm(doc=3401)
        0.0016706948 = weight(_text_:information in 3401) [ClassicSimilarity], result of:
          0.0016706948 = score(doc=3401,freq=2.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.03879095 = fieldWeight in 3401, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.015625 = fieldNorm(doc=3401)
      0.14285715 = coord(2/14)
    
    Content
    What are these two Beowulf translations "expressions" of? I used the term work above, an even more abstract concept in the FRBR model. In this case, the "work" is Beowulf , that ancient intellectual creation or effort that over time has been expressed in multiple ways, each manifested in several different ways itself, with one or more items in each manifestation. This is a pretty gross oversimplification of FRBR, which also details other relationships: among these entities; between these entities and various persons (such as creators, publishers, and owners); and between these entities and their subjects. It also specifies characteristics, or "attributes," of the different types of entities (such as title, physical media, date, availability, and more.). But it should be enough to grasp the possibilities. Now apply it Imagine that you have a patron who needs a copy of Heaney's translation of Beowulf . She doesn't care who published it or when, only that it's Heaney's translation. What if you (or your patron) could place an interlibrary loan call on that expression, instead of looking through multiple bibliographic records (as of March, OCLC's WorldCat had nine regular print editions) for multiple manifestations and then judging which record is the best bet on which to place a request? Combine that with functionality that lets you specify "not Braille, not large print," and it could save you time. Now imagine a patron in want of a copy, any copy, in English, of Romeo and Juliet. Saving staff time means saving money. Whether or not this actually happens depends upon what the library community decides to do with FRBR. It is not a set of cataloging rules or a system design, but it can influence both. Several library system vendors are working with FRBR ideas; VTLS's current integrated library system product Virtua incorporates FRBR concepts in its design. More vendors may follow. How the Joint Steering Committee for Revision of Anglo-American Cataloging Rules develops the Anglo-American Cataloging Rules (AACR) to incorporate FRBR will necessarily be a strong determinant of how records work in a "FRBR-ized" bibliographic database.
    National FRBR experiments The larger the bibliographic database, the greater the effect of "FRBR-like" design in reducing the appearance of duplicate records. LC, RLG, and OCLC, all influenced by FRBR, are experimenting with the redesign of their databases. LC's Network Development and MARC Standards Office has posted at its web site the results of some of its investigations into FRBR and MARC, including possible display options for bibliographic information. The design of RLG's public catalog, RedLightGreen, has been described as "FRBR-ish" by Merrilee Proffitt, RLG's program officer. If you try a search for a prolific author or much-published title in RedLightGreen, you'll probably find that the display of search results is much different than what you would expect. OCLC Research has developed a prototype "frbrized" database for fiction, OCLC FictionFinder. Try a title search for a classic title like Romeo and Juliet and observe that OCLC includes, in the initial display of results (described as "works"), a graphic indicator (stars, ranging from one to five). These show in rough terms how many libraries own the work-Romeo and Juliet clearly gets a five. Indicators like this are something resource sharing staff can consider an "ILL quality rating." If you're intrigued by FRBR's possibilities and what they could mean to resource sharing workflow, start talking. Now is the time to connect with colleagues, your local and/or consortial system vendor, RLG, OCLC, and your professional organizations. Have input into how systems develop in the FRBR world."
  9. Marcum, D.B.: ¬The future of cataloging (2005) 0.00
    0.0011523846 = product of:
      0.016133383 = sum of:
        0.016133383 = weight(_text_:system in 1086) [ClassicSimilarity], result of:
          0.016133383 = score(doc=1086,freq=2.0), product of:
            0.07727166 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02453417 = queryNorm
            0.20878783 = fieldWeight in 1086, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=1086)
      0.071428575 = coord(1/14)
    
    Abstract
    This thought piece on the future of cataloging is long on musings and short on predictions. But that isn't to denigrate it, only to clarify it's role given the possible connotations of the title. Rather than coming up with solutions or predictions, Marcum ponders the proper role of cataloging in a Google age. Marcum cites the Google project to digitize much or all of the contents of a selected set of major research libraries as evidence that the world of cataloging is changing dramatically, and she briefly identifies ways in which the Library of Congress is responding to this new environment. But, Marcum cautions, "the future of cataloging is not something that the Library of Congress, or even the small library group with which we will meet, can or expects to resolve alone." She then poses some specific questions that should be considered, including how we can massively change our current MARC/AACR2 system without creating chaos
  10. Heuvelmann, R.: FRBR-Strukturierung von MAB-Daten, oder : Wieviel MAB passt in FRBR? (2005) 0.00
    9.603204E-4 = product of:
      0.013444485 = sum of:
        0.013444485 = weight(_text_:system in 466) [ClassicSimilarity], result of:
          0.013444485 = score(doc=466,freq=2.0), product of:
            0.07727166 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02453417 = queryNorm
            0.17398985 = fieldWeight in 466, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=466)
      0.071428575 = coord(1/14)
    
    Abstract
    Die Expertengruppe MAB-Ausschuss (seit 2005: Expertengruppe Datenformate) hat sich im Verlauf des Jahres 2004 mit den FRBR und ihren Bezügen zum MABFormat befasst. Es wurde eine Tabelle FRBR => MAB erstellt (veröffentlicht unter http://www.ddb.de/professionell/pdf/frbr_mab.pdf), wichtige Ergebnisse wurden im Artikel "Maschinelles Austauschformat für Bibliotheken und die Functional Requirements for Bibliographic Records : Oder: Wieviel FRBR verträgt MAB?" im "Bibliotheksdienst" 39 (2005), Heft 10 zusammengefasst. Ergänzend dazu wurde bei der Arbeitsstelle Datenformate Der Deutschen Bibliothek versucht, MAB-Daten zu "frbrisieren", d. h. einzelne MAB-Datensätze in die vier Entitäten der Gruppe 1 (work / expression / manifestation / item) zu differenzieren. Ziel war nicht, einen fertigen OPAC-Baustein für die Indexierung, Benutzerführung oder Präsentation zu erstellen. Ziel war vielmehr, anhand von konkreten, in MAB strukturierten Daten die Schichten sichtbar zu machen. Ausgewählt für diesen Zweck wurde BISMAS, das "Bibliographische Informations-System zur Maschinellen Ausgabe und Suche" des BIS Oldenburg (www.bismas.de). In BISMAS ist es mit relativ geringem Aufwand möglich, die Präsentation eines Satzes - basierend auf der intern vorliegenden Datensatzstruktur, z.B. MAB - frei zu definieren. Die Gestaltung der Indices und der Ausgabeformate erfolgt in BISMAS mit Hilfe der Programmiersprache LM. Die Ergebnisse sollen hier anhand von Beispielen dargestellt werden.
  11. Cossham, A.F.: Models of the bibliographic universe (2017) 0.00
    7.234321E-4 = product of:
      0.010128049 = sum of:
        0.010128049 = weight(_text_:information in 3817) [ClassicSimilarity], result of:
          0.010128049 = score(doc=3817,freq=6.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.23515764 = fieldWeight in 3817, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3817)
      0.071428575 = coord(1/14)
    
    Abstract
    What kinds of mental models do library catalogue users have of the bibliographic universe in an age of online and electronic information? Using phenomenography and grounded analysis, it identifies participants' understanding, experience, and conceptualisation of the bibliographic universe, and identifies their expectations when using library catalogues. It contrasts participants' mental models with existing LIS models, and explores the nature of the bibliographic universe. The bibliographic universe can be considered to be a social object that exists because it is inscribed in catalogue records, cataloguing codes, bibliographies, and other bibliographic tools. It is a socially constituted phenomenon.
    Content
    A thesis submitted for the degree of Doctor of Philosophy at Monash University in 2017 Faculty of Information Technology. Vgl.: https://figshare.com/articles/Models_of_the_bibliographic_universe/5216347.
    Imprint
    Melbourne : Monash University; Faculty of Information Technology
  12. Guerrini, M.: Cataloguing based on bibliographic axiology (2010) 0.00
    6.2008464E-4 = product of:
      0.008681185 = sum of:
        0.008681185 = weight(_text_:information in 2624) [ClassicSimilarity], result of:
          0.008681185 = score(doc=2624,freq=6.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.20156369 = fieldWeight in 2624, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2624)
      0.071428575 = coord(1/14)
    
    Abstract
    The article presents the work of Elaine Svenonius The Intellectual Foundation of Information Organization, translated in Italian and published by Le Lettere of Florence, within the series Pinakes, with the title Il fondamento intellettuale dell'organizzazione dell'informazione. The Intellectual Foundation of Information Organization defines the theoretical aspects of library science, its philosophical basics and principles, the purposes that must be kept in mind, abstracting from the technology used in a library. The book deals with information organization and bibliographic universe, in particular using the bibliographic entities defined in FRBR, at first. Then, it analyzes all the specific languages by which works and subjects are treated. This work, already acknowledged as a classic, organizes, synthesizes and make easily understood the whole complex of knowledge, practices and procedures developed in the last 150 years.
  13. FictionFinder : a FRBR-based prototype for fiction in WorldCat (o.J.) 0.00
    5.906798E-4 = product of:
      0.008269517 = sum of:
        0.008269517 = weight(_text_:information in 2432) [ClassicSimilarity], result of:
          0.008269517 = score(doc=2432,freq=4.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.1920054 = fieldWeight in 2432, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2432)
      0.071428575 = coord(1/14)
    
    Abstract
    FictionFinder is a FRBR-based prototype that provides access to over 2.9 million bibliographic records for fiction books, eBooks, and audio materials described in OCLC WorldCat. This project applies principles of the FRBR model to aggregate bibliographic information above the manifestation level. Records are clustered into works using the OCLC FRBR Work-Set Algorithm. The algorithm collects bibliographic records into groups based on author and title information from bibliographic and authority records. Author names and titles are normalized to construct a key. All records with the same key are grouped together in a work set.
  14. O'Neill, E.T.: ¬The FRBRization of Humphry Clinker : a case study in the application of IFLA's Functional Requirements for Bibliographic Records (FRBR) (2002) 0.00
    5.06297E-4 = product of:
      0.0070881573 = sum of:
        0.0070881573 = weight(_text_:information in 2433) [ClassicSimilarity], result of:
          0.0070881573 = score(doc=2433,freq=4.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.16457605 = fieldWeight in 2433, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2433)
      0.071428575 = coord(1/14)
    
    Abstract
    The goal of OCLC's FRBR projects is to examine issues associated with the conversion of a set of bibliographic records to conform to FRBR requirements (a process referred to as "FRBRization"). The goals of this FRBR project were to: - examine issues associated with creating an entity-relationship model for (i.e., "FRBRizing") a non-trivial work - better understand the relationship between the bibliographic records and the bibliographic objects they represent - determine if the information available in the bibliographic record is sufficient to reliably identify the FRBR entities - to develop a data set that could be used to evaluate FRBRization algorithms. Using an exemplary work as a case study, lead scientist Ed O'Neill sought to: - better understand the relationship between bibliographic records and the bibliographic objects they represent - determine if the information available in the bibliographic records is sufficient to reliably identify FRBR entities.
  15. Strunck, K.: ¬Die Anwendung der 'Functional Requirements for Bibliographic Records' im Katalogisierungsunterricht (1999) 0.00
    4.7734138E-4 = product of:
      0.006682779 = sum of:
        0.006682779 = weight(_text_:information in 4181) [ClassicSimilarity], result of:
          0.006682779 = score(doc=4181,freq=2.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.1551638 = fieldWeight in 4181, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=4181)
      0.071428575 = coord(1/14)
    
    Abstract
    Der im Titel genannte Bericht der gleichnamigen Arbeitsgruppe der IFLA wird als theoretische Grundlage für den Katalogisierungsunterricht verwendet. Dieser Vortrag schildert, wie wir die Ideen des Berichts im Katalogisierungsunterricht der Royal School of Library and Information Science, Kopenhagen umgesetzt haben. Unsere Erfahrungen sind befriedigend und wir werden unseren Lehrplan auf dieses Basis weiterentwickeln
  16. Danskin, A.; Gryspeerdt, K.: Changing the Rules? : RDA and cataloguing in Europe. (2014) 0.00
    4.7734138E-4 = product of:
      0.006682779 = sum of:
        0.006682779 = weight(_text_:information in 5137) [ClassicSimilarity], result of:
          0.006682779 = score(doc=5137,freq=2.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.1551638 = fieldWeight in 5137, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=5137)
      0.071428575 = coord(1/14)
    
    Abstract
    This paper provides an overview of plans to implement RDA: Resource Description & Access in Europe to replace existing cataloguing rules. It is based on survey information gathered by EURIG and CILIP CIG. It includes background on the development of RDA as a replacement for AACR2.
  17. Edmunds, J.: Zombrary apocalypse!? : RDA, LRM, and the death of cataloging (2017) 0.00
    4.7734138E-4 = product of:
      0.006682779 = sum of:
        0.006682779 = weight(_text_:information in 3818) [ClassicSimilarity], result of:
          0.006682779 = score(doc=3818,freq=8.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.1551638 = fieldWeight in 3818, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=3818)
      0.071428575 = coord(1/14)
    
    Abstract
    A brochure on RDA issued in 2010 includes the statements that "RDA goes beyond earlier cataloguing codes in that it provides guidelines on cataloguing digital resources and a stronger emphasis on helping users find, identify, select, and obtain the information they want. RDA also supports clustering of bibliographic records to show relationships between works and their creators. This important new feature makes users more aware of a work's different editions, translations, or physical formats - an exciting development." Setting aside the fact that the author(s) of these statements and I differ on the definition of exciting, their claims are, at best, dubious. There is no evidence-empirical or anecdotal-that bibliographic records created using RDA are any better than records created using AACR2 (or AACR, for that matter) in "helping users find, identify, select, and obtain the information they want." The claim is especially unfounded in the context of the current discovery ecosystem, in which users are perfectly capable of finding, identifying, selecting, and obtaining information with absolutely no assistance from libraries or the bibliographic data libraries create.
    Equally fallacious is the statement that support for the "clustering bibliographic records to show relationships between works and their creators" is an "important new feature" of RDA. AACR2 bibliographic records and the systems housing them can, did, and do show such relationships. Finally, whether users want or care to be made "more aware of a work's different editions, translations, or physical formats" is debatable. As an aim, it sounds less like what a user wants and more like what a cataloging librarian thinks a user should want. As Amanda Cossham writes in her recently issued doctoral thesis: "The explicit focus on user needs in the FRBR model, the International Cataloguing Principles, and RDA: Resource Description and Access does not align well with the ways that users use, understand, and experience library catalogues nor with the ways that they understand and experience the wider information environment. User tasks, as constituted in the FRBR model and RDA, are insufficient to meet users' needs." (p. 11, emphasis in the original)
  18. Byrd, J.; Charbonneau, G.; Charbonneau, M.; Courtney, A.; Johnson, E.; Leonard, K.; Morrison, A.; Mudge, S.; O'Bryan, A.; Opasik, S.; Riley, J.; Turchyn, S.: ¬A white paper on the future of cataloging at Indiana University (2006) 0.00
    4.2191416E-4 = product of:
      0.005906798 = sum of:
        0.005906798 = weight(_text_:information in 3225) [ClassicSimilarity], result of:
          0.005906798 = score(doc=3225,freq=4.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.13714671 = fieldWeight in 3225, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3225)
      0.071428575 = coord(1/14)
    
    Abstract
    This is a report by a group "charged to identify current trends that will have a direct impact on cataloging operations and to define possible new roles for the online catalog and cataloging staff at Indiana University." Their one general conclusion after nine months of work is that "The need for cataloging expertise within the I.U. Libraries will not be diminished in the coming years. Rather, catalogers of the future will work in the evolving environment of publishing, scholarly communication, and information technology in new expanded roles. Catalogers will need to be key players in addressing the many challenges facing the libraries and the overall management and organization of information at Indiana University." The report also identifies five strategic directions. The report is an interesting read, and taken with the explosion of related reports (e.g., Calhoun's report to the Library of Congress cited in this issue, the UC Bibliographic Services TF Report), adds yet another perspective to the kinds of changes we must foster to create better library services in a vastly changed environment.
  19. Coyle, K.; Hillmann, D.: Resource Description and Access (RDA) : cataloging rules for the 20th century (2007) 0.00
    4.2191416E-4 = product of:
      0.005906798 = sum of:
        0.005906798 = weight(_text_:information in 2525) [ClassicSimilarity], result of:
          0.005906798 = score(doc=2525,freq=4.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.13714671 = fieldWeight in 2525, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2525)
      0.071428575 = coord(1/14)
    
    Abstract
    There is evidence that many individuals and organizations in the library world do not support the work taking place to develop a next generation of the library cataloging rules. The authors describe the tensions existing between those advocating an incremental change to cataloging process and others who desire a bolder library entry into the digital era. Libraries have lost their place as primary information providers, surpassed by more agile (and in many cases wealthier) purveyors of digital information delivery services. Although libraries still manage materials that are not available elsewhere, the library's approach to user service and the user interface is not competing successfully against services like Amazon or Google. If libraries are to avoid further marginalization, they need to make a fundamental change in their approach to user services. The library's signature service, its catalog, uses rules for cataloging that are remnants of a long departed technology: the card catalog. Modifications to the rules, such as those proposed by the Resource Description and Access (RDA) development effort, can only keep us rooted firmly in the 20th, if not the 19th century. A more radical change is required that will contribute to the library of the future, re-imagined and integrated with the chosen workflow of its users.
  20. Valacchi, F.: Things in the World : the integration process of archival descriptions in intercultural systems (2016) 0.00
    4.176737E-4 = product of:
      0.0058474317 = sum of:
        0.0058474317 = weight(_text_:information in 2957) [ClassicSimilarity], result of:
          0.0058474317 = score(doc=2957,freq=2.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.13576832 = fieldWeight in 2957, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2957)
      0.071428575 = coord(1/14)
    
    Abstract
    This paper conjectures that standard archival descriptions are no longer efficient in order to answer to society needs, mainly in an intercultural perspective. After a brief evaluation of the peculiarities of cultural heritage different domain languages, the specific issues of archival descriptions are discussed, seeking the possible strategies - technological as well as cultural - valid to open to an integration of descriptive languages. A particular focus is proposed on RDA, an approach which shows to be the best candidate to harmonize the separate descriptions typical of archival domain and activating the potential informative integrations with any limitation of information environments and single content quality.