Diese Datenbank enthält über 40.000 Dokumente zu Themen aus den Bereichen Formalerschließung – Inhaltserschließung – Information Retrieval.
© 2015 W. Gödert, TH Köln, Institut für Informationswissenschaft / Powered by litecat, BIS Oldenburg (Stand: 04. Juni 2021)
1Gore, E. ; Bitta, M.D. ; Cohen, D.: ¬The Digital Public Library of America and the National Digital Platform.
In: D-Lib magazine. 23(2017) nos.5/6, xx S.
Abstract: The Digital Public Library of America brings together the riches of America's libraries, archives, and museums, and makes them freely available to the world. In order to do this, DPLA has had to build elements of the national digital platform to connect to those institutions and to serve their digitized materials to audiences. In this article, we detail the construction of two critical elements of our work: the decentralized national network of "hubs," which operate in states across the country; and a version of the Hydra repository software that is tailored to the needs of our community. This technology and the organizations that make use of it serve as the foundation of the future of DPLA and other projects that seek to take advantage of the national digital platform.
Anmerkung: Vgl.: https://doi.org/10.1045/may2017-gore. Vgl. auch: http://webdoc.sub.gwdg.de/edoc/aw/d-lib/dlib/may17/gore/05gore.html.
Themenfeld: Information Gateway ; Internet
Objekt: Digital Public Library of America
Anwendungsfeld: Archive ; Museen ; Öffentliche Bibliotheken
2Cohen, D.J.: From Babel to knowledge : data mining large digital collections.
In: D-Lib magazine. 12(2006) no.3, x S.
Abstract: In Jorge Luis Borges's curious short story The Library of Babel, the narrator describes an endless collection of books stored from floor to ceiling in a labyrinth of countless hexagonal rooms. The pages of the library's books seem to contain random sequences of letters and spaces; occasionally a few intelligible words emerge in the sea of paper and ink. Nevertheless, readers diligently, and exasperatingly, scan the shelves for coherent passages. The narrator himself has wandered numerous rooms in search of enlightenment, but with resignation he simply awaits his death and burial - which Borges explains (with signature dark humor) consists of being tossed unceremoniously over the library's banister. Borges's nightmare, of course, is a cursed vision of the research methods of disciplines such as literature, history, and philosophy, where the careful reading of books, one after the other, is supposed to lead inexorably to knowledge and understanding. Computer scientists would approach Borges's library far differently. Employing the information theory that forms the basis for search engines and other computerized techniques for assessing in one fell swoop large masses of documents, they would quickly realize the collection's incoherence though sampling and statistical methods - and wisely start looking for the library's exit. These computational methods, which allow us to find patterns, determine relationships, categorize documents, and extract information from massive corpuses, will form the basis for new tools for research in the humanities and other disciplines in the coming decade. For the past three years I have been experimenting with how to provide such end-user tools - that is, tools that harness the power of vast electronic collections while hiding much of their complicated technical plumbing. In particular, I have made extensive use of the application programming interfaces (APIs) the leading search engines provide for programmers to query their databases directly (from server to server without using their web interfaces). In addition, I have explored how one might extract information from large digital collections, from the well-curated lexicographic database WordNet to the democratic (and poorly curated) online reference work Wikipedia. While processing these digital corpuses is currently an imperfect science, even now useful tools can be created by combining various collections and methods for searching and analyzing them. And more importantly, these nascent services suggest a future in which information can be gleaned from, and sense can be made out of, even imperfect digital libraries of enormous scale. A brief examination of two approaches to data mining large digital collections hints at this future, while also providing some lessons about how to get there.
Inhalt: Vgl. Syllabus Finder unter: http://chnm.gmu.edu/tools/syllabi/
Anmerkung: Vgl.: http://dlib.ukoln.ac.uk/dlib/march06/cohen/03cohen.html.
Themenfeld: Data Mining
Objekt: Syllabus Finder
3Smith-Cohen, D.: Developing an inhouse database from online sources.
In: Special libraries. 84(1993) no.1, S.9-17.
Abstract: Describes the creation of a database of citations in a narrow subject area from references downloaded from commercial databases, with customised added indexing, for access by end users. The database was on the subject of biological environmental and physical aspects of arctic wetlands and was compiled by the US Army Corps of Engineers, Cold Regions Research and Engineering Laboratory. Databases were selected from the DIALOG and ORBIT hosts and downloaded records were loaded into a database using Pro-Cite software and searched using Inmagic search software
Themenfeld: Bibliographische Software