Search (117 results, page 2 of 6)

  • × theme_ss:"Datenformate"
  1. Cranefield, S.: Networked knowledge representation and exchange using UML and RDF (2001) 0.00
    0.0033329446 = product of:
      0.033329446 = sum of:
        0.033329446 = weight(_text_:web in 5896) [ClassicSimilarity], result of:
          0.033329446 = score(doc=5896,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.35694647 = fieldWeight in 5896, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5896)
      0.1 = coord(1/10)
    
    Abstract
    This paper proposes the use of the Unified Modeling Language (UML) as a language for modelling ontologies for Web resources and the knowledge contained within them. To provide a mechanism for serialising and processing object diagrams representing knowledge, a pair of XSI-T stylesheets have been developed to map from XML Metadata Interchange (XMI) encodings of class diagrams to corresponding RDF schemas and to Java classes representing the concepts in the ontologies. The Java code includes methods for marshalling and unmarshalling object-oriented information between in-memory data structures and RDF serialisations of that information. This provides a convenient mechanism for Java applications to share knowledge on the Web
  2. Johnson, B.C.: XML and MARC : which is "right"? (2001) 0.00
    0.0033329446 = product of:
      0.033329446 = sum of:
        0.033329446 = weight(_text_:web in 5423) [ClassicSimilarity], result of:
          0.033329446 = score(doc=5423,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.35694647 = fieldWeight in 5423, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5423)
      0.1 = coord(1/10)
    
    Abstract
    This article explores recent discussions about appropriate mark-up conventions for library information intended for use on the World Wide Web. In particular, the question of whether the MARC 21 format will continue to be useful and whether the time is right for a full-fledged conversion effort to XML is explored. The author concludes that the MARC format will be relevant well into the future, and its use will not hamper access to bibliographic information via the web. Early XML exploratory efforts carried out at the Stanford University's Lane Medical Library are reported on. Although these efforts are a promising start, much more consultation and investigation is needed to arrive at broadly acceptable standards for XML library information encoding and retrieval.
  3. Nix, M.: ¬Die praktische Einsetzbarkeit des CIDOC CRM in Informationssystemen im Bereich des Kulturerbes (2004) 0.00
    0.0029157193 = product of:
      0.029157192 = sum of:
        0.029157192 = weight(_text_:web in 3742) [ClassicSimilarity], result of:
          0.029157192 = score(doc=3742,freq=6.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.3122631 = fieldWeight in 3742, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3742)
      0.1 = coord(1/10)
    
    Abstract
    Es steht uns eine praktisch unbegrenzte Menge an Informationen über das World Wide Web zur Verfügung. Das Problem, das daraus erwächst, ist, diese Menge zu bewältigen und an die Information zu gelangen, die im Augenblick benötigt wird. Das überwältigende Angebot zwingt sowohl professionelle Anwender als auch Laien zu suchen, ungeachtet ihrer Ansprüche an die gewünschten Informationen. Um dieses Suchen effizienter zu gestalten, gibt es einerseits die Möglichkeit, leistungsstärkere Suchmaschinen zu entwickeln. Eine andere Möglichkeit ist, Daten besser zu strukturieren, um an die darin enthaltenen Informationen zu gelangen. Hoch strukturierte Daten sind maschinell verarbeitbar, sodass ein Teil der Sucharbeit automatisiert werden kann. Das Semantic Web ist die Vision eines weiterentwickelten World Wide Web, in dem derart strukturierten Daten von so genannten Softwareagenten verarbeitet werden. Die fortschreitende inhaltliche Strukturierung von Daten wird Semantisierung genannt. Im ersten Teil der Arbeit sollen einige wichtige Methoden der inhaltlichen Strukturierung von Daten skizziert werden, um die Stellung von Ontologien innerhalb der Semantisierung zu klären. Im dritten Kapitel wird der Aufbau und die Aufgabe des CIDOC Conceptual Reference Model (CRM), einer Domain Ontologie im Bereich des Kulturerbes dargestellt. Im darauf folgenden praktischen Teil werden verschiedene Ansätze zur Verwendung des CRM diskutiert und umgesetzt. Es wird ein Vorschlag zur Implementierung des Modells in XML erarbeitet. Das ist eine Möglichkeit, die dem Datentransport dient. Außerdem wird der Entwurf einer Klassenbibliothek in Java dargelegt, auf die die Verarbeitung und Nutzung des Modells innerhalb eines Informationssystems aufbauen kann.
  4. Horah, J.L.: from cards to the Web : ¬The evolution of a library database (1998) 0.00
    0.0028568096 = product of:
      0.028568096 = sum of:
        0.028568096 = weight(_text_:web in 4842) [ClassicSimilarity], result of:
          0.028568096 = score(doc=4842,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.3059541 = fieldWeight in 4842, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4842)
      0.1 = coord(1/10)
    
    Abstract
    The Jack Brause Library at New York University (NYU) is a special library supporting the curriculum of NYU's Real Estate Institute. The Jack Brause Library (JBL) Real estate Periodical Index was established in 1990 and draws on the library's collection of over 140 real estate periodicals. Describes the conversion of the JBL Index from a 3x5 card index to an online resource. The database was originally created using Rbase for DOS but this quickly became obsolete and in 1993 was replaced with InMagic. In 1997 the JBL Index was made available on NYU's telnet catalogue, BobCat, and the Internet database catalogue, BobCatPlus. The transition of InMagic data to USMARC formatted records involved a 3-step process: data normalization; adding value; and data recording. The Index has been operational through telnet since May 1997 and installing it onto the Web became functional in Oct 1997
  5. Burd, B.: Java und XML für Dummies (2003) 0.00
    0.0028568096 = product of:
      0.028568096 = sum of:
        0.028568096 = weight(_text_:web in 1694) [ClassicSimilarity], result of:
          0.028568096 = score(doc=1694,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.3059541 = fieldWeight in 1694, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=1694)
      0.1 = coord(1/10)
    
    Abstract
    Barry Budd erklärt, wie XML in Verbindung mit Java optimal funktionieren und Daten zwischen Anwendungen transferieren kann. Das Buch ist nicht nur für Anfänger interessant. Insbesondere Javabewanderte Web-Entwickler finden hier einen geradlinigen und lockeren Einstieg in die Zusammenarbeit von Java und XML. Die ersten 30 Seiten widmen sich der Programmiertheorie, der Installation von Java und den zusätzlich nötigen XMLrelevanten Klassen. Im zweiten Teil erfährt der Leser alles über die lineare Untersuchung von XML-Dateien und die Programmierung mit SAX. Es folgt der holistische Ansatz mit DOM und dessen Java-optimierter, eleganterer Alternative JDOM. XSL, die Anzeige von XML-Daten im Web und Java API for XML-Binding (JAXB) mit der man aus XML-Dokumenten maßgeschneiderte Java-Klassen dateien erstellen kann. Teil drei zeigt Spezialwerkzeuge für Webservices Soap, UDDI, WSDL, Java API for XML Messaging (JAXM) und die Arbeit mit XML-Registraren. Zu den Techniken erfährt der Leser sinnvolle Einsatzszenarien. Der Autor verdeutlicht die Theorie mit Code-Beispielen, die in Teilen für eigene Projekte verwendbar sind.
  6. Willner, E.: Preparing data for the Web with SGML/XML (1998) 0.00
    0.0026934259 = product of:
      0.026934259 = sum of:
        0.026934259 = weight(_text_:web in 2894) [ClassicSimilarity], result of:
          0.026934259 = score(doc=2894,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.2884563 = fieldWeight in 2894, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0625 = fieldNorm(doc=2894)
      0.1 = coord(1/10)
    
  7. Xu, A.; Hess, K.; Akerman, L.: From MARC to BIBFRAME 2.0 : Crosswalks (2018) 0.00
    0.0023806747 = product of:
      0.023806747 = sum of:
        0.023806747 = weight(_text_:web in 5172) [ClassicSimilarity], result of:
          0.023806747 = score(doc=5172,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.25496176 = fieldWeight in 5172, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5172)
      0.1 = coord(1/10)
    
    Abstract
    One of the big challenges facing academic libraries today is to increase the relevance of the libraries to their user communities. If the libraries can increase the visibility of their resources on the open web, it will increase the chances of the libraries to reach to their user communities via the user's first search experience. BIBFRAME and library Linked Data will enable libraries to publish their resources in a way that the Web understands, consume Linked Data to enrich their resources relevant to the libraries' user communities, and visualize networks across collections. However, one of the important steps for transitioning to BIBFRAME and library Linked Data involves crosswalks, mapping MARC fields and subfields across data models and performing necessary data reformatting to be in compliance with the specifications of the new model, which is currently BIBFRAME 2.0. This article looks into how the Library of Congress has mapped library bibliographic data from the MARC format to the BIBFRAME 2.0 model and vocabulary published and updated since April 2016, available from http://www.loc.gov/bibframe/docs/index.html based on the recently released conversion specifications and converter, developed by the Library of Congress with input from many community members. The BIBFRAME 2.0 standard and conversion tools will enable libraries to transform bibliographic data from MARC into BIBFRAME 2.0, which introduces a Linked Data model as the improved method of bibliographic control for the future, and make bibliographic information more useful within and beyond library communities.
  8. Cantrall, D.: From MARC to Mosaic : progressing toward data interchangeability at the Oregon State Archives (1994) 0.00
    0.0023567479 = product of:
      0.023567477 = sum of:
        0.023567477 = weight(_text_:web in 8470) [ClassicSimilarity], result of:
          0.023567477 = score(doc=8470,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.25239927 = fieldWeight in 8470, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=8470)
      0.1 = coord(1/10)
    
    Abstract
    Explains the technology used by the Oregon State Archives to relaize the goal of data interchangeability given the prescribed nature of the MARC format. Describes an emergent model of learning and information delivery focusing on the example of World Wide Web, accessed most often by the software client Mosaic, which is the fastest growing segment of the Internet information highway. Also describes The Data Magician, a flexible program which allows for many combinations of input and output formats, and will read unconventional formats such as MARC communications format. Oregon State Archives, using Mosaic and The Data Magician, are consequently able to present valuable electronic information to a variety of users
  9. Oeltjen, W.: Dokumentenstrukturen manipulieren und visualisieren : über das Arbeiten mit der logischen Struktur (1998) 0.00
    0.0023567479 = product of:
      0.023567477 = sum of:
        0.023567477 = weight(_text_:web in 6616) [ClassicSimilarity], result of:
          0.023567477 = score(doc=6616,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.25239927 = fieldWeight in 6616, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=6616)
      0.1 = coord(1/10)
    
    Abstract
    Thema dieses Beitrages sind Dokumentenstrukturen und zwar aus zwei Blickrichtungen: aus der Sicht der Autoren, die ein Dokument mit Computerunterstützung erstellen und die Dokumentenstruktur manipulieren und aus der Sicht der Lesenden, die ein Dokument lesen und die Struktur des Dokumentes wahrnehmen. Bei der Dokumentenstruktur wird unterschieden zwischen der logischen Struktur und der grafischen Struktur eines Dokumentes. Diese Trennung ermöglicht das Manipulieren und Visualisieren der logischen Struktur. Welche Bedeutung das für die Autoren und für die Benutzenden des Dokumentes hat, soll in dem Beitrag u.a. am Beispiel der Auszeichnungssprache HTML, der Dokumentenbeschreibungssprache des World-Wide Web, erörtert werden
  10. Galvão, R.M.: UNIMARC format relevance : maintenance or replacement? (2018) 0.00
    0.0023567479 = product of:
      0.023567477 = sum of:
        0.023567477 = weight(_text_:web in 5163) [ClassicSimilarity], result of:
          0.023567477 = score(doc=5163,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.25239927 = fieldWeight in 5163, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5163)
      0.1 = coord(1/10)
    
    Abstract
    This article presents an empirical study focused on a qualitative analysis of the UNIMARC format. An analysis of the structural quality of the data provided by the format is evaluated to determine its current suitability for meeting the requirements and trends in data architecture for the information network and the Semantic Web. Driven by a set of quality characteristics that identify weaknesses in the data schema that cannot be bridged by simply converting data to MARC XML or RDF/XML, we conclude that the UNIMARC format is not compliant with the current metadata schema desiderata and must be replaced.
  11. Jimenez, V.O.R.: Nuevas perspectivas para la catalogacion : metadatos ver MARC (1999) 0.00
    0.002192856 = product of:
      0.02192856 = sum of:
        0.02192856 = product of:
          0.065785676 = sum of:
            0.065785676 = weight(_text_:22 in 5743) [ClassicSimilarity], result of:
              0.065785676 = score(doc=5743,freq=4.0), product of:
                0.10019246 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.028611459 = queryNorm
                0.6565931 = fieldWeight in 5743, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=5743)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Date
    30. 3.2002 19:45:22
    Source
    Revista Española de Documentaçion Cientifica. 22(1999) no.2, S.198-219
  12. Boßmeyer, C.; Henze, V.: ¬2. MAB-Expertengespräch (1995) 0.00
    0.002086211 = product of:
      0.02086211 = sum of:
        0.02086211 = product of:
          0.06258633 = sum of:
            0.06258633 = weight(_text_:29 in 2914) [ClassicSimilarity], result of:
              0.06258633 = score(doc=2914,freq=2.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.6218451 = fieldWeight in 2914, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.125 = fieldNorm(doc=2914)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Bibliotheksdienst. 29(1995) H.2, S. 322-327
  13. MARC and metadata : METS, MODS, and MARCXML: current and future implications (2004) 0.00
    0.0020674444 = product of:
      0.020674443 = sum of:
        0.020674443 = product of:
          0.062023327 = sum of:
            0.062023327 = weight(_text_:22 in 2840) [ClassicSimilarity], result of:
              0.062023327 = score(doc=2840,freq=2.0), product of:
                0.10019246 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.028611459 = queryNorm
                0.61904186 = fieldWeight in 2840, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=2840)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Library hi tech. 22(2004) no.1
  14. Beaudiquez, M.: ¬L'¬avenir des formats de communication (1996) 0.00
    0.0020200694 = product of:
      0.020200694 = sum of:
        0.020200694 = weight(_text_:web in 92) [ClassicSimilarity], result of:
          0.020200694 = score(doc=92,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.21634221 = fieldWeight in 92, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=92)
      0.1 = coord(1/10)
    
    Abstract
    Over the last decade, the gap between North and South in relation to formats has increasingly limited development of international programmes such as UBC and UAP. At present the need to review formats in the light of network developments such as the Internet is urgent. Presentations covered the Web, limitations of traditional formats eg. MARC, Internet formats eg. Z39.50 and SGML, and multimedia formats, with examples of prototypes for converting traditional systems to international standards. The Internet has defined a new field of action for librarians. It is essential that developing countries share the experience of industrialised countries, with support from governments and collaboration from international organisations, in ensuring harmonious progress to universal compatibility
  15. Martin, P.: Conventions and notations for knowledge representation and retrieval (2000) 0.00
    0.0020200694 = product of:
      0.020200694 = sum of:
        0.020200694 = weight(_text_:web in 5070) [ClassicSimilarity], result of:
          0.020200694 = score(doc=5070,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.21634221 = fieldWeight in 5070, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=5070)
      0.1 = coord(1/10)
    
    Abstract
    Much research has focused on the problem of knowledge accessibility, sharing and reuse. Specific languages (e.g. KIF, CG, RDF) and ontologies have been proposed. Common characteristics, conventions or ontological distinctions are beginning to emerge. Since knowledge providers (humans and software agents) must follow common conventions for the knowledge to be widely accessed and re-used, we propose lexical, structural, semantic and ontological conventions based on various knowledge representation projects and our own research. These are minimal conventions that can be followed by most and cover the most common knowledge representation cases. However, agreement and refinements are still required. We also show that a notation can be both readable and expressive by quickly presenting two new notations -- Formalized English (FE) and Frame-CG (FCG) - derived from the CG linear form [9] and Frame-Logics [4]. These notations support the above conventions, and are implemented in our Web-based knowledge representation and document indexation tool, WebKB¹ [7]
  16. Doerr, M.; Gradmann, S.; Hennicke, S.; Isaac, A.; Meghini, C.; Van de Sompel, H.: ¬The Europeana Data Model (EDM) (2010) 0.00
    0.0020200694 = product of:
      0.020200694 = sum of:
        0.020200694 = weight(_text_:web in 3967) [ClassicSimilarity], result of:
          0.020200694 = score(doc=3967,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.21634221 = fieldWeight in 3967, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=3967)
      0.1 = coord(1/10)
    
    Abstract
    The Europeana Data Model (EDM) is a new approach towards structuring and representing data delivered to Europeana by the various contributing cultural heritage institutions. The model aims at greater expressivity and flexibility in comparison to the current Europeana Semantic Elements (ESE), which it is destined to replace. The design principles underlying the EDM are based on the core principles and best practices of the Semantic Web and Linked Data efforts to which Europeana wants to contribute. The model itself builds upon established standards like RDF(S), OAI-ORE, SKOS, and Dublin Core. It acts as a common top-level ontology which retains original data models and information perspectives while at the same time enabling interoperability. The paper elaborates on the aforementioned aspects and the design principles which drove the development of the EDM.
  17. SKOS Core Guide (2005) 0.00
    0.0020200694 = product of:
      0.020200694 = sum of:
        0.020200694 = weight(_text_:web in 4689) [ClassicSimilarity], result of:
          0.020200694 = score(doc=4689,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.21634221 = fieldWeight in 4689, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4689)
      0.1 = coord(1/10)
    
    Abstract
    SKOS Core provides a model for expressing the basic structure and content of concept schemes such as thesauri, classification schemes, subject heading lists, taxonomies, 'folksonomies', other types of controlled vocabulary, and also concept schemes embedded in glossaries and terminologies. The SKOS Core Vocabulary is an application of the Resource Description Framework (RDF), that can be used to express a concept scheme as an RDF graph. Using RDF allows data to be linked to and/or merged with other data, enabling data sources to be distributed across the web, but still be meaningfully composed and integrated. This document is a guide using the SKOS Core Vocabulary, for readers who already have a basic understanding of RDF concepts. This edition of the SKOS Core Guide [SKOS Core Guide] is a W3C Public Working Draft. It is the authoritative guide to recommended usage of the SKOS Core Vocabulary at the time of publication.
  18. Salgáné, M.M.: Our electronic era and bibliographic informations computer-related bibliographic data formats, metadata formats and BDML (2005) 0.00
    0.0019045398 = product of:
      0.019045398 = sum of:
        0.019045398 = weight(_text_:web in 3005) [ClassicSimilarity], result of:
          0.019045398 = score(doc=3005,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.2039694 = fieldWeight in 3005, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=3005)
      0.1 = coord(1/10)
    
    Abstract
    Using new communication technologies libraries must face continuously new questions, possibilities and expectations. This study discusses library-related aspects of our electronic era and how computer-related data formats affect bibliographic dataprocessing to give a summary of the most important results. First bibliographic formats for the exchange of bibliographic and related information in the machine-readable form between different types of computer systems were created more than 30 years ago. The evolution of information technologies leads to the improvement of computer systems. In addition to the development of computers and media types Internet has a great influence on data structure as well. Since the introduction of MARC bibliographic format, technology of data exchange between computers and between different computer systems has reached a very sophisticated stage and has contributed to the creation of new standards in this field. Today libraries work with this new infrastructure that induces many challenges. One of the most significant challenges is moving from a relatively homogenous bibliographic environment to a diverse one. Despite these challenges such changes are achievable and necessary to exploit possibilities of new metadata and technologies like the Internet and XML (Extensible Markup Language). XML is an open standard, a universal language for data on the Web. XML is nearly six-years-old standard designed for the description and computer-based management of (semi)-structured data and structured texts. XML gives developers the power to deliver structured data from a wide variety of applications and it is also an ideal format from server-to-server transfer of structured data. XML also isn't limited for Internet use and is an especially valuable tool in the field of library. In fact, XML's main strength - organizing information - makes it perfect for exchanging data between different systems. Tools that work with the XML can be used to process XML records without incurring additional costs associated with one's own software development. In addition, XML is also a suitable format for library web services. The Department of Computer-related Graphic Design and Library and Information Sciences of Debrecen University launched the BDML (Bibliographic Description Markup Language) development project in order to standardize bibliogrphic description with the help of XML.
  19. IFLA Cataloguing Principles : steps towards an International Cataloguing Code. Report from the 1st Meeting of Experts on an International Cataloguing Code, Frankfurt 2003 (2004) 0.00
    0.0018682657 = product of:
      0.009341328 = sum of:
        0.0067335647 = weight(_text_:web in 2312) [ClassicSimilarity], result of:
          0.0067335647 = score(doc=2312,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.07211407 = fieldWeight in 2312, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.015625 = fieldNorm(doc=2312)
        0.0026077637 = product of:
          0.007823291 = sum of:
            0.007823291 = weight(_text_:29 in 2312) [ClassicSimilarity], result of:
              0.007823291 = score(doc=2312,freq=2.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.07773064 = fieldWeight in 2312, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.015625 = fieldNorm(doc=2312)
          0.33333334 = coord(1/3)
      0.2 = coord(2/10)
    
    Date
    30. 7.2004 14:20:29
    Footnote
    Rez. in: KO 31(2004) no.4, S.255-257: (P. Riva): "Cataloguing standardization at the international level can be viewed as proceeding in a series of milestone conferences. This meeting, the first in a series which will cover different regions of the world, will take its place in that progression. The first IFLA Meeting of Experts an an International Cataloguing Code (IME ICC), held July 28-30, 2003 at Die Deutsche Bibliothek in Frankfurt, gathered representatives of almost all European countries as well as three of the four AACR author countries. As explained in the introduction by Barbara Tillett, chair of the IME ICC planning committee, the plan is for five meetings in total. Subsequent meetings are to take place in Buenos Aires, Argentina (held August 17-18, 2004) for Latin America and the Carribean, to be followed by Alexandria, Egypt (2005) for the Middle East, Seoul, South Korea (2006) for Asia, and Durban, South Africa (2007) for Africa. The impetus for planning these meetings was triggered by the 40th anniversary of the Paris Principles, approved at the International Conference an Cataloguing Principles held in 1961. Many will welcome the timely publication of the reports and papers from this important conference in book form. The original conference website (details given an p. 176) which includes most of the same material, is still extant, but the reports and papers gathered into this volume will be referred to by cataloguing rule makers long after the web as we know it has transformed itself into a new (and quite possibly not backwards compatible) environment.
  20. Devadason, F.J.: Common format for machine-readable bibliographic records for India : a proposal (1978) 0.00
    0.0018254347 = product of:
      0.018254347 = sum of:
        0.018254347 = product of:
          0.054763038 = sum of:
            0.054763038 = weight(_text_:29 in 5539) [ClassicSimilarity], result of:
              0.054763038 = score(doc=5539,freq=2.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.5441145 = fieldWeight in 5539, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=5539)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Source
    Towards a common bibliographic exchange format? International Symposium on Bibliographic Exchange Formats, Taormina, Sicily, 27-29 April 1978

Authors

Years

Languages

  • e 70
  • d 31
  • f 9
  • pl 1
  • sp 1
  • More… Less…

Types

  • a 99
  • s 9
  • m 8
  • el 7
  • b 2
  • l 1
  • n 1
  • x 1
  • More… Less…