Search (32 results, page 1 of 2)

  • × year_i:[2000 TO 2010}
  • × theme_ss:"Auszeichnungssprachen"
  1. as: XML: Extensible Markup Language : I: Was ist XML? (2001) 0.03
    0.034582928 = product of:
      0.069165856 = sum of:
        0.069165856 = sum of:
          0.006765375 = weight(_text_:a in 4950) [ClassicSimilarity], result of:
            0.006765375 = score(doc=4950,freq=2.0), product of:
              0.053105544 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046056706 = queryNorm
              0.12739488 = fieldWeight in 4950, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.078125 = fieldNorm(doc=4950)
          0.06240048 = weight(_text_:22 in 4950) [ClassicSimilarity], result of:
            0.06240048 = score(doc=4950,freq=2.0), product of:
              0.16128273 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046056706 = queryNorm
              0.38690117 = fieldWeight in 4950, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.078125 = fieldNorm(doc=4950)
      0.5 = coord(1/2)
    
    Date
    30. 3.2003 11:06:22
    Type
    a
  2. Trotman, A.: Searching structured documents (2004) 0.03
    0.027640268 = product of:
      0.055280536 = sum of:
        0.055280536 = sum of:
          0.011600202 = weight(_text_:a in 2538) [ClassicSimilarity], result of:
            0.011600202 = score(doc=2538,freq=12.0), product of:
              0.053105544 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046056706 = queryNorm
              0.21843673 = fieldWeight in 2538, product of:
                3.4641016 = tf(freq=12.0), with freq of:
                  12.0 = termFreq=12.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2538)
          0.043680333 = weight(_text_:22 in 2538) [ClassicSimilarity], result of:
            0.043680333 = score(doc=2538,freq=2.0), product of:
              0.16128273 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046056706 = queryNorm
              0.2708308 = fieldWeight in 2538, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2538)
      0.5 = coord(1/2)
    
    Abstract
    Structured document interchange formats such as XML and SGML are ubiquitous, however, information retrieval systems supporting structured searching are not. Structured searching can result in increased precision. A search for the author "Smith" in an unstructured corpus of documents specializing in iron-working could have a lower precision than a structured search for "Smith as author" in the same corpus. Analysis of XML retrieval languages identifies additional functionality that must be supported including searching at, and broken across multiple nodes in the document tree. A data structure is developed to support structured document searching. Application of this structure to information retrieval is then demonstrated. Document ranking is examined and adapted specifically for structured searching.
    Date
    14. 8.2004 10:39:22
    Type
    a
  3. Ioannides, D.: XML schema languages : beyond DTD (2000) 0.02
    0.022235535 = product of:
      0.04447107 = sum of:
        0.04447107 = sum of:
          0.007030784 = weight(_text_:a in 720) [ClassicSimilarity], result of:
            0.007030784 = score(doc=720,freq=6.0), product of:
              0.053105544 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046056706 = queryNorm
              0.13239266 = fieldWeight in 720, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=720)
          0.037440285 = weight(_text_:22 in 720) [ClassicSimilarity], result of:
            0.037440285 = score(doc=720,freq=2.0), product of:
              0.16128273 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046056706 = queryNorm
              0.23214069 = fieldWeight in 720, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=720)
      0.5 = coord(1/2)
    
    Abstract
    The flexibility and extensibility of XML have largely contributed to its wide acceptance beyond the traditional realm of SGML. Yet, there is still one more obstacle to be overcome before XML is able to become the evangelized universal data/document format. The obstacle is posed by the limitations of the legacy standard for constraining the contents of an XML document. The traditionally used DTD (document type definition) format does not lend itself to be used in the wide variety of applications XML is capable of handling. The World Wide Web Consortium (W3C) has charged the XML schema working group with the task of developing a schema language to replace DTD. This XML schema language is evolving based on early drafts of XML schema languages. Each one of these early efforts adopted a slightly different approach, but all of them were moving in the same direction.
    Date
    28. 1.2006 19:01:22
    Type
    a
  4. Vonhoegen, H.: Einstieg in XML (2002) 0.01
    0.0054600416 = product of:
      0.010920083 = sum of:
        0.010920083 = product of:
          0.021840166 = sum of:
            0.021840166 = weight(_text_:22 in 4002) [ClassicSimilarity], result of:
              0.021840166 = score(doc=4002,freq=2.0), product of:
                0.16128273 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046056706 = queryNorm
                0.1354154 = fieldWeight in 4002, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=4002)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Footnote
    Rez. in: XML Magazin und Web Services 2003, H.1, S.14 (S. Meyen): "Seit dem 22. Februar 1999 ist das Resource Description Framework (RDF) als W3C-Empfehlung verfügbar. Doch was steckt hinter diesem Standard, der das Zeitalter des Semantischen Webs einläuten soll? Was RDF bedeutet, wozu man es einsetzt, welche Vorteile es gegenüber XML hat und wie man RDF anwendet, soll in diesem Artikel erläutert werden. Schlägt man das Buch auf und beginnt, im EinleitungsKapitel zu schmökern, fällt sogleich ins Auge, dass der Leser nicht mit Lektionen im Stile von "bei XML sind die spitzen Klammern ganz wichtig" belehrt wird, obgleich es sich um ein Buch für Anfänger handelt. Im Gegenteil: Es geht gleich zur Sache und eine gesunde Mischung an Vorkenntnissen wird vorausgesetzt. Wer sich heute für XML interessiert, der hat ja mit 99-prozentiger Wahrscheinlichkeit schon seine einschlägigen Erfahrungen mit HTML und dem Web gemacht und ist kein Newbie in dem Reich der spitzen Klammern und der (einigermaßen) wohlformatierten Dokumente. Und hier liegt eine deutliche Stärke des Werkes Helmut Vonhoegens, der seinen Einsteiger-Leser recht gut einzuschätzen weiß und ihn daher praxisnah und verständlich ans Thema heranführt. Das dritte Kapitel beschäftigt sich mit der Document Type Definition (DTD) und beschreibt deren Einsatzziele und Verwendungsweisen. Doch betont der Autor hier unablässig die Begrenztheit dieses Ansatzes, welche den Ruf nach einem neuen Konzept deutlich macht: XML Schema, welches er im folgenden Kapitel darstellt. Ein recht ausführliches Kapitel widmet sich dann dem relativ aktuellen XML Schema-Konzept und erläutert dessen Vorzüge gegenüber der DTD (Modellierung komplexer Datenstrukturen, Unterstützung zahlreicher Datentypen, Zeichenbegrenzungen u.v.m.). XML Schema legt, so erfährt der Leser, wie die alte DTD, das Vokabular und die zulässige Grammatik eines XML-Dokuments fest, ist aber seinerseits ebenfalls ein XML-Dokument und kann (bzw. sollte) wie jedes andere XML auf Wohlgeformtheit überprüft werden. Weitere Kapitel behandeln die Navigations-Standards XPath, XLink und XPointer, Transformationen mit XSLT und XSL und natürlich die XML-Programmierschnittstellen DOM und SAX. Dabei kommen verschiedene Implementierungen zum Einsatz und erfreulicherweise werden Microsoft-Ansätze auf der einen und Java/Apache-Projekte auf der anderen Seite in ungefähr vergleichbarem Umfang vorgestellt. Im letzten Kapitel schließlich behandelt Vonhoegen die obligatorischen Web Services ("Webdienste") als Anwendungsfall von XML und demonstriert ein kleines C#- und ASP-basiertes Beispiel (das Java-Äquivalent mit Apache Axis fehlt leider). "Einstieg in XML" präsentiert seinen Stoff in klar verständlicher Form und versteht es, seine Leser auf einem guten Niveau "abzuholen". Es bietet einen guten Überblick über die Grundlagen von XML und kann - zumindest derzeit noch - mit recht hoher Aktualität aufwarten."
  5. Anderson, R.; Birbeck, M.; Kay, M.; Livingstone, S.; Loesgen, B.; Martin, D.; Mohr, S.; Ozu, N.; Peat, B.; Pinnock, J.; Stark, P.; Williams, K.: XML professionell : behandelt W3C DOM, SAX, CSS, XSLT, DTDs, XML Schemas, XLink, XPointer, XPath, E-Commerce, BizTalk, B2B, SOAP, WAP, WML (2000) 0.00
    0.0046800356 = product of:
      0.009360071 = sum of:
        0.009360071 = product of:
          0.018720143 = sum of:
            0.018720143 = weight(_text_:22 in 729) [ClassicSimilarity], result of:
              0.018720143 = score(doc=729,freq=2.0), product of:
                0.16128273 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046056706 = queryNorm
                0.116070345 = fieldWeight in 729, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=729)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 6.2005 15:12:11
  6. Lee, M.; Baillie, S.; Dell'Oro, J.: TML: a Thesaural Markpup Language (200?) 0.00
    0.0030444188 = product of:
      0.0060888375 = sum of:
        0.0060888375 = product of:
          0.012177675 = sum of:
            0.012177675 = weight(_text_:a in 1622) [ClassicSimilarity], result of:
              0.012177675 = score(doc=1622,freq=18.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.22931081 = fieldWeight in 1622, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1622)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Thesauri are used to provide controlled vocabularies for resource classification. Their use can greatly assist document discovery because thesauri man date a consistent shared terminology for describing documents. A particular thesauras classifies documents according to an information community's needs. As a result, there are many different thesaural schemas. This has led to a proliferation of schema-specific thesaural systems. In our research, we exploit schematic regularities to design a generic thesaural ontology and specfiy it as a markup language. The language provides a common representational framework in which to encode the idiosyncrasies of specific thesauri. This approach has several advantages: it offers consistent syntax and semantics in which to express thesauri; it allows general purpose thesaural applications to leverage many thesauri; and it supports a single thesaural user interface by which information communities can consistently organise, score and retrieve electronic documents.
  7. Chang, M.: ¬An electronic finding aid using extensible markup language (XML) and encoded archival description (EAD) (2000) 0.00
    0.0026473717 = product of:
      0.0052947435 = sum of:
        0.0052947435 = product of:
          0.010589487 = sum of:
            0.010589487 = weight(_text_:a in 4886) [ClassicSimilarity], result of:
              0.010589487 = score(doc=4886,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.19940455 = fieldWeight in 4886, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4886)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Increasingly, XML applications are appearing on the World Wide Web, from e-commerce to information management. In the case of libraries and archives, XML enables more flexible information management and retrieval than using MARC or a relational database management system. Describes a project to explore the use of XML and the EAD, and the development of a prototype electronic finding aid. It focuses on the technical aspects, and reviews the options available and the choices made. This is done within the setting of a small- to medium-sized archive with minimal tools and resources.
    Type
    a
  8. Peis, E.; Moya, F. de; Fernández-Molina, J.C.: Encoded archival description (EAD) conversion : a methodological proposal (2000) 0.00
    0.0023678814 = product of:
      0.0047357627 = sum of:
        0.0047357627 = product of:
          0.009471525 = sum of:
            0.009471525 = weight(_text_:a in 4899) [ClassicSimilarity], result of:
              0.009471525 = score(doc=4899,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.17835285 = fieldWeight in 4899, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4899)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The eventual adaptation of archives to new technological possibilities could begin with the creation of digital versions of archival finding aids, which would allow the international diffusion of descriptive information. The Standard Generalized Markup Language (SGML), document type definition (DTD) for archival description known as encoded archival description (EAD) is an appropriate tool for this purpose. Presents a methodological strategy that begins with an analysis of EAD and the informational object to be marked up, allowing the semiautomatic creation of a digital version.
    Type
    a
  9. Warwick, C.; Pritchard, E.: 'Hyped' text markup language : XML and the future of web markup (2000) 0.00
    0.002269176 = product of:
      0.004538352 = sum of:
        0.004538352 = product of:
          0.009076704 = sum of:
            0.009076704 = weight(_text_:a in 718) [ClassicSimilarity], result of:
              0.009076704 = score(doc=718,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.1709182 = fieldWeight in 718, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=718)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    There is a widespread perception that, in terms of web-technology, XML is going to be the 'next big thing'. Given the amount of comment that it has generated, it seems to be on its way to achieving that status. But how much of the praise should be taken at face value, and how much of the hype is credible? In the following article we examine some of the claims made about the importance of XML and consider how far the enthusiasm about it can be justified. Will XML cause a revolution that will change the way that everyone uses the Internet, whether as searchers or data creators? Or is it a tool for certain types of e-commerce and large-scale markup, which may not have a significant impact on the majority of web users?
    Type
    a
  10. Marcoux, Y.; Rizkallah, E.: Knowledge organization in the light of intertextual semantics : a natural-language analysis of controlled vocabularies (2008) 0.00
    0.002269176 = product of:
      0.004538352 = sum of:
        0.004538352 = product of:
          0.009076704 = sum of:
            0.009076704 = weight(_text_:a in 2241) [ClassicSimilarity], result of:
              0.009076704 = score(doc=2241,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.1709182 = fieldWeight in 2241, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2241)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Content
    Intertextual semantics is a semiotics-based approach to the design of communication artefacts primarily aimed at modeling XML structured documents. SKOS (Simple Knowledge Organization System) is a specification currently under development at the W3C that allows expressing various types of controlled vocabularies in XML. In this article, we show through an example how intertextual semantics could be applied to controlled vocabularies expressed in SKOS, and argue that it could facilitate the communication of meaning among the various persons who interact with a controlled vocabulary.
    Type
    a
  11. Learning XML : [creating self describing data] (2001) 0.00
    0.0022438213 = product of:
      0.0044876426 = sum of:
        0.0044876426 = product of:
          0.008975285 = sum of:
            0.008975285 = weight(_text_:a in 1744) [ClassicSimilarity], result of:
              0.008975285 = score(doc=1744,freq=22.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.16900843 = fieldWeight in 1744, product of:
                  4.690416 = tf(freq=22.0), with freq of:
                    22.0 = termFreq=22.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1744)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Although Learning XML covers XML with a broad brush, it nevertheless presents the key elements of the technology with enough detail to familiarise the reader with the crucial markup language. This guide is brief enough to tackle in a weekend. Author Erik T Ray begins with an excellent summary of XML's history as an outgrowth of SGML and HTML. He outlines very clearly the elements of markup, demystifying concepts such as attributes, entities and namespaces with numerous clear examples. To illustrate a real-world XML application, he gives the reader a look at a document written in DocBook--a publicly available XML document type for publishing technical writings--and explains the sections of the document step by step. A simplified version of DocBook is used later in the book to illustrate transformation--a powerful benefit of XML. The all-important Document Type Definition (DTD) is covered in depth, but the still-unofficial alternative--XML Schema--is only briefly addressed. The author makes liberal use of graphical illustrations, tables and code to demonstrate concepts along the way, keeping the reader engaged and on track. Ray also gets into a deep discussion of programming XML utilities with Perl. Learning XML is a highly readable introduction to XML for readers with existing knowledge of markup and Web technologies, and it meets its goals very well--to deliver a broad perspective of XML and its potential.
  12. Salminen, A.: Modeling documents in their context (2009) 0.00
    0.0020506454 = product of:
      0.004101291 = sum of:
        0.004101291 = product of:
          0.008202582 = sum of:
            0.008202582 = weight(_text_:a in 3847) [ClassicSimilarity], result of:
              0.008202582 = score(doc=3847,freq=6.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.1544581 = fieldWeight in 3847, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3847)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This entry describes notions and methods for analyzing and modeling documents in an organizational context. A model for the analysis process is provided and methods for data gathering, modeling, and user needs analysis described. The methods have been originally developed and tested during document standardization activities carried out in the Finnish Parliament and ministries. Later the methods have been adopted and adapted in other Finnish organizations in their document management development projects. The methods are intended especially for cases where the goal is to develop an Extensible Markup Language (XML)-based solution for document management. This entry emphasizes the importance of analyzing and describing documents in their organizational context.
    Type
    a
  13. Mintert, S.; Wilde, E.; Behme, H.: Auszeichnend : Editoren für Extensible Markup Language (2002) 0.00
    0.0020296127 = product of:
      0.0040592253 = sum of:
        0.0040592253 = product of:
          0.008118451 = sum of:
            0.008118451 = weight(_text_:a in 4759) [ClassicSimilarity], result of:
              0.008118451 = score(doc=4759,freq=2.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15287387 = fieldWeight in 4759, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4759)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  14. Qin, J.: Representation and organization of information in the Web space : from MARC to XML (2000) 0.00
    0.0020296127 = product of:
      0.0040592253 = sum of:
        0.0040592253 = product of:
          0.008118451 = sum of:
            0.008118451 = weight(_text_:a in 3918) [ClassicSimilarity], result of:
              0.008118451 = score(doc=3918,freq=2.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15287387 = fieldWeight in 3918, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3918)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  15. Fiander, D. J.: Applying XML to the bibliographic description (2001) 0.00
    0.0020296127 = product of:
      0.0040592253 = sum of:
        0.0040592253 = product of:
          0.008118451 = sum of:
            0.008118451 = weight(_text_:a in 5441) [ClassicSimilarity], result of:
              0.008118451 = score(doc=5441,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15287387 = fieldWeight in 5441, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5441)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Over the past few years there has been a significant amount of work in the area of cataloging internet resources, primarily using new metadata standards like the Dublin Core, but there has been little work on applying new data description formats like SGML and XML to traditional cataloging practices. What little work has been done in the area of using SGML and XML for traditional bibliographic description has primarily been based on the concept of converting MARC tagging into XML tagging. I suggest that, rather than attempting to convert existing MARC tagging into a new syntax based on SGML or XML, a more fruitful possibility is to return to the cataloging standards and describe their inherent structure, learning from how MARC has been used successfully in modern OPAC while attempting to avoid MARC's rigid field-based restrictions.
    Type
    a
  16. Clarke, K.S.: Extensible Markup Language (XML) (2009) 0.00
    0.0020296127 = product of:
      0.0040592253 = sum of:
        0.0040592253 = product of:
          0.008118451 = sum of:
            0.008118451 = weight(_text_:a in 3781) [ClassicSimilarity], result of:
              0.008118451 = score(doc=3781,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15287387 = fieldWeight in 3781, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3781)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    XML, the Extensible Markup Language is a syntax for tagging, or marking-up, textual information. It is a standard, established by the World Wide Web Consortium (W3C) that many use when sharing or working with structured information. XML isn't used by itself, but as a tool to create other data-specific markup languages. One benefit to using XML is that it enables these languages to distinguish the content that is being marked up from its presentation, allowing for greater flexibility and data reuse. The library community has embraced XML and uses it as the foundation for many of their own data-specific markup languages. Perhaps the greatest strength of XML is that it is very easy to start working with and yet, in conjunction with many other XML-related standards and technologies, can also be used to develop complex applications.
    Type
    a
  17. Salgáné, M.M.: Our electronic era and bibliographic informations computer-related bibliographic data formats, metadata formats and BDML (2005) 0.00
    0.0020296127 = product of:
      0.0040592253 = sum of:
        0.0040592253 = product of:
          0.008118451 = sum of:
            0.008118451 = weight(_text_:a in 3005) [ClassicSimilarity], result of:
              0.008118451 = score(doc=3005,freq=18.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15287387 = fieldWeight in 3005, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3005)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Using new communication technologies libraries must face continuously new questions, possibilities and expectations. This study discusses library-related aspects of our electronic era and how computer-related data formats affect bibliographic dataprocessing to give a summary of the most important results. First bibliographic formats for the exchange of bibliographic and related information in the machine-readable form between different types of computer systems were created more than 30 years ago. The evolution of information technologies leads to the improvement of computer systems. In addition to the development of computers and media types Internet has a great influence on data structure as well. Since the introduction of MARC bibliographic format, technology of data exchange between computers and between different computer systems has reached a very sophisticated stage and has contributed to the creation of new standards in this field. Today libraries work with this new infrastructure that induces many challenges. One of the most significant challenges is moving from a relatively homogenous bibliographic environment to a diverse one. Despite these challenges such changes are achievable and necessary to exploit possibilities of new metadata and technologies like the Internet and XML (Extensible Markup Language). XML is an open standard, a universal language for data on the Web. XML is nearly six-years-old standard designed for the description and computer-based management of (semi)-structured data and structured texts. XML gives developers the power to deliver structured data from a wide variety of applications and it is also an ideal format from server-to-server transfer of structured data. XML also isn't limited for Internet use and is an especially valuable tool in the field of library. In fact, XML's main strength - organizing information - makes it perfect for exchanging data between different systems. Tools that work with the XML can be used to process XML records without incurring additional costs associated with one's own software development. In addition, XML is also a suitable format for library web services. The Department of Computer-related Graphic Design and Library and Information Sciences of Debrecen University launched the BDML (Bibliographic Description Markup Language) development project in order to standardize bibliogrphic description with the help of XML.
    Type
    a
  18. Pharo, N.: ¬The effect of granularity and order in XML element retrieval (2008) 0.00
    0.001913537 = product of:
      0.003827074 = sum of:
        0.003827074 = product of:
          0.007654148 = sum of:
            0.007654148 = weight(_text_:a in 2118) [ClassicSimilarity], result of:
              0.007654148 = score(doc=2118,freq=4.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.14413087 = fieldWeight in 2118, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2118)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The article presents an analysis of the effect of granularity and order in an XML encoded collection of full text journal articles. Two-hundred and eighteen sessions of searchers performing simulated work tasks in the collection have been analysed. The results show that searchers prefer to use smaller sections of the article as their source of information. In interaction sessions during which articles are assessed, however, they are to a large degree evaluated as more important than the articles' sections and subsections.
    Type
    a
  19. Miller, D.R.; Clarke, K.S.: Putting XML to work in the library : tools for improving access and management (2004) 0.00
    0.001757696 = product of:
      0.003515392 = sum of:
        0.003515392 = product of:
          0.007030784 = sum of:
            0.007030784 = weight(_text_:a in 1438) [ClassicSimilarity], result of:
              0.007030784 = score(doc=1438,freq=6.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.13239266 = fieldWeight in 1438, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1438)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The authors, hoping to stimulate interest in XML (Extensible Markup Language) and explain its value to the library community, offer a fine introduction to the topic. The opening chapter defines XML as "a system for electronically tagging or marking up documents in order to label, organize, and categorize their content" and then goes on to describe its origins and fundamental building blocks. Subsequent chapters address related technologies, schema development, XML-based tools, and current and future library uses. The authors argue persuasively for increased XML use, emphasizing its advantages over HTML in flexibility, interoperability, extensibility, and internationalization. Information is detailed, deftly written, and supported by numerous examples. Readers without a technological bent may find the text daunting, but their perseverance will be richly rewarded. Particularly recommended for webmasters and those working in library information systems and technical services.
  20. Miller, D.R.: XML: Libraries' strategic opportunity (2001) 0.00
    0.0016913437 = product of:
      0.0033826875 = sum of:
        0.0033826875 = product of:
          0.006765375 = sum of:
            0.006765375 = weight(_text_:a in 1467) [ClassicSimilarity], result of:
              0.006765375 = score(doc=1467,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.12739488 = fieldWeight in 1467, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1467)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    XML (eXtensible Markup Language) is fast gaining favor as the universal format for data and document exchange -- in effect becoming the lingua franca of the Information Age. Currently, "library information" is at a particular disadvantage on the rapidly evolving World Wide Web. Why? Despite libraries'explorations of web catalogs, scanning projects, digital data repositories, and creation of web pages galore, there remains a digital divide. The core of libraries' data troves are stored in proprietary formats of integrated library systems (ILS) and in the complex and arcane MARC formats -- both restricted chiefly to the province of technical services and systems librarians. Even they are hard-pressed to extract and integrate this wealth of data with resources from outside this rarefied environment. Segregation of library information underlies many difficulties: producing standard bibliographic citations from MARC data, automatically creating new materials lists (including new web resources) on a particular topic, exchanging data with our vendors, and even migrating from one ILS to another. Why do we continue to hobble our potential by embracing these self-imposed limitations? Most ILSs began in libraries, which soon recognized the pitfalls of do-it-yourself solutions. Thus, we wisely anticipated the necessity for standards. However, with the advent of the web, we soon found "our" collections and a flood of new resources appearing in digital format on opposite sides of the divide. If we do not act quickly to integrate library resources with mainstream web resources, we are in grave danger of becoming marginalized