Search (37 results, page 1 of 2)

  • × theme_ss:"Auszeichnungssprachen"
  1. Clarke, K.S.: Extensible Markup Language (XML) (2009) 0.04
    0.042114194 = product of:
      0.08422839 = sum of:
        0.017839102 = weight(_text_:information in 3781) [ClassicSimilarity], result of:
          0.017839102 = score(doc=3781,freq=6.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.20156369 = fieldWeight in 3781, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3781)
        0.066389285 = weight(_text_:standards in 3781) [ClassicSimilarity], result of:
          0.066389285 = score(doc=3781,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.29545712 = fieldWeight in 3781, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.046875 = fieldNorm(doc=3781)
      0.5 = coord(2/4)
    
    Abstract
    XML, the Extensible Markup Language is a syntax for tagging, or marking-up, textual information. It is a standard, established by the World Wide Web Consortium (W3C) that many use when sharing or working with structured information. XML isn't used by itself, but as a tool to create other data-specific markup languages. One benefit to using XML is that it enables these languages to distinguish the content that is being marked up from its presentation, allowing for greater flexibility and data reuse. The library community has embraced XML and uses it as the foundation for many of their own data-specific markup languages. Perhaps the greatest strength of XML is that it is very easy to start working with and yet, in conjunction with many other XML-related standards and technologies, can also be used to develop complex applications.
    Source
    Encyclopedia of library and information sciences. 3rd ed. Ed.: M.J. Bates
  2. XML & Co : Die W3C-Spezifikationen für Dokumenten- und Datenarchitektur (2002) 0.04
    0.04176301 = product of:
      0.08352602 = sum of:
        0.006866273 = weight(_text_:information in 197) [ClassicSimilarity], result of:
          0.006866273 = score(doc=197,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.0775819 = fieldWeight in 197, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=197)
        0.07665975 = weight(_text_:standards in 197) [ClassicSimilarity], result of:
          0.07665975 = score(doc=197,freq=6.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.34116453 = fieldWeight in 197, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.03125 = fieldNorm(doc=197)
      0.5 = coord(2/4)
    
    Footnote
    Rez. in: XML Magazin und Web Services 2003, H.1, S.13 (T. Weitzel): "Die Standards der XML-Familie sind nun auch in deutscher Übersetzung und Kommentierung verfügbar. Unter der Leitung von Stefan Mintert, der 1998 bereits die XML-Spezifikation selbst übersetzte, ist zusammen mit einem Team namhafter XML-Spezialisten und mit Unterstützung des deutsch-österreichischen W3C-Büros eine Sammlung aller wichtigen Specs aus der XML-Familie bei Addison-Wesley erschienen. Genauer gesagt handelt es sich um die kommentierten und übersetzten Versionen der Standards aus der W3C Activity Domain Architecture/XML, und zwar, in chronologischer Reihenfolge, XML, Namensräume, Verknüpfen von Stylesheets mit XML-Dokumenten, XPath, XSLT, XML Schema 0-3 (Einführung, Strukturen und Datentypen), XLink, XML Base und XML Information Set. Ziel des Projekts "edition W3C" war es, den Inhalt der jeweiligen Spezifikationen besser verständlich zu machen und so auch einen Beitrag zur Verbreitung der Standards zu leisten. Daher liegen die Übersetzungen nicht nur in Buchform vor, sondern sind sie über das normative Original hinaus in unterschiedlichem Umfang durch die jeweiligen Experten kommentiert. Insbesondere XPath ist so ausführlich und liebevoll mit vielen Erläuterungen und Abbildungen versehen, dass eine analoge Behandlung etwa der Schema-Specs sicherlich den Rahmen des Buchs gesprengt hätte. Das erhoffte Feedback auf die unterschiedlichen Übersetzungen und Kommentierungen soll helfen, das geplante zweite Buch der edition W3C-Reihe zu XHTML und CSS an den Leserwünschen auszurichten. Interessant ist, dass das Buch selbst in XML erstellt wurde (XMLspec-DTD). Wie schon in "XML in der Praxis" (zusammen mit Henning Behme) hat Stefan Mintert für die edition W3C in lehrbuchmäßiger Cross-Media-Manier XML eingesetzt. So konnte beispielsweise das Stichwortverzeichnis (www.edition-w3c.de/ gesamtindex.html) direkt per Single Source aus den jeweiligen Spezifikationen generiert werden. Eine Besonderheit des 1999 initiierten Projekts liegt darin, dass es sich um die einzige durch das W3C legitimierte PrintPublikation überhaupt handelt, nachdem 1997 das WWW Journal (W3J) bei O'Reilly eingestellt wurde. Darüber hinaus ist die edition W3 C das einzige offizielle, also vom örtlichen W3C-Büro unterstützte, Übersetzungsprojekt. Bedenkt man, dass das W3C-neben Boston-Sitze in Frankreich und Japan hat, stellt sich die Frage, ob dort Übersetzungen nicht noch viel stärker nachgefragt werden könnten und somit auch zur weiteren Durchsetzung der XML-Familie beitragen könnten. Die Übersetzungen sind, allerdings ohne Kommentierungen, auch über die Website des Projekts bei www.edition-w3c.de/ verlinkt. Dort gibt es auch weitere Übersetzungen, etwa zu XHTML oder CSS L2."
  3. Bold, M.: ¬Die Zukunft des Web : Standards für das Web der Zukunft (2004) 0.04
    0.03912026 = product of:
      0.15648104 = sum of:
        0.15648104 = weight(_text_:standards in 1725) [ClassicSimilarity], result of:
          0.15648104 = score(doc=1725,freq=4.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.69639915 = fieldWeight in 1725, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.078125 = fieldNorm(doc=1725)
      0.25 = coord(1/4)
    
    Abstract
    Neue Technologien und Standards sollen die Zukunft des Web prägen. Internet Professionell erklärt, was es mit XML, XSLT, XHTML, XPath und XLink auf sich hat
  4. Miller, D.R.: XML: Libraries' strategic opportunity (2001) 0.04
    0.03509516 = product of:
      0.07019032 = sum of:
        0.014865918 = weight(_text_:information in 1467) [ClassicSimilarity], result of:
          0.014865918 = score(doc=1467,freq=6.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.16796975 = fieldWeight in 1467, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1467)
        0.0553244 = weight(_text_:standards in 1467) [ClassicSimilarity], result of:
          0.0553244 = score(doc=1467,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.24621427 = fieldWeight in 1467, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1467)
      0.5 = coord(2/4)
    
    Abstract
    XML (eXtensible Markup Language) is fast gaining favor as the universal format for data and document exchange -- in effect becoming the lingua franca of the Information Age. Currently, "library information" is at a particular disadvantage on the rapidly evolving World Wide Web. Why? Despite libraries'explorations of web catalogs, scanning projects, digital data repositories, and creation of web pages galore, there remains a digital divide. The core of libraries' data troves are stored in proprietary formats of integrated library systems (ILS) and in the complex and arcane MARC formats -- both restricted chiefly to the province of technical services and systems librarians. Even they are hard-pressed to extract and integrate this wealth of data with resources from outside this rarefied environment. Segregation of library information underlies many difficulties: producing standard bibliographic citations from MARC data, automatically creating new materials lists (including new web resources) on a particular topic, exchanging data with our vendors, and even migrating from one ILS to another. Why do we continue to hobble our potential by embracing these self-imposed limitations? Most ILSs began in libraries, which soon recognized the pitfalls of do-it-yourself solutions. Thus, we wisely anticipated the necessity for standards. However, with the advent of the web, we soon found "our" collections and a flood of new resources appearing in digital format on opposite sides of the divide. If we do not act quickly to integrate library resources with mainstream web resources, we are in grave danger of becoming marginalized
  5. Mayo, D.; Bowers, K.: ¬The devil's shoehorn : a case study of EAD to ArchivesSpace migration at a large university (2017) 0.03
    0.033731185 = product of:
      0.06746237 = sum of:
        0.01213797 = weight(_text_:information in 3373) [ClassicSimilarity], result of:
          0.01213797 = score(doc=3373,freq=4.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.13714671 = fieldWeight in 3373, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3373)
        0.0553244 = weight(_text_:standards in 3373) [ClassicSimilarity], result of:
          0.0553244 = score(doc=3373,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.24621427 = fieldWeight in 3373, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3373)
      0.5 = coord(2/4)
    
    Abstract
    A band of archivists and IT professionals at Harvard took on a project to convert nearly two million descriptions of archival collection components from marked-up text into the ArchivesSpace archival metadata management system. Starting in the mid-1990s, Harvard was an alpha implementer of EAD, an SGML (later XML) text markup language for electronic inventories, indexes, and finding aids that archivists use to wend their way through the sometimes quirky filing systems that bureaucracies establish for their records or the utter chaos in which some individuals keep their personal archives. These pathfinder documents, designed to cope with messy reality, can themselves be difficult to classify. Portions of them are rigorously structured, while other parts are narrative. Early documents predate the establishment of the standard; many feature idiosyncratic encoding that had been through several machine conversions, while others were freshly encoded and fairly consistent. In this paper, we will cover the practical and technical challenges involved in preparing a large (900MiB) corpus of XML for ingest into an open-source archival information system (ArchivesSpace). This case study will give an overview of the project, discuss problem discovery and problem solving, and address the technical challenges, analysis, solutions, and decisions and provide information on the tools produced and lessons learned. The authors of this piece are Kate Bowers, Collections Services Archivist for Metadata, Systems, and Standards at the Harvard University Archive, and Dave Mayo, a Digital Library Software Engineer for Harvard's Library and Technology Services. Kate was heavily involved in both metadata analysis and later problem solving, while Dave was the sole full-time developer assigned to the migration project.
  6. Vonhoegen, H.: Einstieg in XML (2002) 0.03
    0.033360977 = product of:
      0.06672195 = sum of:
        0.054768365 = weight(_text_:standards in 4002) [ClassicSimilarity], result of:
          0.054768365 = score(doc=4002,freq=4.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.2437397 = fieldWeight in 4002, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4002)
        0.011953591 = product of:
          0.023907183 = sum of:
            0.023907183 = weight(_text_:22 in 4002) [ClassicSimilarity], result of:
              0.023907183 = score(doc=4002,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.1354154 = fieldWeight in 4002, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=4002)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Dieses Buch richtet sich an alle, die eine kompetente Einführung in XML benötigen - praxisnah und verständlich aufbereitet. Die referenzartige Darstellung der eXtensible Markup Language XML, ihrer Dialekte und Technologien wird dabei durch viele Beispiele vertieft. »Einstieg in XML« ist kein theoretisches Buch zu verschiedenen Standards der Sprachfamilie XML. Hier bekommen Sie in konzentrierter Form genau das, was Sie zur Entwicklung eigener XML-Lösungen brauchen. Die im Buch enthaltene CD enthält alle nötigen Tools, um sofort starten zu können.
    Footnote
    Rez. in: XML Magazin und Web Services 2003, H.1, S.14 (S. Meyen): "Seit dem 22. Februar 1999 ist das Resource Description Framework (RDF) als W3C-Empfehlung verfügbar. Doch was steckt hinter diesem Standard, der das Zeitalter des Semantischen Webs einläuten soll? Was RDF bedeutet, wozu man es einsetzt, welche Vorteile es gegenüber XML hat und wie man RDF anwendet, soll in diesem Artikel erläutert werden. Schlägt man das Buch auf und beginnt, im EinleitungsKapitel zu schmökern, fällt sogleich ins Auge, dass der Leser nicht mit Lektionen im Stile von "bei XML sind die spitzen Klammern ganz wichtig" belehrt wird, obgleich es sich um ein Buch für Anfänger handelt. Im Gegenteil: Es geht gleich zur Sache und eine gesunde Mischung an Vorkenntnissen wird vorausgesetzt. Wer sich heute für XML interessiert, der hat ja mit 99-prozentiger Wahrscheinlichkeit schon seine einschlägigen Erfahrungen mit HTML und dem Web gemacht und ist kein Newbie in dem Reich der spitzen Klammern und der (einigermaßen) wohlformatierten Dokumente. Und hier liegt eine deutliche Stärke des Werkes Helmut Vonhoegens, der seinen Einsteiger-Leser recht gut einzuschätzen weiß und ihn daher praxisnah und verständlich ans Thema heranführt. Das dritte Kapitel beschäftigt sich mit der Document Type Definition (DTD) und beschreibt deren Einsatzziele und Verwendungsweisen. Doch betont der Autor hier unablässig die Begrenztheit dieses Ansatzes, welche den Ruf nach einem neuen Konzept deutlich macht: XML Schema, welches er im folgenden Kapitel darstellt. Ein recht ausführliches Kapitel widmet sich dann dem relativ aktuellen XML Schema-Konzept und erläutert dessen Vorzüge gegenüber der DTD (Modellierung komplexer Datenstrukturen, Unterstützung zahlreicher Datentypen, Zeichenbegrenzungen u.v.m.). XML Schema legt, so erfährt der Leser, wie die alte DTD, das Vokabular und die zulässige Grammatik eines XML-Dokuments fest, ist aber seinerseits ebenfalls ein XML-Dokument und kann (bzw. sollte) wie jedes andere XML auf Wohlgeformtheit überprüft werden. Weitere Kapitel behandeln die Navigations-Standards XPath, XLink und XPointer, Transformationen mit XSLT und XSL und natürlich die XML-Programmierschnittstellen DOM und SAX. Dabei kommen verschiedene Implementierungen zum Einsatz und erfreulicherweise werden Microsoft-Ansätze auf der einen und Java/Apache-Projekte auf der anderen Seite in ungefähr vergleichbarem Umfang vorgestellt. Im letzten Kapitel schließlich behandelt Vonhoegen die obligatorischen Web Services ("Webdienste") als Anwendungsfall von XML und demonstriert ein kleines C#- und ASP-basiertes Beispiel (das Java-Äquivalent mit Apache Axis fehlt leider). "Einstieg in XML" präsentiert seinen Stoff in klar verständlicher Form und versteht es, seine Leser auf einem guten Niveau "abzuholen". Es bietet einen guten Überblick über die Grundlagen von XML und kann - zumindest derzeit noch - mit recht hoher Aktualität aufwarten."
  7. Qin, J.: Representation and organization of information in the Web space : from MARC to XML (2000) 0.03
    0.03154151 = product of:
      0.06308302 = sum of:
        0.02059882 = weight(_text_:information in 3918) [ClassicSimilarity], result of:
          0.02059882 = score(doc=3918,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.23274569 = fieldWeight in 3918, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=3918)
        0.042484205 = product of:
          0.08496841 = sum of:
            0.08496841 = weight(_text_:organization in 3918) [ClassicSimilarity], result of:
              0.08496841 = score(doc=3918,freq=2.0), product of:
                0.17974974 = queryWeight, product of:
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.050415643 = queryNorm
                0.47270393 = fieldWeight in 3918, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3918)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
  8. Salgáné, M.M.: Our electronic era and bibliographic informations computer-related bibliographic data formats, metadata formats and BDML (2005) 0.03
    0.029806487 = product of:
      0.059612975 = sum of:
        0.015353453 = weight(_text_:information in 3005) [ClassicSimilarity], result of:
          0.015353453 = score(doc=3005,freq=10.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.1734784 = fieldWeight in 3005, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=3005)
        0.044259522 = weight(_text_:standards in 3005) [ClassicSimilarity], result of:
          0.044259522 = score(doc=3005,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.19697142 = fieldWeight in 3005, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.03125 = fieldNorm(doc=3005)
      0.5 = coord(2/4)
    
    Abstract
    Using new communication technologies libraries must face continuously new questions, possibilities and expectations. This study discusses library-related aspects of our electronic era and how computer-related data formats affect bibliographic dataprocessing to give a summary of the most important results. First bibliographic formats for the exchange of bibliographic and related information in the machine-readable form between different types of computer systems were created more than 30 years ago. The evolution of information technologies leads to the improvement of computer systems. In addition to the development of computers and media types Internet has a great influence on data structure as well. Since the introduction of MARC bibliographic format, technology of data exchange between computers and between different computer systems has reached a very sophisticated stage and has contributed to the creation of new standards in this field. Today libraries work with this new infrastructure that induces many challenges. One of the most significant challenges is moving from a relatively homogenous bibliographic environment to a diverse one. Despite these challenges such changes are achievable and necessary to exploit possibilities of new metadata and technologies like the Internet and XML (Extensible Markup Language). XML is an open standard, a universal language for data on the Web. XML is nearly six-years-old standard designed for the description and computer-based management of (semi)-structured data and structured texts. XML gives developers the power to deliver structured data from a wide variety of applications and it is also an ideal format from server-to-server transfer of structured data. XML also isn't limited for Internet use and is an especially valuable tool in the field of library. In fact, XML's main strength - organizing information - makes it perfect for exchanging data between different systems. Tools that work with the XML can be used to process XML records without incurring additional costs associated with one's own software development. In addition, XML is also a suitable format for library web services. The Department of Computer-related Graphic Design and Library and Information Sciences of Debrecen University launched the BDML (Bibliographic Description Markup Language) development project in order to standardize bibliogrphic description with the help of XML.
    Source
    Librarianship in the information age: Proceedings of the 13th BOBCATSSS Symposium, 31 January - 2 February 2005 in Budapest, Hungary. Eds.: Marte Langeland u.a
  9. Fiander, D. J.: Applying XML to the bibliographic description (2001) 0.02
    0.023472156 = product of:
      0.093888626 = sum of:
        0.093888626 = weight(_text_:standards in 5441) [ClassicSimilarity], result of:
          0.093888626 = score(doc=5441,freq=4.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.41783947 = fieldWeight in 5441, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.046875 = fieldNorm(doc=5441)
      0.25 = coord(1/4)
    
    Abstract
    Over the past few years there has been a significant amount of work in the area of cataloging internet resources, primarily using new metadata standards like the Dublin Core, but there has been little work on applying new data description formats like SGML and XML to traditional cataloging practices. What little work has been done in the area of using SGML and XML for traditional bibliographic description has primarily been based on the concept of converting MARC tagging into XML tagging. I suggest that, rather than attempting to convert existing MARC tagging into a new syntax based on SGML or XML, a more fruitful possibility is to return to the cataloging standards and describe their inherent structure, learning from how MARC has been used successfully in modern OPAC while attempting to avoid MARC's rigid field-based restrictions.
  10. Patrick, D.A.: XML in der Praxis : Unternehmensübergreifende Vorteile durch Enterprise Content Management (1999) 0.02
    0.022359734 = product of:
      0.04471947 = sum of:
        0.020812286 = weight(_text_:information in 1461) [ClassicSimilarity], result of:
          0.020812286 = score(doc=1461,freq=6.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.23515764 = fieldWeight in 1461, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1461)
        0.023907183 = product of:
          0.047814365 = sum of:
            0.047814365 = weight(_text_:22 in 1461) [ClassicSimilarity], result of:
              0.047814365 = score(doc=1461,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.2708308 = fieldWeight in 1461, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1461)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    In dem Maße, in dem Unternehmen ihren Erfolg in einem zunehmend von Konkurrenz geprägten Weltmarkt suchen, ist ds Content Management als Informationslösung interessant geworden. Content Management-Systeme können dabei helfen, die enormen betrieblichen Investitionen in die Information zu verringern. Wie bei jeder neuartigen Technologie bestehen auch hier noch keine klaren Vorstellungen darüber, was ContentnManagement eigentlich ausmacht. In diesem Beitrag werden die Probleme und Technologien im Zusammenhang mit dem Content Management untersucht und der aktuelle Stand in Sachen Content Management beschrieben. Content Management ist mehr als nur eine neue Technologie. Im Kern erlaubt Content Management Unternehmen, Informationen zum aufbau intensiverer Beziehungen entlang der Wertschöpfungskette aufzubauen, wobei Kunden, Vertriebspartner, Zulieferer und Hersteller verbunden werden
    Date
    30. 3.2003 10:50:22
    Source
    nfd Information - Wissenschaft und Praxis. 50(1999) H.1, S.5-12
    Theme
    Information Resources Management
  11. Trotman, A.: Searching structured documents (2004) 0.02
    0.022359734 = product of:
      0.04471947 = sum of:
        0.020812286 = weight(_text_:information in 2538) [ClassicSimilarity], result of:
          0.020812286 = score(doc=2538,freq=6.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.23515764 = fieldWeight in 2538, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2538)
        0.023907183 = product of:
          0.047814365 = sum of:
            0.047814365 = weight(_text_:22 in 2538) [ClassicSimilarity], result of:
              0.047814365 = score(doc=2538,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.2708308 = fieldWeight in 2538, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2538)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Structured document interchange formats such as XML and SGML are ubiquitous, however, information retrieval systems supporting structured searching are not. Structured searching can result in increased precision. A search for the author "Smith" in an unstructured corpus of documents specializing in iron-working could have a lower precision than a structured search for "Smith as author" in the same corpus. Analysis of XML retrieval languages identifies additional functionality that must be supported including searching at, and broken across multiple nodes in the document tree. A data structure is developed to support structured document searching. Application of this structure to information retrieval is then demonstrated. Document ranking is examined and adapted specifically for structured searching.
    Date
    14. 8.2004 10:39:22
    Source
    Information processing and management. 40(2004) no.4, S.619-632
  12. Anderson, R.; Birbeck, M.; Kay, M.; Livingstone, S.; Loesgen, B.; Martin, D.; Mohr, S.; Ozu, N.; Peat, B.; Pinnock, J.; Stark, P.; Williams, K.: XML professionell : behandelt W3C DOM, SAX, CSS, XSLT, DTDs, XML Schemas, XLink, XPointer, XPath, E-Commerce, BizTalk, B2B, SOAP, WAP, WML (2000) 0.02
    0.021720288 = product of:
      0.043440577 = sum of:
        0.033194643 = weight(_text_:standards in 729) [ClassicSimilarity], result of:
          0.033194643 = score(doc=729,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.14772856 = fieldWeight in 729, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.0234375 = fieldNorm(doc=729)
        0.010245935 = product of:
          0.02049187 = sum of:
            0.02049187 = weight(_text_:22 in 729) [ClassicSimilarity], result of:
              0.02049187 = score(doc=729,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.116070345 = fieldWeight in 729, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=729)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    In diesem Buch sollen die grundlegenden Techniken zur Erstellung, Anwendung und nicht zuletzt Darstellung von XML-Dokumenten erklärt und demonstriert werden. Die wichtigste und vornehmste Aufgabe dieses Buches ist es jedoch, die Grundlagen von XML, wie sie vom World Wide Web Consortium (W3C) festgelegt sind, darzustellen. Das W3C hat nicht nur die Entwicklung von XML initiiert und ist die zuständige Organisation für alle XML-Standards, es werden auch weiterhin XML-Spezifikationen vom W3C entwickelt. Auch wenn immer mehr Vorschläge für neue XML-basierte Techniken aus dem weiteren Umfeld der an XML Interessierten kommen, so spielt doch weiterhin das W3C die zentrale und wichtigste Rolle für die Entwicklung von XML. Der Schwerpunkt dieses Buches liegt darin, zu lernen, wie man XML als tragende Technologie in echten Alltags-Anwendungen verwendet. Wir wollen Ihnen gute Design-Techniken vorstellen und demonstrieren, wie man XML-fähige Anwendungen mit Applikationen für das WWW oder mit Datenbanksystemen verknüpft. Wir wollen die Grenzen und Möglichkeiten von XML ausloten und eine Vorausschau auf einige "nascent"-Technologien werfen. Egal ob Ihre Anforderungen sich mehr an dem Austausch von Daten orientieren oder bei der visuellen Gestaltung liegen, dieses Buch behandelt alle relevanten Techniken. jedes Kapitel enthält ein Anwendungsbeispiel. Da XML eine Plattform-neutrale Technologie ist, werden in den Beispielen eine breite Palette von Sprachen, Parsern und Servern behandelt. Jede der vorgestellten Techniken und Methoden ist auf allen Plattformen und Betriebssystemen relevant. Auf diese Weise erhalten Sie wichtige Einsichten durch diese Beispiele, auch wenn die konkrete Implementierung nicht auf dem von Ihnen bevorzugten System durchgeführt wurde.
    Date
    22. 6.2005 15:12:11
  13. Vanhoutte, E.; Branden, R. van den: Text Encoding Initiative (TEI) (2009) 0.02
    0.021027675 = product of:
      0.04205535 = sum of:
        0.013732546 = weight(_text_:information in 3889) [ClassicSimilarity], result of:
          0.013732546 = score(doc=3889,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.1551638 = fieldWeight in 3889, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=3889)
        0.028322803 = product of:
          0.056645606 = sum of:
            0.056645606 = weight(_text_:organization in 3889) [ClassicSimilarity], result of:
              0.056645606 = score(doc=3889,freq=2.0), product of:
                0.17974974 = queryWeight, product of:
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.050415643 = queryNorm
                0.31513596 = fieldWeight in 3889, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3889)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    The result of community efforts among computing humanists, the Text Encoding Initiative or TEI is the de facto standard for the encoding of texts in the humanities. This entry explains the historical context of the TEI, its fundamental principles, history, and organization.
    Source
    Encyclopedia of library and information sciences. 3rd ed. Ed.: M.J. Bates
  14. XML in libraries (2002) 0.01
    0.014903244 = product of:
      0.029806487 = sum of:
        0.0076767267 = weight(_text_:information in 3100) [ClassicSimilarity], result of:
          0.0076767267 = score(doc=3100,freq=10.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.0867392 = fieldWeight in 3100, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.015625 = fieldNorm(doc=3100)
        0.022129761 = weight(_text_:standards in 3100) [ClassicSimilarity], result of:
          0.022129761 = score(doc=3100,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.09848571 = fieldWeight in 3100, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.015625 = fieldNorm(doc=3100)
      0.5 = coord(2/4)
    
    Footnote
    Rez. in: JASIST 55(2004) no.14, S.1304-1305 (Z. Holbrooks):"The eXtensible Markup Language (XML) and its family of enabling technologies (XPath, XPointer, XLink, XSLT, et al.) were the new "new thing" only a couple of years ago. Happily, XML is now a W3C standard, and its enabling technologies are rapidly proliferating and maturing. Together, they are changing the way data is handled an the Web, how legacy data is accessed and leveraged in corporate archives, and offering the Semantic Web community a powerful toolset. Library and information professionals need a basic understanding of what XML is, and what its impacts will be an the library community as content vendors and publishers convert to the new standards. Norman Desmarais aims to provide librarians with an overview of XML and some potential library applications. The ABCs of XML contains the useful basic information that most general XML works cover. It is addressed to librarians, as evidenced by the occasional reference to periodical vendors, MARC, and OPACs. However, librarians without SGML, HTML, database, or programming experience may find the work daunting. The snippets of code-most incomplete and unattended by screenshots to illustrate the result of the code's execution-obscure more often than they enlighten. A single code sample (p. 91, a book purchase order) is immediately recognizable and sensible. There are no figures, illustrations, or screenshots. Subsection headings are used conservatively. Readers are confronted with page after page of unbroken technical text, and occasionally oddly formatted text (in some of the code samples). The author concentrates an commercial products and projects. Library and agency initiatives-for example, the National Institutes of Health HL-7 and U.S. Department of Education's GEM project-are notable for their absence. The Library of Congress USMARC to SGML effort is discussed in chapter 1, which covers the relationship of XML to its parent SGML, the XML processor, and data type definitions, using MARC as its illustrative example. Chapter 3 addresses the stylesheet options for XML, including DSSSL, CSS, and XSL. The Document Style Semantics and Specification Language (DSSSL) was created for use with SGML, and pruned into DSSSL-Lite and further (DSSSL-online). Cascading Style Sheets (CSS) were created for use with HTML. Extensible Style Language (XSL) is a further revision (and extension) of DSSSL-o specifically for use with XML. Discussion of aural stylesheets and Synchronized Multimedia Integration Language (SMIL) round out the chapter.
    Chapter 4 introduces XML internal and external pointing and linking technologies. XML Link Language (XLL, now XLink) provides unidirectional, multi-ended, and typed linking. XPointer, used with XLink, provides addressing into the interior of XML documents. XPath operates an the logical structure of an XML document, creating a tree of nodes. Used with both XPointer and XSLT, it permits operations an strings, numbers, and Boolean expressions in the document. The final chapter, "Getting Started" argues for the adoption of a tool for XML production. The features and functionality of various tools for content development, application development, databases, and schema development provide an introduction to some of the available options. Roy Tennant is weIl known in the library community as an author (bis column "Digital Libraries" has appeared in Library Journal since 1997 and he has published Current Cites each month for more than a decade), an electronic discussion list manager (Web4Lib and XML4Lib), and as the creator and manager of UC/Berkeley's Digital Library SunSITE. Librarians have wondered what use they might make of XML since its beginnings. Tennant suggests one answer: "The Extensible Markup Language (XML) has the potential to exceed the impact of MARC an librarianship. While MARC is limited to bibliographic descriptionand arguably a subset at that, as any archivist will tell you-XML provides a highly-effective framework for encoding anything from a bibliographic record for a book to the book itself." (Tennant, p. vii) This slim paperback volume offers librarians and library managers concerned with automation projects "show and teIl" examples of XML technologies used as solutions to everyday tasks and challenges. What distinguishes this work is the editor and contributors' commitment to providing messy details. This book's target audience is technically savvy. While not a "cookbook" per se, the information provided an each project serves as a draft blueprint complete with acronyms and jargon. The inclusion of "lessons learned" (including failures as well as successes) is refreshing and commendable. Experienced IT and automation project veterans will appreciate the technical specifics more fully than the general reader.
    Tennant's collection covers a variety of well- and lesser-known XML-based pilot and prototype projects undertaken by libraries around the world. Some of the projects included are: Stanford's XMLMARC conversion, Oregon State's use of XML in interlibrary loaning, e-books (California Digital Library) and electronic scholarly publishing (University of Michigan), the Washington Research Library Consortium's XML-based Web Services, and using TEI Lite to support indexing (Halton Hills Public Library). Of the 13 projects presented, nine are sited in academe, three are state library endeavors, and one is an American public library initiative. The projects are gathered into sections grouped by seven library applications: the use of XML in library catalog records, interlibrary loan, cataloging and indexing, collection building, databases, data migration, and systems interoperability. Each project is introduced with a few paragraphs of background information. The project reports-averaging about 13 pages each-include project goals and justification, project description, challenges and lessons learned (successes and failures), future plans, implications of the work, contact information for individual(s) responsible for the project, and relevant Web links and resources. The clear strengths of this collection are in the details and the consistency of presentation. The concise project write-ups flow well and encourage interested readers to follow-up via personal contacts and URLs. The sole weakness is the price. XML in Libraries will excite and inspire institutions and organizations with technically adept staff resources and visionary leaders. Erik Ray has written a how-to book. Unlike most, Learning XML is not aimed at the professional programming community. The intended audience is readers familiar with a structured markup (HTML, TEX, etc.) and Web concepts (hypertext links, data representation). In the first six chapters, Ray introduces XMUs main concepts and tools for writing, viewing, testing, and transforming XML (chapter 1), describes basic syntax (chapter 2), discusses linking with XLink and XPointer (chapter 3), introduces Cascading Style Sheets for use with XML (chapter 4), explains document type definitions (DTDs) and schemas (chapter 5), and covers XSLT stylesheets and XPath (chapter 6). Chapter 7 introduces Unicode, internationalization and language support, including CSS and XSLT encoding. Chapter 8 is an overview of writing software for processing XML, and includes the Perl code for an XML syntax checker. This work is written very accessibly for nonprogrammers. Writers, designers, and students just starting to acquire Web technology skills will find Ray's style approachable. Concepts are introduced in a logical flow, and explained clearly. Code samples (130+), illustrations and screen shots (50+), and numerous tables are distributed throughout the text. Ray uses a modified DocBook DTD and a checkbook example throughout, introducing concepts in early chapters and adding new concepts to them. Readers become familiar with the code and its evolution through repeated exposure. The code for converting the "barebones DocBook" DTD (10 pages of code) to HTML via XSLT stylesheet occupies 19 pages. Both code examples allow the learner to sec an accumulation of snippets incorporated into a sensible whole. While experienced programmers might not need this type of support, nonprogrammers certainly do. Using the checkbook example is an inspired choice: Most of us are familiar with personal checking, even if few of us world build an XML application for it. Learning XML is an excellent textbook. I've used it for several years as a recommended text for adult continuing education courses and workshops."
  15. Learning XML (2003) 0.01
    0.014903244 = product of:
      0.029806487 = sum of:
        0.0076767267 = weight(_text_:information in 3101) [ClassicSimilarity], result of:
          0.0076767267 = score(doc=3101,freq=10.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.0867392 = fieldWeight in 3101, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.015625 = fieldNorm(doc=3101)
        0.022129761 = weight(_text_:standards in 3101) [ClassicSimilarity], result of:
          0.022129761 = score(doc=3101,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.09848571 = fieldWeight in 3101, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.015625 = fieldNorm(doc=3101)
      0.5 = coord(2/4)
    
    Footnote
    Rez. in: JASIST 55(2004) no.14, S.1304-1305 (Z. Holbrooks):"The eXtensible Markup Language (XML) and its family of enabling technologies (XPath, XPointer, XLink, XSLT, et al.) were the new "new thing" only a couple of years ago. Happily, XML is now a W3C standard, and its enabling technologies are rapidly proliferating and maturing. Together, they are changing the way data is handled an the Web, how legacy data is accessed and leveraged in corporate archives, and offering the Semantic Web community a powerful toolset. Library and information professionals need a basic understanding of what XML is, and what its impacts will be an the library community as content vendors and publishers convert to the new standards. Norman Desmarais aims to provide librarians with an overview of XML and some potential library applications. The ABCs of XML contains the useful basic information that most general XML works cover. It is addressed to librarians, as evidenced by the occasional reference to periodical vendors, MARC, and OPACs. However, librarians without SGML, HTML, database, or programming experience may find the work daunting. The snippets of code-most incomplete and unattended by screenshots to illustrate the result of the code's execution-obscure more often than they enlighten. A single code sample (p. 91, a book purchase order) is immediately recognizable and sensible. There are no figures, illustrations, or screenshots. Subsection headings are used conservatively. Readers are confronted with page after page of unbroken technical text, and occasionally oddly formatted text (in some of the code samples). The author concentrates an commercial products and projects. Library and agency initiatives-for example, the National Institutes of Health HL-7 and U.S. Department of Education's GEM project-are notable for their absence. The Library of Congress USMARC to SGML effort is discussed in chapter 1, which covers the relationship of XML to its parent SGML, the XML processor, and data type definitions, using MARC as its illustrative example. Chapter 3 addresses the stylesheet options for XML, including DSSSL, CSS, and XSL. The Document Style Semantics and Specification Language (DSSSL) was created for use with SGML, and pruned into DSSSL-Lite and further (DSSSL-online). Cascading Style Sheets (CSS) were created for use with HTML. Extensible Style Language (XSL) is a further revision (and extension) of DSSSL-o specifically for use with XML. Discussion of aural stylesheets and Synchronized Multimedia Integration Language (SMIL) round out the chapter.
    Chapter 4 introduces XML internal and external pointing and linking technologies. XML Link Language (XLL, now XLink) provides unidirectional, multi-ended, and typed linking. XPointer, used with XLink, provides addressing into the interior of XML documents. XPath operates an the logical structure of an XML document, creating a tree of nodes. Used with both XPointer and XSLT, it permits operations an strings, numbers, and Boolean expressions in the document. The final chapter, "Getting Started" argues for the adoption of a tool for XML production. The features and functionality of various tools for content development, application development, databases, and schema development provide an introduction to some of the available options. Roy Tennant is weIl known in the library community as an author (bis column "Digital Libraries" has appeared in Library Journal since 1997 and he has published Current Cites each month for more than a decade), an electronic discussion list manager (Web4Lib and XML4Lib), and as the creator and manager of UC/Berkeley's Digital Library SunSITE. Librarians have wondered what use they might make of XML since its beginnings. Tennant suggests one answer: "The Extensible Markup Language (XML) has the potential to exceed the impact of MARC an librarianship. While MARC is limited to bibliographic descriptionand arguably a subset at that, as any archivist will tell you-XML provides a highly-effective framework for encoding anything from a bibliographic record for a book to the book itself." (Tennant, p. vii) This slim paperback volume offers librarians and library managers concerned with automation projects "show and teIl" examples of XML technologies used as solutions to everyday tasks and challenges. What distinguishes this work is the editor and contributors' commitment to providing messy details. This book's target audience is technically savvy. While not a "cookbook" per se, the information provided an each project serves as a draft blueprint complete with acronyms and jargon. The inclusion of "lessons learned" (including failures as well as successes) is refreshing and commendable. Experienced IT and automation project veterans will appreciate the technical specifics more fully than the general reader.
    Tennant's collection covers a variety of well- and lesser-known XML-based pilot and prototype projects undertaken by libraries around the world. Some of the projects included are: Stanford's XMLMARC conversion, Oregon State's use of XML in interlibrary loaning, e-books (California Digital Library) and electronic scholarly publishing (University of Michigan), the Washington Research Library Consortium's XML-based Web Services, and using TEI Lite to support indexing (Halton Hills Public Library). Of the 13 projects presented, nine are sited in academe, three are state library endeavors, and one is an American public library initiative. The projects are gathered into sections grouped by seven library applications: the use of XML in library catalog records, interlibrary loan, cataloging and indexing, collection building, databases, data migration, and systems interoperability. Each project is introduced with a few paragraphs of background information. The project reports-averaging about 13 pages each-include project goals and justification, project description, challenges and lessons learned (successes and failures), future plans, implications of the work, contact information for individual(s) responsible for the project, and relevant Web links and resources. The clear strengths of this collection are in the details and the consistency of presentation. The concise project write-ups flow well and encourage interested readers to follow-up via personal contacts and URLs. The sole weakness is the price. XML in Libraries will excite and inspire institutions and organizations with technically adept staff resources and visionary leaders. Erik Ray has written a how-to book. Unlike most, Learning XML is not aimed at the professional programming community. The intended audience is readers familiar with a structured markup (HTML, TEX, etc.) and Web concepts (hypertext links, data representation). In the first six chapters, Ray introduces XMUs main concepts and tools for writing, viewing, testing, and transforming XML (chapter 1), describes basic syntax (chapter 2), discusses linking with XLink and XPointer (chapter 3), introduces Cascading Style Sheets for use with XML (chapter 4), explains document type definitions (DTDs) and schemas (chapter 5), and covers XSLT stylesheets and XPath (chapter 6). Chapter 7 introduces Unicode, internationalization and language support, including CSS and XSLT encoding. Chapter 8 is an overview of writing software for processing XML, and includes the Perl code for an XML syntax checker. This work is written very accessibly for nonprogrammers. Writers, designers, and students just starting to acquire Web technology skills will find Ray's style approachable. Concepts are introduced in a logical flow, and explained clearly. Code samples (130+), illustrations and screen shots (50+), and numerous tables are distributed throughout the text. Ray uses a modified DocBook DTD and a checkbook example throughout, introducing concepts in early chapters and adding new concepts to them. Readers become familiar with the code and its evolution through repeated exposure. The code for converting the "barebones DocBook" DTD (10 pages of code) to HTML via XSLT stylesheet occupies 19 pages. Both code examples allow the learner to sec an accumulation of snippets incorporated into a sensible whole. While experienced programmers might not need this type of support, nonprogrammers certainly do. Using the checkbook example is an inspired choice: Most of us are familiar with personal checking, even if few of us world build an XML application for it. Learning XML is an excellent textbook. I've used it for several years as a recommended text for adult continuing education courses and workshops."
  16. ¬The ABCs of XML : the librarian's guide to the eXtensible Markup Language (2000) 0.01
    0.014903244 = product of:
      0.029806487 = sum of:
        0.0076767267 = weight(_text_:information in 3102) [ClassicSimilarity], result of:
          0.0076767267 = score(doc=3102,freq=10.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.0867392 = fieldWeight in 3102, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.015625 = fieldNorm(doc=3102)
        0.022129761 = weight(_text_:standards in 3102) [ClassicSimilarity], result of:
          0.022129761 = score(doc=3102,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.09848571 = fieldWeight in 3102, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.015625 = fieldNorm(doc=3102)
      0.5 = coord(2/4)
    
    Footnote
    Rez. in: JASIST 55(2004) no.14, S.1304-1305 (Z. Holbrooks):"The eXtensible Markup Language (XML) and its family of enabling technologies (XPath, XPointer, XLink, XSLT, et al.) were the new "new thing" only a couple of years ago. Happily, XML is now a W3C standard, and its enabling technologies are rapidly proliferating and maturing. Together, they are changing the way data is handled an the Web, how legacy data is accessed and leveraged in corporate archives, and offering the Semantic Web community a powerful toolset. Library and information professionals need a basic understanding of what XML is, and what its impacts will be an the library community as content vendors and publishers convert to the new standards. Norman Desmarais aims to provide librarians with an overview of XML and some potential library applications. The ABCs of XML contains the useful basic information that most general XML works cover. It is addressed to librarians, as evidenced by the occasional reference to periodical vendors, MARC, and OPACs. However, librarians without SGML, HTML, database, or programming experience may find the work daunting. The snippets of code-most incomplete and unattended by screenshots to illustrate the result of the code's execution-obscure more often than they enlighten. A single code sample (p. 91, a book purchase order) is immediately recognizable and sensible. There are no figures, illustrations, or screenshots. Subsection headings are used conservatively. Readers are confronted with page after page of unbroken technical text, and occasionally oddly formatted text (in some of the code samples). The author concentrates an commercial products and projects. Library and agency initiatives-for example, the National Institutes of Health HL-7 and U.S. Department of Education's GEM project-are notable for their absence. The Library of Congress USMARC to SGML effort is discussed in chapter 1, which covers the relationship of XML to its parent SGML, the XML processor, and data type definitions, using MARC as its illustrative example. Chapter 3 addresses the stylesheet options for XML, including DSSSL, CSS, and XSL. The Document Style Semantics and Specification Language (DSSSL) was created for use with SGML, and pruned into DSSSL-Lite and further (DSSSL-online). Cascading Style Sheets (CSS) were created for use with HTML. Extensible Style Language (XSL) is a further revision (and extension) of DSSSL-o specifically for use with XML. Discussion of aural stylesheets and Synchronized Multimedia Integration Language (SMIL) round out the chapter.
    Chapter 4 introduces XML internal and external pointing and linking technologies. XML Link Language (XLL, now XLink) provides unidirectional, multi-ended, and typed linking. XPointer, used with XLink, provides addressing into the interior of XML documents. XPath operates an the logical structure of an XML document, creating a tree of nodes. Used with both XPointer and XSLT, it permits operations an strings, numbers, and Boolean expressions in the document. The final chapter, "Getting Started" argues for the adoption of a tool for XML production. The features and functionality of various tools for content development, application development, databases, and schema development provide an introduction to some of the available options. Roy Tennant is weIl known in the library community as an author (bis column "Digital Libraries" has appeared in Library Journal since 1997 and he has published Current Cites each month for more than a decade), an electronic discussion list manager (Web4Lib and XML4Lib), and as the creator and manager of UC/Berkeley's Digital Library SunSITE. Librarians have wondered what use they might make of XML since its beginnings. Tennant suggests one answer: "The Extensible Markup Language (XML) has the potential to exceed the impact of MARC an librarianship. While MARC is limited to bibliographic descriptionand arguably a subset at that, as any archivist will tell you-XML provides a highly-effective framework for encoding anything from a bibliographic record for a book to the book itself." (Tennant, p. vii) This slim paperback volume offers librarians and library managers concerned with automation projects "show and teIl" examples of XML technologies used as solutions to everyday tasks and challenges. What distinguishes this work is the editor and contributors' commitment to providing messy details. This book's target audience is technically savvy. While not a "cookbook" per se, the information provided an each project serves as a draft blueprint complete with acronyms and jargon. The inclusion of "lessons learned" (including failures as well as successes) is refreshing and commendable. Experienced IT and automation project veterans will appreciate the technical specifics more fully than the general reader.
    Tennant's collection covers a variety of well- and lesser-known XML-based pilot and prototype projects undertaken by libraries around the world. Some of the projects included are: Stanford's XMLMARC conversion, Oregon State's use of XML in interlibrary loaning, e-books (California Digital Library) and electronic scholarly publishing (University of Michigan), the Washington Research Library Consortium's XML-based Web Services, and using TEI Lite to support indexing (Halton Hills Public Library). Of the 13 projects presented, nine are sited in academe, three are state library endeavors, and one is an American public library initiative. The projects are gathered into sections grouped by seven library applications: the use of XML in library catalog records, interlibrary loan, cataloging and indexing, collection building, databases, data migration, and systems interoperability. Each project is introduced with a few paragraphs of background information. The project reports-averaging about 13 pages each-include project goals and justification, project description, challenges and lessons learned (successes and failures), future plans, implications of the work, contact information for individual(s) responsible for the project, and relevant Web links and resources. The clear strengths of this collection are in the details and the consistency of presentation. The concise project write-ups flow well and encourage interested readers to follow-up via personal contacts and URLs. The sole weakness is the price. XML in Libraries will excite and inspire institutions and organizations with technically adept staff resources and visionary leaders. Erik Ray has written a how-to book. Unlike most, Learning XML is not aimed at the professional programming community. The intended audience is readers familiar with a structured markup (HTML, TEX, etc.) and Web concepts (hypertext links, data representation). In the first six chapters, Ray introduces XMUs main concepts and tools for writing, viewing, testing, and transforming XML (chapter 1), describes basic syntax (chapter 2), discusses linking with XLink and XPointer (chapter 3), introduces Cascading Style Sheets for use with XML (chapter 4), explains document type definitions (DTDs) and schemas (chapter 5), and covers XSLT stylesheets and XPath (chapter 6). Chapter 7 introduces Unicode, internationalization and language support, including CSS and XSLT encoding. Chapter 8 is an overview of writing software for processing XML, and includes the Perl code for an XML syntax checker. This work is written very accessibly for nonprogrammers. Writers, designers, and students just starting to acquire Web technology skills will find Ray's style approachable. Concepts are introduced in a logical flow, and explained clearly. Code samples (130+), illustrations and screen shots (50+), and numerous tables are distributed throughout the text. Ray uses a modified DocBook DTD and a checkbook example throughout, introducing concepts in early chapters and adding new concepts to them. Readers become familiar with the code and its evolution through repeated exposure. The code for converting the "barebones DocBook" DTD (10 pages of code) to HTML via XSLT stylesheet occupies 19 pages. Both code examples allow the learner to sec an accumulation of snippets incorporated into a sensible whole. While experienced programmers might not need this type of support, nonprogrammers certainly do. Using the checkbook example is an inspired choice: Most of us are familiar with personal checking, even if few of us world build an XML application for it. Learning XML is an excellent textbook. I've used it for several years as a recommended text for adult continuing education courses and workshops."
  17. Marcoux, Y.; Rizkallah, E.: Knowledge organization in the light of intertextual semantics : a natural-language analysis of controlled vocabularies (2008) 0.01
    0.010621051 = product of:
      0.042484205 = sum of:
        0.042484205 = product of:
          0.08496841 = sum of:
            0.08496841 = weight(_text_:organization in 2241) [ClassicSimilarity], result of:
              0.08496841 = score(doc=2241,freq=8.0), product of:
                0.17974974 = queryWeight, product of:
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.050415643 = queryNorm
                0.47270393 = fieldWeight in 2241, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2241)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Content
    Intertextual semantics is a semiotics-based approach to the design of communication artefacts primarily aimed at modeling XML structured documents. SKOS (Simple Knowledge Organization System) is a specification currently under development at the W3C that allows expressing various types of controlled vocabularies in XML. In this article, we show through an example how intertextual semantics could be applied to controlled vocabularies expressed in SKOS, and argue that it could facilitate the communication of meaning among the various persons who interact with a controlled vocabulary.
    Series
    Advances in knowledge organization; vol.11
    Source
    Culture and identity in knowledge organization: Proceedings of the Tenth International ISKO Conference 5-8 August 2008, Montreal, Canada. Ed. by Clément Arsenault and Joseph T. Tennis
  18. as: XML: Extensible Markup Language : I: Was ist XML? (2001) 0.01
    0.00853828 = product of:
      0.03415312 = sum of:
        0.03415312 = product of:
          0.06830624 = sum of:
            0.06830624 = weight(_text_:22 in 4950) [ClassicSimilarity], result of:
              0.06830624 = score(doc=4950,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.38690117 = fieldWeight in 4950, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4950)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    30. 3.2003 11:06:22
  19. Jackenkroll, M.: Nutzen von XML für die Herstellung verschiedener medialer Varianten von Informationsmitteln : dargestellt am Beispiel eines geografischen Lexikonartikels (2002) 0.01
    0.008298661 = product of:
      0.033194643 = sum of:
        0.033194643 = weight(_text_:standards in 4804) [ClassicSimilarity], result of:
          0.033194643 = score(doc=4804,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.14772856 = fieldWeight in 4804, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.0234375 = fieldNorm(doc=4804)
      0.25 = coord(1/4)
    
    Content
    "Die Extensible Markup Language (XML) ist eine Metaauszeichnungssprache, die 1998 vom World Wide Web Consortium (W3C), einer Organisation, die sich mit der Erstellung von Web Standards und neuen Technologien für das Internet beschäftigt, als neue Empfehlung für Web-Anwendungen festgesetzt wurde. Seitdem ist viel über XML und die sich durch diese Sprache ergebenden neuen Möglichkeiten des Datenaustausches über das Internet publiziert worden. In XML-Dokumenten werden die hierarchische Struktur und der Inhalt der Dokumente festgelegt, aber keinerlei Angaben zum Layout gemacht. Dieses wird in so genannten Stylesheets definiert. Mit Hilfe mehrerer Stylesheets, die sich alle auf ein XML-Dokument beziehen, ist es möglich, aus einem Datenbestand verschiedene Ausgabeprodukte, z.B. eine Online-Version und eine druckbare Ausgabe eines Dokuments, zu erzeugen. Diese Möglichkeit der Herstellung verschiedener medialer Varianten eines Produkts ist auch für die Herstellung von Informationsmitteln interessant. Im Bereich der Produktion von Informationsmitteln, vor allem von Lexika und Enzyklopädien, ist in den letzten Jahren zu beobachten gewesen, dass neben der traditionellen, gedruckten Ausgabe des Nachschlagewerks zunehmend auch elektronische Varianten, die durch multimediale Elemente angereichert sind, angeboten werden. Diese elektronischen Nachschlagewerke werden sowohl offline, d.h. auf CD-ROM bzw. DVD, als auch online im Internet veröffentlicht. Im Gegensatz zu den gedruckten Versionen werden die neuen Produkte fast jährlich aktualisiert. Diese neue Situation erforderte Veränderungen im Herstellungsprozess. Ein Verfahren, das die Erzeugung verschiedener medialer Varianten eines Produkts möglichst einfach und problemlos ermöglicht, wurde benötigt. XML und ihr Vorgänger, die Standard Generalized Markup Language (SGML), schienen die perfekte Lösung für dieses Problem zu sein. Die Erwartungen an den Nutzen, den SGML und XML bringen könnten, waren hoch: "Allein dieses Spitzklammerformat, eingespeist in einen Datenpool, soll auf Knopfdruck die Generierung der verschiedensten Medienprodukte ermöglichen". Ziel dieser Arbeit ist es, darzustellen, wie der neue Standard XML bei der Publikation von Informationsmitteln eingesetzt werden kann, um aus einem einmal erfassten Datenbestand mit möglichst geringem Aufwand mehrere Ausgabeprodukte zu generieren. Es wird darauf eingegangen, welche Ausgabeformen sich in diesem Bereich für XML-Dokumente anbieten und mit welchen Verfahren und Hilfsmitteln die jeweiligen Ausgabeformate erstellt werden können. In diesem Zusammenhang sollen auch die Aspekte behandelt werden, die sich bei der Umwandlung von XML-Dokumenten in andere For mate unter Umständen als problematisch erweisen könnten.
  20. Schröder, A.: Web der Zukunft : RDF - Der erste Schritt zum semantischen Web 0.01
    0.0068306234 = product of:
      0.027322493 = sum of:
        0.027322493 = product of:
          0.054644987 = sum of:
            0.054644987 = weight(_text_:22 in 1457) [ClassicSimilarity], result of:
              0.054644987 = score(doc=1457,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.30952093 = fieldWeight in 1457, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1457)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Seit dem 22. Februar 1999 ist das Resource Description Framework (RDF) als W3C-Empfehlung verfügbar. Doch was steckt hinter diesem Standard, der das Zeitalter des Semantischen Webs einläuten soll? Was RDF bedeutet, wozu man es einsetzt, welche Vorteile es gegenüber XML hat und wie man RDF anwendet, soll in diesem Artikel erläutert werden.

Languages

  • e 23
  • d 13

Types