Search (15 results, page 1 of 1)

  • × type_ss:"el"
  • × type_ss:"s"
  1. nestor-Handbuch : eine kleine Enzyklopädie der digitalen Langzeitarchivierung (2010) 0.03
    0.033348635 = product of:
      0.06669727 = sum of:
        0.061649315 = weight(_text_:digitale in 3716) [ClassicSimilarity], result of:
          0.061649315 = score(doc=3716,freq=8.0), product of:
            0.18027179 = queryWeight, product of:
              5.158747 = idf(docFreq=690, maxDocs=44218)
              0.034944877 = queryNorm
            0.34197983 = fieldWeight in 3716, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              5.158747 = idf(docFreq=690, maxDocs=44218)
              0.0234375 = fieldNorm(doc=3716)
        0.0050479556 = weight(_text_:information in 3716) [ClassicSimilarity], result of:
          0.0050479556 = score(doc=3716,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.08228803 = fieldWeight in 3716, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0234375 = fieldNorm(doc=3716)
      0.5 = coord(2/4)
    
    Abstract
    Glücklicher als Sie mit Ihren privaten digitalen Daten sind da die Astronomen, wenn sie nach Daten von Himmels-Beobachtungen fahnden, die bereits Jahrzehnte zurückliegen. Obwohl die Bild- und Datenarchive dieser Beobachtungen in vielfältigen und sehr unterschiedlichen Formaten abgespeichert wurden, gibt es immer die Möglichkeit, über geeignete Interface-Verfahren die Originaldaten zu lesen und zu interpretieren. Dies ist der Fall, weil durch das sogenannte Virtuelle Observatorium weltweit die Archive für astronomische Beobachtungen vernetzt und immer in den neuesten digitalen Formaten zugänglich sind, seien es digitale Aufnahmen von Asteroiden, Planetenbewegungen, der Milchstrasse oder auch Simulationen des Urknalls. Selbst Photoplatten von Beginn des 20. Jahrhunderts wurden systematisch digitalisiert und stehen zur Wiederverwendung bereit. So sind ältere und neue digitale Daten und Bilder gemeinsam nutzbar und gewähren einen Blick in das Universum, der sich über weit mehr Wellenlängen erstreckt als die Sinne des Menschen allein wahrnehmen können. Wir freuen uns, Ihnen mit dem nestor Handbuch "Eine kleine Enzyklopädie der digitalen Langzeitarchivierung" den aktuellen Wissensstand über die Langzeitarchivierung digitaler Objekte im Überblick sowie aus vielen Teilbereichen nun auch in gedruckter Form präsentieren zu können. Schon seit Frühjahr 2007 ist das Handbuch in digitaler Version unter http://nestor.sub.uni-goettingen.de/handbuch/ verfügbar und seitdem in mehreren Intervallen aktualisiert worden. Die nun vorliegende Version 2.0 - hier gedruckt und unter o.g. URL auch weiterhin entgeltfrei herunterladbar - wurde neu strukturiert, um neue Themenfelder ergänzt und bislang schon vorhandene Beiträge wurden, wo fachlich geboten, überarbeitet. Aus seiner Entstehung ergibt sich eine gewisse Heterogenität der einzelnen Kapitel untereinander, z.B. bezüglich der Ausführlichkeit des behandelten Themas oder des Schreibstils. Der Herausgeberkreis hat nicht primär das Ziel verfolgt, dies redaktionell lektorierend auszugleichen oder ein insgesamt kohärentes Gesamtwerk vorzulegen. Vielmehr geht es ihm darum, der deutschsprachigen Gemeinschaft eine möglichst aktuelle "Kleine Enzyklopädie der digitalen Langzeitarchivierung" anbieten zu können.
    Die parallel verfügbare entgeltfreie, digitale Version des Handbuchs wird bei Bedarf aktualisiert und erweitert, eine zweite Druckauflage ist bereits geplant. Gerne nehmen wir Ihre Anregungen auf und berücksichtigen sie bei zukünftigen Aktualisierungen! Unser Dank gilt insbesondere den Autorinnen und Autoren, ohne die es nur bei der Idee eines solchen Handbuches geblieben wäre. Mein Dank gilt aber auch den Mitherausgebern dieser Ausgabe, durch deren engagiertes Stimulieren und "Bändigen" der Autoren die vielen Beiträge erst zu einem Gesamtwerk zusammengeführt werden konnten. Zusammen mit allen Beteiligten hoffe ich, dass dieses Handbuch Ihnen hilfreiche Anregungen und Anleitungen zu einem erfolgreichen Einstieg in die Theorie und Praxis der Langzeitarchivierung digitaler Objekte bietet!
    Content
    Einführung - State of the Art - Rahmenbedingungen für die LZA digitaler Objekte - Das Referenzmodell OAIS - Open Archival Information System - Das Referenzmodell OAIS - Open Archival Information System - Metadatenstandards im Bereich der digitalen LZA - Formate - Digitale Erhaltungsstrategien - Access - Hardware - Speichersysteme mit Langzeitarchivierungsanspruch - Technischer Workflow - Tools - Geschäftsmodelle - Organisation - Recht - Vorgehensweise für ausgewählte Objekttypen - Praxisbeispiele - Qualifizierung im Themenbereich "Langzeitarchivierung digitaler Objekte"
  2. Vernetztes Wissen - Daten, Menschen, Systeme : 6. Konferenz der Zentralbibliothek Forschungszentrum Jülich. 5. - 7. November 2012 - Proceedingsband: WissKom 2012 (2012) 0.03
    0.02866175 = product of:
      0.0573235 = sum of:
        0.051374428 = weight(_text_:digitale in 482) [ClassicSimilarity], result of:
          0.051374428 = score(doc=482,freq=2.0), product of:
            0.18027179 = queryWeight, product of:
              5.158747 = idf(docFreq=690, maxDocs=44218)
              0.034944877 = queryNorm
            0.2849832 = fieldWeight in 482, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.158747 = idf(docFreq=690, maxDocs=44218)
              0.0390625 = fieldNorm(doc=482)
        0.0059490725 = weight(_text_:information in 482) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=482,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 482, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=482)
      0.5 = coord(2/4)
    
    Abstract
    Informations- und Wissensvermittlung verlagern sich immer stärker in die digitale Welt. Möglich wird dies nicht zuletzt durch die voranschreitende Durchdringung aller Lebensbereiche durch das Internet. Wissen wird mehr und mehr zu vernetztem Wissen. Die Jülicher Konferenz WissKom2012 thematisiert die Anpassung an diese Entwicklung und ihre Mitgestaltung durch innovative Bibliotheksdienstleistungen. Der Konferenztitel "Vernetztes Wissen: Daten, Menschen, Systeme" deutet die wechselseitige Vernetzung unter- und miteinander an. Ziel ist, vorhandene Insellösungen zu verbinden und neue Konzepte für inhärent vernetzte Strukturen zu entwickeln. Mit der WissKom2012 "Vernetztes Wissen - Daten, Menschen, Systeme" greift die Zentralbibliothek des Forschungszentrums Jülich erneut Themen im Spannungsfeld von "Bibliothek - Information - Wissenschaft" in einer Konferenz interdisziplinär auf und versucht, neue Handlungsfelder für Bibliotheken aufzuzeigen. Diese sechste Konferenz der Zentralbibliothek thematisiert den immer wichtiger werdenden Bereich der Forschungsdaten und den nachhaltigen Umgang mit ihnen. Sie zeigt auf, was Interdisziplinarität konkret bedeutet und wie bislang isolierte Systeme vernetzt werden können und so Mehrwert entsteht. Der Konferenzband enthält neben den Ausführungen der Referenten zudem die Beiträge der Poster Session sowie den Festvortrag von Prof. Viktor Mayer-Schönberger mit dem Titel "Delete: Die Tugend des Vergessens in digitalen Zeiten".
  3. Networked knowledge organization systems (2001) 0.00
    0.0035694437 = product of:
      0.014277775 = sum of:
        0.014277775 = weight(_text_:information in 6473) [ClassicSimilarity], result of:
          0.014277775 = score(doc=6473,freq=8.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.23274569 = fieldWeight in 6473, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=6473)
      0.25 = coord(1/4)
    
    Abstract
    Knowledge Organization Systems can comprise thesauri and other controlled lists of keywords, ontologies, classification systems, clustering approaches, taxonomies, gazetteers, dictionaries, lexical databases, concept maps/spaces, semantic road maps, etc. These schemas enable knowledge structuring and management, knowledge-based data processing and systematic access to knowledge structures in individual collections and digital libraries. Used as interactive information services on the Internet they have an increased potential to support the description, discovery and retrieval of heterogeneous information resources and to contribute to an overall resource discovery infrastructure
    Content
    This issue of the Journal of Digital Information evolved from a workshop on Networked Knowledge Organization Systems (NKOS) held at the Fourth European Conference on Research and Advanced Technology for Digital Libraries (ECDL2000) in Lisbon during September 2000. The focus of the workshop was European NKOS initiatives and projects and options for global cooperation. Workshop organizers were Martin Doerr, Traugott Koch, Dougles Tudhope and Repke de Vries. This group has, with Traugott Koch as the main editor and with the help of Linda Hill, cooperated in the editorial tasks for this special issue
    Source
    Journal of digital information. 1(2001) no.8
  4. Tudhope, D.: New Applications of Knowledge Organization Systems : introduction to a special issue (2004) 0.00
    0.0035694437 = product of:
      0.014277775 = sum of:
        0.014277775 = weight(_text_:information in 2344) [ClassicSimilarity], result of:
          0.014277775 = score(doc=2344,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.23274569 = fieldWeight in 2344, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=2344)
      0.25 = coord(1/4)
    
    Footnote
    Journal of digital information. 4(2004) no.4.
  5. ¬Third International World Wide Web Conference, Darmstadt 1995 : [Inhaltsverzeichnis] (1995) 0.00
    0.003091229 = product of:
      0.012364916 = sum of:
        0.012364916 = weight(_text_:information in 3458) [ClassicSimilarity], result of:
          0.012364916 = score(doc=3458,freq=6.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.20156369 = fieldWeight in 3458, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3458)
      0.25 = coord(1/4)
    
    Abstract
    ANDREW, K. u. F. KAPPE: Serving information to the Web with Hyper-G; BARBIERI, K., H.M. DOERR u. D. DWYER: Creating a virtual classroom for interactive education on the Web; CAMPBELL, J.K., S.B. JONES, N.M. STEPHENS u. S. HURLEY: Constructing educational courseware using NCSA Mosaic and the World Wide Web; CATLEDGE, L.L. u. J.E. PITKOW: Characterizing browsing strategies in the World-Wide Web; CLAUSNITZER, A. u. P. VOGEL: A WWW interface to the OMNIS/Myriad literature retrieval engine; FISCHER, R. u. L. PERROCHON: IDLE: Unified W3-access to interactive information servers; FOLEY, J.D.: Visualizing the World-Wide Web with the navigational view builder; FRANKLIN, S.D. u. B. IBRAHIM: Advanced educational uses of the World-Wide Web; FUHR, N., U. PFEIFER u. T. HUYNH: Searching structured documents with the enhanced retrieval functionality of free WAIS-sf and SFgate; FIORITO, M., J. OKSANEN u. D.R. IOIVANE: An educational environment using WWW; KENT, R.E. u. C. NEUSS: Conceptual analysis of resource meta-information; SHELDON, M.A. u. R. WEISS: Discover: a resource discovery system based on content routing; WINOGRAD, T.: Beyond browsing: shared comments, SOAPs, Trails, and On-line communities
  6. New applications of knowledge organization systems (2004) 0.00
    0.0029446408 = product of:
      0.011778563 = sum of:
        0.011778563 = weight(_text_:information in 2343) [ClassicSimilarity], result of:
          0.011778563 = score(doc=2343,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.1920054 = fieldWeight in 2343, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2343)
      0.25 = coord(1/4)
    
    Abstract
    Knowledge Organization Systems/Services (KOS), such as classifications, gazetteers, lexical databases, ontologies, taxonomies and thesauri, model the underlying semantic structure of a domain. They can support subject indexing and facilitate resource discovery and retrieval, whether by humans or by machines. New networked KOS services and applications are emerging and we are reaching the stage where we can prepare the work for future exploitation of common representations and protocols for distributed use. A number of technologies could be combined to yield new solutions. The papers published here are concerned with different types of KOS, discuss various standards issues and span the information lifecycle.
    Source
    Journal of digital information. 4(2004) no.4,
  7. Rauber, A.: Digital preservation in data-driven science : on the importance of process capture, preservation and validation (2012) 0.00
    0.0025239778 = product of:
      0.010095911 = sum of:
        0.010095911 = weight(_text_:information in 469) [ClassicSimilarity], result of:
          0.010095911 = score(doc=469,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.16457605 = fieldWeight in 469, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=469)
      0.25 = coord(1/4)
    
    Abstract
    Current digital preservation is strongly biased towards data objects: digital files of document-style objects, or encapsulated and largely self-contained objects. To provide authenticity and provenance information, comprehensive metadata models are deployed to document information on an object's context. Yet, we claim that simply documenting an objects context may not be sufficient to ensure proper provenance and to fulfill the stated preservation goals. Specifically in e-Science and business settings, capturing, documenting and preserving entire processes may be necessary to meet the preservation goals. We thus present an approach for capturing, documenting and preserving processes, and means to assess their authenticity upon re-execution. We will discuss options as well as limitations and open challenges to achieve sound preservation, speci?cally within scientific processes.
  8. Proceedings of the 2nd International Workshop on Semantic Digital Archives held in conjunction with the 16th Int. Conference on Theory and Practice of Digital Libraries (TPDL) on September 27, 2012 in Paphos, Cyprus (2012) 0.00
    0.0023609651 = product of:
      0.0094438605 = sum of:
        0.0094438605 = weight(_text_:information in 468) [ClassicSimilarity], result of:
          0.0094438605 = score(doc=468,freq=14.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.1539468 = fieldWeight in 468, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0234375 = fieldNorm(doc=468)
      0.25 = coord(1/4)
    
    Abstract
    Archival Information Systems (AIS) are becoming increasingly important. For decades, the amount of content created digitally is growing and its complete life cycle nowadays tends to remain digital. A selection of this content is expected to be of value for the future and can thus be considered being part of our cultural heritage. However, digital content poses many challenges for long-term or indefinite preservation, e.g. digital publications become increasingly complex by the embedding of different kinds of multimedia, data in arbitrary formats and software. As soon as these digital publications become obsolete, but are still deemed to be of value in the future, they have to be transferred smoothly into appropriate AIS where they need to be kept accessible even through changing technologies. The successful previous SDA workshop in 2011 showed: Both, the library and the archiving community have made valuable contributions to the management of huge amounts of knowledge and data. However, both are approaching this topic from different views which shall be brought together to cross-fertilize each other. There are promising combinations of pertinence and provenance models since those are traditionally the prevailing knowledge organization principles of the library and archiving community, respectively. Another scientific discipline providing promising technical solutions for knowledge representation and knowledge management is semantic technologies, which is supported by appropriate W3C recommendations and a large user community. At the forefront of making the semantic web a mature and applicable reality is the linked data initiative, which already has started to be adopted by the library community. It can be expected that using semantic (web) technologies in general and linked data in particular can mature the area of digital archiving as well as technologically tighten the natural bond between digital libraries and digital archives. Semantic representations of contextual knowledge about cultural heritage objects will enhance organization and access of data and knowledge. In order to achieve a comprehensive investigation, the information seeking and document triage behaviors of users (an area also classified under the field of Human Computer Interaction) will also be included in the research.
    Semantic search & semantic information retrieval in digital archives and digital libraries Semantic multimedia archives Ontologies & linked data for digital archives and digital libraries Ontologies & linked data for multimedia archives Implementations and evaluations of semantic digital archives Visualization and exploration of digital content User interfaces for semantic digital libraries User interfaces for intelligent multimedia information retrieval User studies focusing on end-user needs and information seeking behavior of end-users Theoretical and practical archiving frameworks using Semantic (Web) technologies Logical theories for digital archives Semantic (Web) services implementing the OAIS standard Semantic or logical provenance models for digital archives or digital libraries Information integration/semantic ingest (e.g. from digital libraries) Trust for ingest and data security/integrity check for long-term storage of archival records Semantic extensions of emulation/virtualization methodologies tailored for digital archives Semantic long-term storage and hardware organization tailored for AIS Migration strategies based on Semantic (Web) technologies Knowledge evolution We expect new insights and results for sustainable technical solutions for digital archiving using knowledge management techniques based on semantic technologies. The workshop emphasizes interdisciplinarity and aims at an audience consisting of scientists and scholars from the digital library, digital archiving, multimedia technology and semantic web community, the information and library sciences, as well as, from the social sciences and (digital) humanities, in particular people working on the mentioned topics. We encourage end-users, practitioners and policy-makers from cultural heritage institutions to participate as well.
  9. Dietze, S.; Maynard, D.; Demidova, E.; Risse, T.; Stavrakas, Y.: Entity extraction and consolidation for social Web content preservation (2012) 0.00
    0.0021033147 = product of:
      0.008413259 = sum of:
        0.008413259 = weight(_text_:information in 470) [ClassicSimilarity], result of:
          0.008413259 = score(doc=470,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.13714671 = fieldWeight in 470, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=470)
      0.25 = coord(1/4)
    
    Abstract
    With the rapidly increasing pace at which Web content is evolving, particularly social media, preserving the Web and its evolution over time becomes an important challenge. Meaningful analysis of Web content lends itself to an entity-centric view to organise Web resources according to the information objects related to them. Therefore, the crucial challenge is to extract, detect and correlate entities from a vast number of heterogeneous Web resources where the nature and quality of the content may vary heavily. While a wealth of information extraction tools aid this process, we believe that, the consolidation of automatically extracted data has to be treated as an equally important step in order to ensure high quality and non-ambiguity of generated data. In this paper we present an approach which is based on an iterative cycle exploiting Web data for (1) targeted archiving/crawling of Web objects, (2) entity extraction, and detection, and (3) entity correlation. The long-term goal is to preserve Web content over time and allow its navigation and analysis based on well-formed structured RDF data about entities.
  10. Bozzato, L.; Braghin, S.; Trombetta, A.: ¬A method and guidelines for the cooperation of ontologies and relational databases in Semantic Web applications (2012) 0.00
    0.0021033147 = product of:
      0.008413259 = sum of:
        0.008413259 = weight(_text_:information in 475) [ClassicSimilarity], result of:
          0.008413259 = score(doc=475,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.13714671 = fieldWeight in 475, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=475)
      0.25 = coord(1/4)
    
    Abstract
    Ontologies are a well-affirmed way of representing complex structured information and they provide a sound conceptual foundation to Semantic Web technologies. On the other hand, a huge amount of information available on the web is stored in legacy relational databases. The issues raised by the collaboration between such worlds are well known and addressed by consolidated mapping languages. Nevertheless, to the best of our knowledge, a best practice for such cooperation is missing: in this work we thus present a method to guide the definition of cooperations between ontology-based and relational databases systems. Our method, mainly based on ideas from knowledge reuse and re-engineering, is aimed at the separation of data between database and ontology instances and at the definition of suitable mappings in both directions, taking advantage of the representation possibilities offered by both models. We present the steps of our method along with guidelines for their application. Finally, we propose an example of its deployment in the context of a large repository of bio-medical images we developed.
  11. Alexiev, V.: Implementing CIDOC CRM search based on fundamental relations and OWLIM rules (2012) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 467) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=467,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 467, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=467)
      0.25 = coord(1/4)
    
    Abstract
    The CIDOC CRM provides an ontology for describing entities, properties and relationships appearing in cultural heritage (CH) documentation, history and archeology. CRM promotes shared understanding by providing an extensible semantic framework that any CH information can be mapped to. CRM data is usually represented in semantic web format (RDF) and comprises complex graphs of nodes and properties. An important question is how a user can search through such complex graphs, since the number of possible combinations is staggering. One approach "compresses" the semantic network by mapping many CRM entity classes to a few "Fundamental Concepts" (FC), and mapping whole networks of CRM properties to fewer "Fundamental Relations" (FR). These FC and FRs serve as a "search index" over the CRM semantic web and allow the user to use a simpler query vocabulary. We describe an implementation of CRM FR Search based on OWLIM Rules, done as part of the ResearchSpace (RS) project. We describe the technical details, problems and difficulties encountered, benefits and disadvantages of using OWLIM rules, and preliminary performance results. We provide implementation experience that can be valuable for further implementation, definition and maintenance of CRM FRs.
  12. Wartena, C.; Sommer, M.: Automatic classification of scientific records using the German Subject Heading Authority File (SWD) (2012) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 472) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=472,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 472, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=472)
      0.25 = coord(1/4)
    
    Abstract
    The following paper deals with an automatic text classification method which does not require training documents. For this method the German Subject Heading Authority File (SWD), provided by the linked data service of the German National Library is used. Recently the SWD was enriched with notations of the Dewey Decimal Classification (DDC). In consequence it became possible to utilize the subject headings as textual representations for the notations of the DDC. Basically, we we derive the classification of a text from the classification of the words in the text given by the thesaurus. The method was tested by classifying 3826 OAI-Records from 7 different repositories. Mean reciprocal rank and recall were chosen as evaluation measure. Direct comparison to a machine learning method has shown that this method is definitely competitive. Thus we can conclude that the enriched version of the SWD provides high quality information with a broad coverage for classification of German scientific articles.
  13. Grassi, M.; Morbidoni, C.; Nucci, M.; Fonda, S.; Ledda, G.: Pundit: semantically structured annotations for Web contents and digital libraries (2012) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 473) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=473,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 473, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=473)
      0.25 = coord(1/4)
    
    Abstract
    This paper introduces Pundit: a novel semantic annotation tool that allows users to create structured data while annotating Web pages relying on stand-off mark-up techniques. Pundit provides support for different types of annotations, ranging from simple comments to semantic links to Web of data entities and fine granular cross-references and citations. In addition, it can be configured to include custom controlled vocabularies and has been designed to enable groups of users to share their annotations and collaboratively create structured knowledge. Pundit allows creating semantically typed relations among heterogeneous resources, both having different multimedia formats and belonging to different pages and domains. In this way, annotations can reinforce existing data connections or create new ones and augment original information generating new semantically structured aggregations of knowledge. These can later be exploited both by other users to better navigate DL and Web content, and by applications to improve data management.
  14. Bahls, D.; Scherp, G.; Tochtermann, K.; Hasselbring, W.: Towards a recommender system for statistical research data (2012) 0.00
    0.0014872681 = product of:
      0.0059490725 = sum of:
        0.0059490725 = weight(_text_:information in 474) [ClassicSimilarity], result of:
          0.0059490725 = score(doc=474,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.09697737 = fieldWeight in 474, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=474)
      0.25 = coord(1/4)
    
    Abstract
    To effectively promote the exchange of scientific data, retrieval services are required to suit the needs of the research community. A large amount of research in the field of economics is based on statistical data, which is often drawn from external sources like data agencies, statistical offices or affiated institutes. Since producing such data for a particular research question is expensive in time and money-if possible at all- research activities are often influenced by the availability of suitable data. Researchers choose or adjust their questions, so that the empirical foundation to support their results is given. As a consequence, researchers look out and poll for newly available data in all sorts of directions due to a lacking information infrastructure for this domain. This circumstance and a recent report from the High Level Expert Group on Scientific Data motivate recommendation and notification services for research data sets. In this paper, we elaborate on a case-based recommender system for statistical data, which allows for precise query specification. We discuss required similarity measures on the basis of cross-domain code lists and propose a system architecture. To address the problem of continuous polling, we elaborate on a notification service to inform researchers on newly avaible data sets based on their personal request.
  15. OWLED 2009; OWL: Experiences and Directions, Sixth International Workshop, Chantilly, Virginia, USA, 23-24 October 2009, Co-located with ISWC 2009. (2009) 0.00
    8.923609E-4 = product of:
      0.0035694437 = sum of:
        0.0035694437 = weight(_text_:information in 3391) [ClassicSimilarity], result of:
          0.0035694437 = score(doc=3391,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.058186423 = fieldWeight in 3391, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0234375 = fieldNorm(doc=3391)
      0.25 = coord(1/4)
    
    Content
    Short Papers * A Database Backend for OWL, Jörg Henss, Joachim Kleb and Stephan Grimm. * Unifying SysML and OWL, Henson Graves. * The OWLlink Protocol, Thorsten Liebig, Marko Luther and Olaf Noppens. * A Reasoning Broker Framework for OWL, Juergen Bock, Tuvshintur Tserendorj, Yongchun Xu, Jens Wissmann and Stephan Grimm. * Change Representation For OWL 2 Ontologies, Raul Palma, Peter Haase, Oscar Corcho and Asunción Gómez-Pérez. * Practical Aspects of Query Rewriting for OWL 2, Héctor Pérez-Urbina, Ian Horrocks and Boris Motik. * CSage: Use of a Configurable Semantically Attributed Graph Editor as Framework for Editing and Visualization, Lawrence Levin. * A Conformance Test Suite for the OWL 2 RL/RDF Rules Language and the OWL 2 RDF-Based Semantics, Michael Schneider and Kai Mainzer. * Improving the Data Quality of Relational Databases using OBDA and OWL 2 QL, Olivier Cure. * Temporal Classes and OWL, Natalya Keberle. * Using Ontologies for Medical Image Retrieval - An Experiment, Jasmin Opitz, Bijan Parsia and Ulrike Sattler. * Task Representation and Retrieval in an Ontology-Guided Modelling System, Yuan Ren, Jens Lemcke, Andreas Friesen, Tirdad Rahmani, Srdjan Zivkovic, Boris Gregorcic, Andreas Bartho, Yuting Zhao and Jeff Z. Pan. * A platform for reasoning with OWL-EL knowledge bases in a Peer-to-Peer environment, Alexander De Leon and Michel Dumontier. * Axiomé: a Tool for the Elicitation and Management of SWRL Rules, Saeed Hassanpour, Martin O'Connor and Amar Das. * SQWRL: A Query Language for OWL, Martin O'Connor and Amar Das. * Classifying ELH Ontologies In SQL Databases, Vincent Delaitre and Yevgeny Kazakov. * A Semantic Web Approach to Represent and Retrieve Information in a Corporate Memory, Ana B. Rios-Alvarado, R. Carolina Medina-Ramirez and Ricardo Marcelin-Jimenez. * Towards a Graphical Notation for OWL 2, Elisa Kendall, Roy Bell, Roger Burkhart, Mark Dutra and Evan Wallace.