Search (4 results, page 1 of 1)

  • × theme_ss:"Semantic Web"
  • × theme_ss:"Semantische Interoperabilität"
  • × type_ss:"el"
  1. Carbonaro, A.; Santandrea, L.: ¬A general Semantic Web approach for data analysis on graduates statistics 0.01
    0.00831538 = product of:
      0.01663076 = sum of:
        0.01663076 = product of:
          0.049892277 = sum of:
            0.049892277 = weight(_text_:i in 5309) [ClassicSimilarity], result of:
              0.049892277 = score(doc=5309,freq=4.0), product of:
                0.16931784 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.044891298 = queryNorm
                0.29466638 = fieldWeight in 5309, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5309)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Abstract
    Currently, several datasets released in a Linked Open Data format are available at a national and international level, but the lack of shared strategies concerning the definition of concepts related to the statistical publishing community makes difficult a comparison among given facts starting from different data sources. In order to guarantee a shared representation framework for what concerns the dissemination of statistical concepts about graduates, we developed SW4AL, an ontology-based system for graduate's surveys domain. The developed system transforms low-level data into an enriched information model and is based on the AlmaLaurea surveys covering more than 90% of Italian graduates. SW4AL: i) semantically describes the different peculiarities of the graduates; ii) promotes the structured definition of the AlmaLaurea data and the following publication in the Linked Open Data context; iii) provides their reuse in the open data scope; iv) enables logical reasoning about knowledge representation. SW4AL establishes a common semantic for addressing the concept of graduate's surveys domain by proposing the creation of a SPARQL endpoint and a Web based interface for the query and the visualization of the structured data.
    Location
    I
  2. Heflin, J.; Hendler, J.: Semantic interoperability on the Web (2000) 0.01
    0.007095843 = product of:
      0.014191686 = sum of:
        0.014191686 = product of:
          0.042575058 = sum of:
            0.042575058 = weight(_text_:22 in 759) [ClassicSimilarity], result of:
              0.042575058 = score(doc=759,freq=2.0), product of:
                0.15720168 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.044891298 = queryNorm
                0.2708308 = fieldWeight in 759, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=759)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Date
    11. 5.2013 19:22:18
  3. Proceedings of the 2nd International Workshop on Evaluation of Ontology-based Tools (2004) 0.01
    0.0058798613 = product of:
      0.011759723 = sum of:
        0.011759723 = product of:
          0.035279166 = sum of:
            0.035279166 = weight(_text_:i in 3152) [ClassicSimilarity], result of:
              0.035279166 = score(doc=3152,freq=2.0), product of:
                0.16931784 = queryWeight, product of:
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.044891298 = queryNorm
                0.20836058 = fieldWeight in 3152, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7717297 = idf(docFreq=2765, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3152)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Content
    Table of Contents Part I: Accepted Papers Christoph Tempich and Raphael Volz: Towards a benchmark for Semantic Web reasoners - an analysis of the DAML ontology library M. Carmen Suarez-Figueroa and Asuncion Gomez-Perez: Results of Taxonomic Evaluation of RDF(S) and DAML+OIL ontologies using RDF(S) and DAML+OIL Validation Tools and Ontology Platforms import services Volker Haarslev and Ralf Möller: Racer: A Core Inference Engine for the Semantic Web Mikhail Kazakov and Habib Abdulrab: DL-workbench: a metamodeling approach to ontology manipulation Thorsten Liebig and Olaf Noppens: OntoTrack: Fast Browsing and Easy Editing of Large Ontologie Frederic Fürst, Michel Leclere, and Francky Trichet: TooCoM : a Tool to Operationalize an Ontology with the Conceptual Graph Model Naoki Sugiura, Masaki Kurematsu, Naoki Fukuta, Noriaki Izumi, and Takahira Yamaguchi: A domain ontology engineering tool with general ontologies and text corpus Howard Goldberg, Alfredo Morales, David MacMillan, and Matthew Quinlan: An Ontology-Driven Application to Improve the Prescription of Educational Resources to Parents of Premature Infants Part II: Experiment Contributions Domain natural language description for the experiment Raphael Troncy, Antoine Isaac, and Veronique Malaise: Using XSLT for Interoperability: DOE and The Travelling Domain Experiment Christian Fillies: SemTalk EON2003 Semantic Web Export / Import Interface Test Óscar Corcho, Asunción Gómez-Pérez, Danilo José Guerrero-Rodríguez, David Pérez-Rey, Alberto Ruiz-Cristina, Teresa Sastre-Toral, M. Carmen Suárez-Figueroa: Evaluation experiment of ontology tools' interoperability with the WebODE ontology engineering workbench Holger Knublauch: Case Study: Using Protege to Convert the Travel Ontology to UML and OWL Franz Calvo and John Gennari: Interoperability of Protege 2.0 beta and OilEd 3.5 in the Domain Knowledge of Osteoporosis
  4. Isaac, A.: Aligning thesauri for an integrated access to Cultural Heritage Resources (2007) 0.00
    0.0048684133 = product of:
      0.009736827 = sum of:
        0.009736827 = product of:
          0.029210478 = sum of:
            0.029210478 = weight(_text_:c in 553) [ClassicSimilarity], result of:
              0.029210478 = score(doc=553,freq=4.0), product of:
                0.15484828 = queryWeight, product of:
                  3.4494052 = idf(docFreq=3817, maxDocs=44218)
                  0.044891298 = queryNorm
                0.18863934 = fieldWeight in 553, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.4494052 = idf(docFreq=3817, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=553)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Abstract
    Currently, a number of efforts are being carried out to integrate collections from different institutions and containing heterogeneous material. Examples of such projects are The European Library [1] and the Memory of the Netherlands [2]. A crucial point for the success of these is the availability to provide a unified access on top of the different collections, e.g. using one single vocabulary for querying or browsing the objects they contain. This is made difficult by the fact that the objects from different collections are often described using different vocabularies - thesauri, classification schemes - and are therefore not interoperable at the semantic level. To solve this problem, one can turn to semantic links - mappings - between the elements of the different vocabularies. If one knows that a concept C from a vocabulary V is semantically equivalent to a concept to a concept D from vocabulary W, then an appropriate search engine can return all the objects that were indexed against D for a query for objects described using C. We thus have an access to other collections, using a single one vocabulary. This is however an ideal situation, and hard alignment work is required to reach it. Several projects in the past have tried to implement such a solution, like MACS [3] and Renardus [4]. They have demonstrated very interesting results, but also highlighted the difficulty of aligning manually all the different vocabularies involved in practical cases, which sometimes contain hundreds of thousands of concepts. To alleviate this problem, a number of tools have been proposed in order to provide with candidate mappings between two input vocabularies, making alignment a (semi-) automatic task. Recently, the Semantic Web community has produced a lot of these alignment tools'. Several techniques are found, depending on the material they exploit: labels of concepts, structure of vocabularies, collection objects and external knowledge sources. Throughout our presentation, we will present a concrete heterogeneity case where alignment techniques have been applied to build a (pilot) browser, developed in the context of the STITCH project [5]. This browser enables a unified access to two collections of illuminated manuscripts, using the description vocabulary used in the first collection, Mandragore [6], or the one used by the second, Iconclass [7]. In our talk, we will also make the point for using unified representations the vocabulary semantic and lexical information. Additionally to ease the use of the alignment tools that have these vocabularies as input, turning to a standard representation format helps designing applications that are more generic, like the browser we demonstrate. We give pointers to SKOS [8], an open and web-enabled format currently developed by the Semantic Web community.