Search (53 results, page 1 of 3)

  • × theme_ss:"Semantic Web"
  1. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.13
    0.13311341 = product of:
      0.1996701 = sum of:
        0.049917527 = product of:
          0.14975257 = sum of:
            0.14975257 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.14975257 = score(doc=701,freq=2.0), product of:
                0.39968264 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.047143444 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
        0.14975257 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.14975257 = score(doc=701,freq=2.0), product of:
            0.39968264 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.047143444 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
      0.6666667 = coord(2/3)
    
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  2. Zhitomirsky-Geffet, M.; Bar-Ilan, J.: Towards maximal unification of semantically diverse ontologies for controversial domains (2014) 0.04
    0.036702015 = product of:
      0.110106036 = sum of:
        0.110106036 = sum of:
          0.08455689 = weight(_text_:methodology in 1634) [ClassicSimilarity], result of:
            0.08455689 = score(doc=1634,freq=8.0), product of:
              0.21236731 = queryWeight, product of:
                4.504705 = idf(docFreq=1328, maxDocs=44218)
                0.047143444 = queryNorm
              0.3981634 = fieldWeight in 1634, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                4.504705 = idf(docFreq=1328, maxDocs=44218)
                0.03125 = fieldNorm(doc=1634)
          0.025549144 = weight(_text_:22 in 1634) [ClassicSimilarity], result of:
            0.025549144 = score(doc=1634,freq=2.0), product of:
              0.16508831 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.047143444 = queryNorm
              0.15476047 = fieldWeight in 1634, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1634)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - Ontologies are prone to wide semantic variability due to subjective points of view of their composers. The purpose of this paper is to propose a new approach for maximal unification of diverse ontologies for controversial domains by their relations. Design/methodology/approach - Effective matching or unification of multiple ontologies for a specific domain is crucial for the success of many semantic web applications, such as semantic information retrieval and organization, document tagging, summarization and search. To this end, numerous automatic and semi-automatic techniques were proposed in the past decade that attempt to identify similar entities, mostly classes, in diverse ontologies for similar domains. Apparently, matching individual entities cannot result in full integration of ontologies' semantics without matching their inter-relations with all other-related classes (and instances). However, semantic matching of ontological relations still constitutes a major research challenge. Therefore, in this paper the authors propose a new paradigm for assessment of maximal possible matching and unification of ontological relations. To this end, several unification rules for ontological relations were devised based on ontological reference rules, and lexical and textual entailment. These rules were semi-automatically implemented to extend a given ontology with semantically matching relations from another ontology for a similar domain. Then, the ontologies were unified through these similar pairs of relations. The authors observe that these rules can be also facilitated to reveal the contradictory relations in different ontologies. Findings - To assess the feasibility of the approach two experiments were conducted with different sets of multiple personal ontologies on controversial domains constructed by trained subjects. The results for about 50 distinct ontology pairs demonstrate a good potential of the methodology for increasing inter-ontology agreement. Furthermore, the authors show that the presented methodology can lead to a complete unification of multiple semantically heterogeneous ontologies. Research limitations/implications - This is a conceptual study that presents a new approach for semantic unification of ontologies by a devised set of rules along with the initial experimental evidence of its feasibility and effectiveness. However, this methodology has to be fully automatically implemented and tested on a larger dataset in future research. Practical implications - This result has implication for semantic search, since a richer ontology, comprised of multiple aspects and viewpoints of the domain of knowledge, enhances discoverability and improves search results. Originality/value - To the best of the knowledge, this is the first study to examine and assess the maximal level of semantic relation-based ontology unification.
    Date
    20. 1.2015 18:30:22
  3. Brunetti, J.M.; Roberto García, R.: User-centered design and evaluation of overview components for semantic data exploration (2014) 0.03
    0.028446635 = product of:
      0.085339904 = sum of:
        0.085339904 = sum of:
          0.059790757 = weight(_text_:methodology in 1626) [ClassicSimilarity], result of:
            0.059790757 = score(doc=1626,freq=4.0), product of:
              0.21236731 = queryWeight, product of:
                4.504705 = idf(docFreq=1328, maxDocs=44218)
                0.047143444 = queryNorm
              0.28154406 = fieldWeight in 1626, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                4.504705 = idf(docFreq=1328, maxDocs=44218)
                0.03125 = fieldNorm(doc=1626)
          0.025549144 = weight(_text_:22 in 1626) [ClassicSimilarity], result of:
            0.025549144 = score(doc=1626,freq=2.0), product of:
              0.16508831 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.047143444 = queryNorm
              0.15476047 = fieldWeight in 1626, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1626)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The growing volumes of semantic data available in the web result in the need for handling the information overload phenomenon. The potential of this amount of data is enormous but in most cases it is very difficult for users to visualize, explore and use this data, especially for lay-users without experience with Semantic Web technologies. The paper aims to discuss these issues. Design/methodology/approach - The Visual Information-Seeking Mantra "Overview first, zoom and filter, then details-on-demand" proposed by Shneiderman describes how data should be presented in different stages to achieve an effective exploration. The overview is the first user task when dealing with a data set. The objective is that the user is capable of getting an idea about the overall structure of the data set. Different information architecture (IA) components supporting the overview tasks have been developed, so they are automatically generated from semantic data, and evaluated with end-users. Findings - The chosen IA components are well known to web users, as they are present in most web pages: navigation bars, site maps and site indexes. The authors complement them with Treemaps, a visualization technique for displaying hierarchical data. These components have been developed following an iterative User-Centered Design methodology. Evaluations with end-users have shown that they get easily used to them despite the fact that they are generated automatically from structured data, without requiring knowledge about the underlying semantic technologies, and that the different overview components complement each other as they focus on different information search needs. Originality/value - Obtaining semantic data sets overviews cannot be easily done with the current semantic web browsers. Overviews become difficult to achieve with large heterogeneous data sets, which is typical in the Semantic Web, because traditional IA techniques do not easily scale to large data sets. There is little or no support to obtain overview information quickly and easily at the beginning of the exploration of a new data set. This can be a serious limitation when exploring a data set for the first time, especially for lay-users. The proposal is to reuse and adapt existing IA components to provide this overview to users and show that they can be generated automatically from the thesaurus and ontologies that structure semantic data while providing a comparable user experience to traditional web sites.
    Date
    20. 1.2015 18:30:22
  4. Boer, V. de; Wielemaker, J.; Gent, J. van; Hildebrand, M.; Isaac, A.; Ossenbruggen, J. van; Schreiber, G.: Supporting linked data production for cultural heritage institutes : the Amsterdam Museum case study (2012) 0.02
    0.015255922 = product of:
      0.045767765 = sum of:
        0.045767765 = product of:
          0.09153553 = sum of:
            0.09153553 = weight(_text_:methodology in 265) [ClassicSimilarity], result of:
              0.09153553 = score(doc=265,freq=6.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.43102458 = fieldWeight in 265, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=265)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Within the cultural heritage field, proprietary metadata and vocabularies are being transformed into public Linked Data. These efforts have mostly been at the level of large-scale aggregators such as Europeana where the original data is abstracted to a common format and schema. Although this approach ensures a level of consistency and interoperability, the richness of the original data is lost in the process. In this paper, we present a transparent and interactive methodology for ingesting, converting and linking cultural heritage metadata into Linked Data. The methodology is designed to maintain the richness and detail of the original metadata. We introduce the XMLRDF conversion tool and describe how it is integrated in the ClioPatria semantic web toolkit. The methodology and the tools have been validated by converting the Amsterdam Museum metadata to a Linked Data version. In this way, the Amsterdam Museum became the first 'small' cultural heritage institution with a node in the Linked Data cloud.
  5. Dextre Clarke, S.G.: Challenges and opportunities for KOS standards (2007) 0.01
    0.014903667 = product of:
      0.044711 = sum of:
        0.044711 = product of:
          0.089422 = sum of:
            0.089422 = weight(_text_:22 in 4643) [ClassicSimilarity], result of:
              0.089422 = score(doc=4643,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.5416616 = fieldWeight in 4643, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4643)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 9.2007 15:41:14
  6. Tennis, J.T.: Scheme versioning in the Semantic Web (2006) 0.01
    0.014092816 = product of:
      0.042278446 = sum of:
        0.042278446 = product of:
          0.08455689 = sum of:
            0.08455689 = weight(_text_:methodology in 4939) [ClassicSimilarity], result of:
              0.08455689 = score(doc=4939,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.3981634 = fieldWeight in 4939, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4939)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This paper describes a conceptual framework and methodology for managing scheme versioning for the Semantic Web. The first part of the paper introduces the concept of vocabulary encoding schemes, distinguished from metadata schemas, and discusses the characteristics of changes in schemes. The paper then presents a proposal to use a value record-similar to a term record in thesaurus management techniques-to manage scheme versioning challenges for the Semantic Web. The conclusion identifies future research directions.
  7. Sure, Y.; Erdmann, M.; Studer, R.: OntoEdit: collaborative engineering of ontologies (2004) 0.01
    0.014092816 = product of:
      0.042278446 = sum of:
        0.042278446 = product of:
          0.08455689 = sum of:
            0.08455689 = weight(_text_:methodology in 4405) [ClassicSimilarity], result of:
              0.08455689 = score(doc=4405,freq=8.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.3981634 = fieldWeight in 4405, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4405)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Developing ontologies is central to our vision of Semantic Web-based knowledge management. The methodology described in Chapter 3 guides the development of ontologies for different applications. However, because of the size of ontologies, their complexity, their formal underpinnings and the necessity to come towards a shared understanding within a group of people when defining an ontology, ontology construction is still far from being a well-understood process. Concerning the methodology, OntoEdit focuses on three of the main steps for ontology development (the methodology is described in Chapter 3), viz. the kick off, refinement, and evaluation. We describe the steps supported by OntoEdit and focus on collaborative aspects that occur during each of the step. First, all requirements of the envisaged ontology are collected during the kick off phase. Typically for ontology engineering, ontology engineers and domain experts are joined in a team that works together on a description of the domain and the goal of the ontology, design guidelines, available knowledge sources (e.g. re-usable ontologies and thesauri, etc.), potential users and use cases and applications supported by the ontology. The output of this phase is a semiformal description of the ontology. Second, during the refinement phase, the team extends the semi-formal description in several iterations and formalizes it in an appropriate representation language like RDF(S) or, more advanced, DAML1OIL. The output of this phase is a mature ontology (the 'target ontology'). Third, the target ontology needs to be evaluated according to the requirement specifications. Typically this phase serves as a proof for the usefulness of ontologies (and ontology-based applications) and may involve the engineering team as well as end users of the targeted application. The output of this phase is an evaluated ontology, ready for roll-out into a productive environment. Support for these collaborative development steps within the ontology development methodology is crucial in order to meet the conflicting needs for ease of use and construction of complex ontology structures. We now illustrate OntoEdit's support for each of the supported steps. The examples shown are taken from the Swiss Life case study on skills management (cf. Chapter 12).
  8. Broughton, V.: Automatic metadata generation : Digital resource description without human intervention (2007) 0.01
    0.012774572 = product of:
      0.038323715 = sum of:
        0.038323715 = product of:
          0.07664743 = sum of:
            0.07664743 = weight(_text_:22 in 6048) [ClassicSimilarity], result of:
              0.07664743 = score(doc=6048,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.46428138 = fieldWeight in 6048, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6048)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 9.2007 15:41:14
  9. Tudhope, D.: Knowledge Organization System Services : brief review of NKOS activities and possibility of KOS registries (2007) 0.01
    0.012774572 = product of:
      0.038323715 = sum of:
        0.038323715 = product of:
          0.07664743 = sum of:
            0.07664743 = weight(_text_:22 in 100) [ClassicSimilarity], result of:
              0.07664743 = score(doc=100,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.46428138 = fieldWeight in 100, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=100)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 9.2007 15:41:14
  10. Ghorbel, H.; Bahri, A.; Bouaziz, R.: Fuzzy ontologies building platform for Semantic Web : FOB platform (2012) 0.01
    0.012456408 = product of:
      0.03736922 = sum of:
        0.03736922 = product of:
          0.07473844 = sum of:
            0.07473844 = weight(_text_:methodology in 98) [ClassicSimilarity], result of:
              0.07473844 = score(doc=98,freq=4.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.35193008 = fieldWeight in 98, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=98)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The unstructured design of Web resources favors human comprehension, but makes difficult the automatic exploitation of the contents of these resources by machines. So, the Semantic Web aims at making the cooperation between human and machine possible, by giving any information a well defined meaning. The first weavings of the Semantic Web are already prepared. Machines become able to treat and understand the data that were accustomed to only visualization, by using ontologies constitute an essential element of the Semantic Web, as they serve as a form of knowledge representation, sharing, and reuse. However, the Web content is subject to imperfection, and crisp ontologies become less suitable to represent concepts with imprecise definitions. To overcome this problem, fuzzy ontologies constitute a promising research orientation. Indeed, the definition of fuzzy ontologies components constitutes an issue that needs to be well treated. It is necessary to have an appropriate methodology of building an operationalization of fuzzy ontological models. This chapter defines a fuzzy ontological model based on fuzzy description logic. This model uses a new approach for the formal description of fuzzy ontologies. This new methodology shows how all the basic components defined for fuzzy ontologies can be constructed.
  11. Pattuelli, C.; Rubinow, S.: ¬The knowledge organization of DBpedia : a case study (2013) 0.01
    0.012456408 = product of:
      0.03736922 = sum of:
        0.03736922 = product of:
          0.07473844 = sum of:
            0.07473844 = weight(_text_:methodology in 1776) [ClassicSimilarity], result of:
              0.07473844 = score(doc=1776,freq=4.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.35193008 = fieldWeight in 1776, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1776)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - This paper investigates the semantic structure underlying DBpedia, one of the largest and most heavily used datasets in the current Linked Open Data (LOD) landscape. The analysis attempts to shed light on this new type of knowledge organization tool. Design/methodology/approach - The research followed a case study methodology to analyze DBpedia using the domain of jazz as the application scenario. Findings - The study reveals an evolving knowledge organization tool where different descriptive and classification approaches are employed concurrently. The semantic constructs employed in the DBpedia knowledge base vary significantly in terms of their degree of formalization, stability, cohesiveness and consistency. As such, they challenge the tolerance threshold for data quality and the traditional notion of authority control. Research limitations/implications - The analysis is conducted on a limited portion of a large knowledge base. Initial findings provide a basis for further research and study. Practical implications - Revealing the knowledge organization underlying DBpedia increases the understanding of its power, its limitations and its implications for the new semantic context provided by LOD. Having an understanding of the range of entities and properties available enables LOD users to formulate queries with higher precision. Originality/value - This study is the first conducted from the perspective of the knowledge organization community.
  12. Fripp, D.: Using linked data to classify web documents (2010) 0.01
    0.012331214 = product of:
      0.03699364 = sum of:
        0.03699364 = product of:
          0.07398728 = sum of:
            0.07398728 = weight(_text_:methodology in 4172) [ClassicSimilarity], result of:
              0.07398728 = score(doc=4172,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.348393 = fieldWeight in 4172, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4172)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The purpose of this paper is to find a relationship between traditional faceted classification schemes and semantic web document annotators, particularly in the linked data environment. Design/methodology/approach - A consideration of the conceptual ideas behind faceted classification and linked data architecture is made. Analysis of selected web documents is performed using Calais' Semantic Proxy to support the considerations. Findings - Technical language aside, the principles of both approaches are very similar. Modern classification techniques have the potential to automatically generate metadata to drive more precise information recall by including a semantic layer. Originality/value - Linked data have not been explicitly considered in this context before in the published literature.
  13. Papadakis, I. et al.: Highlighting timely information in libraries through social and semantic Web technologies (2016) 0.01
    0.010645477 = product of:
      0.03193643 = sum of:
        0.03193643 = product of:
          0.06387286 = sum of:
            0.06387286 = weight(_text_:22 in 2090) [ClassicSimilarity], result of:
              0.06387286 = score(doc=2090,freq=2.0), product of:
                0.16508831 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047143444 = queryNorm
                0.38690117 = fieldWeight in 2090, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2090)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    Metadata and semantics research: 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings. Eds.: E. Garoufallou
  14. Macgregor, G.: Introduction to a special issue on digital libraries and the semantic web : context, applications and research (2008) 0.01
    0.010569612 = product of:
      0.031708833 = sum of:
        0.031708833 = product of:
          0.063417666 = sum of:
            0.063417666 = weight(_text_:methodology in 1904) [ClassicSimilarity], result of:
              0.063417666 = score(doc=1904,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.29862255 = fieldWeight in 1904, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1904)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The purpose of this article is to introduce the papers in the special issue which explores some of the potential, opportunities and challenges to be found in greater library and information science alignment with semantic web developments. Design/methodology/approach - The article is a general review of the papers in the issue. Findings - For many digital libraries or cultural institutions, the semantic web offers an opportunity to better expose valuable digital resources pertaining to research, culture or history, using common standards and technologies in a collaborative and "joined up" way. The papers in this issue "paint a rainbow", exploring the issues through elements of case studies, reviews research and conceptual expositions and viewpoints. Originality/value - The article emphasises how the practical implications of semantic web research or developments for digital libraries and repositories is important for LIS professionals.
  15. Joint, N.: ¬The practitioner librarian and the semantic web : ANTAEUS (2008) 0.01
    0.010569612 = product of:
      0.031708833 = sum of:
        0.031708833 = product of:
          0.063417666 = sum of:
            0.063417666 = weight(_text_:methodology in 2012) [ClassicSimilarity], result of:
              0.063417666 = score(doc=2012,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.29862255 = fieldWeight in 2012, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2012)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose -To describe and evoke the potential impact of semantic web systems at the level of library practice. Design/methodology/approach - A general outline of some of the broad issues associated with the semantic web, together with a brief, simple explanation of basic semantic web procedures with some examples of specific practical outcomes of semantic web development. Findings - That the semantic web is of central relevance to contemporary LIS practitioners, whose involvement in its development is necessary in order to determine what will be the true benefits of this form of information service innovation. Research limitations/implications - Since much of the initial discussion of this topic has been developmental and futuristic, applied practitioner-oriented research is required to ground these discussions in a firm bedrock of applications. Practical implications - semantic web technologies are of great practical relevance to areas of LIS practice such as digital repository development and open access services. Originality/value - The paper attempts to bridge the gap between the abstractions of theoretical writing in this area and the concerns of the working library professional.
  16. Towards the Semantic Web : ontology-driven knowledge management (2004) 0.01
    0.009153553 = product of:
      0.027460659 = sum of:
        0.027460659 = product of:
          0.054921318 = sum of:
            0.054921318 = weight(_text_:methodology in 4401) [ClassicSimilarity], result of:
              0.054921318 = score(doc=4401,freq=6.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.25861475 = fieldWeight in 4401, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=4401)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    With the current changes driven by the expansion of the World Wide Web, this book uses a different approach from other books on the market: it applies ontologies to electronically available information to improve the quality of knowledge management in large and distributed organizations. Ontologies are formal theories supporting knowledge sharing and reuse. They can be used to explicitly represent semantics of semi-structured information. These enable sophisticated automatic support for acquiring, maintaining and accessing information. Methodology and tools are developed for intelligent access to large volumes of semi-structured and textual information sources in intra- and extra-, and internet-based environments to employ the full power of ontologies in supporting knowledge management from the information client perspective and the information provider. The aim of the book is to support efficient and effective knowledge management and focuses on weakly-structured online information sources. It is aimed primarily at researchers in the area of knowledge management and information retrieval and will also be a useful reference for students in computer science at the postgraduate level and for business managers who are aiming to increase the corporations' information infrastructure. The Semantic Web is a very important initiative affecting the future of the WWW that is currently generating huge interest. The book covers several highly significant contributions to the semantic web research effort, including a new language for defining ontologies, several novel software tools and a coherent methodology for the application of the tools for business advantage. It also provides 3 case studies which give examples of the real benefits to be derived from the adoption of semantic-web based ontologies in "real world" situations. As such, the book is an excellent mixture of theory, tools and applications in an important area of WWW research. * Provides guidelines for introducing knowledge management concepts and tools into enterprises, to help knowledge providers present their knowledge efficiently and effectively. * Introduces an intelligent search tool that supports users in accessing information and a tool environment for maintenance, conversion and acquisition of information sources. * Discusses three large case studies which will help to develop the technology according to the actual needs of large and or virtual organisations and will provide a testbed for evaluating tools and methods. The book is aimed at people with at least a good understanding of existing WWW technology and some level of technical understanding of the underpinning technologies (XML/RDF). It will be of interest to graduate students, academic and industrial researchers in the field, and the many industrial personnel who are tracking WWW technology developments in order to understand the business implications. It could also be used to support undergraduate courses in the area but is not itself an introductory text.
    Content
    Inhalt: OIL and DAML + OIL: Ontology Languages for the Semantic Web (pages 11-31) / Dieter Fensel, Frank van Harmelen and Ian Horrocks A Methodology for Ontology-Based Knowledge Management (pages 33-46) / York Sure and Rudi Studer Ontology Management: Storing, Aligning and Maintaining Ontologies (pages 47-69) / Michel Klein, Ying Ding, Dieter Fensel and Borys Omelayenko Sesame: A Generic Architecture for Storing and Querying RDF and RDF Schema (pages 71-89) / Jeen Broekstra, Arjohn Kampman and Frank van Harmelen Generating Ontologies for the Semantic Web: OntoBuilder (pages 91-115) / R. H. P. Engels and T. Ch. Lech OntoEdit: Collaborative Engineering of Ontologies (pages 117-132) / York Sure, Michael Erdmann and Rudi Studer QuizRDF: Search Technology for the Semantic Web (pages 133-144) / John Davies, Richard Weeks and Uwe Krohn Spectacle (pages 145-159) / Christiaan Fluit, Herko ter Horst, Jos van der Meer, Marta Sabou and Peter Mika OntoShare: Evolving Ontologies in a Knowledge Sharing System (pages 161-177) / John Davies, Alistair Duke and Audrius Stonkus Ontology Middleware and Reasoning (pages 179-196) / Atanas Kiryakov, Kiril Simov and Damyan Ognyanov Ontology-Based Knowledge Management at Work: The Swiss Life Case Studies (pages 197-218) / Ulrich Reimer, Peter Brockhausen, Thorsten Lau and Jacqueline R. Reich Field Experimenting with Semantic Web Tools in a Virtual Organization (pages 219-244) / Victor Iosif, Peter Mika, Rikard Larsson and Hans Akkermans A Future Perspective: Exploiting Peer-To-Peer and the Semantic Web for Knowledge Management (pages 245-264) / Dieter Fensel, Steffen Staab, Rudi Studer, Frank van Harmelen and John Davies Conclusions: Ontology-driven Knowledge Management - Towards the Semantic Web? (pages 265-266) / John Davies, Dieter Fensel and Frank van Harmelen
  17. Prasad, A.R.D.; Madalli, D.P.: Faceted infrastructure for semantic digital libraries (2008) 0.01
    0.00880801 = product of:
      0.02642403 = sum of:
        0.02642403 = product of:
          0.05284806 = sum of:
            0.05284806 = weight(_text_:methodology in 1905) [ClassicSimilarity], result of:
              0.05284806 = score(doc=1905,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.24885213 = fieldWeight in 1905, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1905)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The paper aims to argue that digital library retrieval should be based on semantic representations and propose a semantic infrastructure for digital libraries. Design/methodology/approach - The approach taken is formal model based on subject representation for digital libraries. Findings - Search engines and search techniques have fallen short of user expectations as they do not give context based retrieval. Deploying semantic web technologies would lead to efficient and more precise representation of digital library content and hence better retrieval. Though digital libraries often have metadata of information resources which can be accessed through OAI-PMH, much remains to be accomplished in making digital libraries semantic web compliant. This paper presents a semantic infrastructure for digital libraries, that will go a long way in providing them and web based information services with products highly customised to users needs. Research limitations/implications - Here only a model for semantic infrastructure is proposed. This model is proposed after studying current user-centric, top-down models adopted in digital library service architectures. Originality/value - This paper gives a generic model for building semantic infrastructure for digital libraries. Faceted ontologies for digital libraries is just one approach. But the same may be adopted by groups working with different approaches in building ontologies to realise efficient retrieval in digital libraries.
  18. Sini, M.; Lauser, B.; Salokhe, G.; Keizer, J.; Katz, S.: ¬The AGROVOC concept server : rationale, goals and usage (2008) 0.01
    0.00880801 = product of:
      0.02642403 = sum of:
        0.02642403 = product of:
          0.05284806 = sum of:
            0.05284806 = weight(_text_:methodology in 1907) [ClassicSimilarity], result of:
              0.05284806 = score(doc=1907,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.24885213 = fieldWeight in 1907, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1907)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The main objective of the AGROVOC Concept Server (CS) is to create a collaborative reference platform and a "one-stop" shop for a pool of commonly used concepts related to agriculture, containing terms, definitions and relationships between terms in multiple languages derived from various sources. This paper aims to address the issues. Design/methodology/approach - The CS offers a centralised facility where the agricultural information management community can build and share agricultural knowledge in a collaborative environment. Findings - The advantages of the CS are its extensibility and modularity that provide the possibility to extend the type of information that can be stored in this system based on user/community needs. Research limitations/implications - Further investigation still needs to be done on the modularisation of the CS (i.e. the creation of separated ontologies that can still be connected, in order to have domain-related ontologies and to allow for better performance of the CS). Practical implications - The CS serves as starting point for the development of specific domain ontologies where multilinguality and the localised representation of knowledge are essential issues. Furthermore, it will offer additional services in order to expose the knowledge to be consumed by other applications. Originality/value - The CS Workbench provides the AGROVOC partners with the possibility to directly and collaboratively edit the AGROVOC CS. It thus provides the opportunity for direct and open "many-to-many" communication links between communities, avoiding decentralised communication between partners and duplication of effort. For the international community, it may allow users to manage, re-use or extend agriculture-related knowledge for better interoperability and for improved services.
  19. Krause, J.: Semantic heterogeneity : comparing new semantic web approaches with those of digital libraries (2008) 0.01
    0.00880801 = product of:
      0.02642403 = sum of:
        0.02642403 = product of:
          0.05284806 = sum of:
            0.05284806 = weight(_text_:methodology in 1908) [ClassicSimilarity], result of:
              0.05284806 = score(doc=1908,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.24885213 = fieldWeight in 1908, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1908)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - To demonstrate that newer developments in the semantic web community, particularly those based on ontologies (simple knowledge organization system and others) mitigate common arguments from the digital library (DL) community against participation in the Semantic web. Design/methodology/approach - The approach is a semantic web discussion focusing on the weak structure of the Web and the lack of consideration given to the semantic content during indexing. Findings - The points criticised by the semantic web and ontology approaches are the same as those of the DL "Shell model approach" from the mid-1990s, with emphasis on the centrality of its heterogeneity components (used, for example, in vascoda). The Shell model argument began with the "invisible web", necessitating the restructuring of DL approaches. The conclusion is that both approaches fit well together and that the Shell model, with its semantic heterogeneity components, can be reformulated on the semantic web basis. Practical implications - A reinterpretation of the DL approaches of semantic heterogeneity and adapting to standards and tools supported by the W3C should be the best solution. It is therefore recommended that - although most of the semantic web standards are not technologically refined for commercial applications at present - all individual DL developments should be checked for their adaptability to the W3C standards of the semantic web. Originality/value - A unique conceptual analysis of the parallel developments emanating from the digital library and semantic web communities.
  20. Mayr, P.; Mutschke, P.; Petras, V.: Reducing semantic complexity in distributed digital libraries : Treatment of term vagueness and document re-ranking (2008) 0.01
    0.00880801 = product of:
      0.02642403 = sum of:
        0.02642403 = product of:
          0.05284806 = sum of:
            0.05284806 = weight(_text_:methodology in 1909) [ClassicSimilarity], result of:
              0.05284806 = score(doc=1909,freq=2.0), product of:
                0.21236731 = queryWeight, product of:
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.047143444 = queryNorm
                0.24885213 = fieldWeight in 1909, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.504705 = idf(docFreq=1328, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1909)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The general science portal "vascoda" merges structured, high-quality information collections from more than 40 providers on the basis of search engine technology (FAST) and a concept which treats semantic heterogeneity between different controlled vocabularies. First experiences with the portal show some weaknesses of this approach which come out in most metadata-driven Digital Libraries (DLs) or subject specific portals. The purpose of the paper is to propose models to reduce the semantic complexity in heterogeneous DLs. The aim is to introduce value-added services (treatment of term vagueness and document re-ranking) that gain a certain quality in DLs if they are combined with heterogeneity components established in the project "Competence Center Modeling and Treatment of Semantic Heterogeneity". Design/methodology/approach - Two methods, which are derived from scientometrics and network analysis, will be implemented with the objective to re-rank result sets by the following structural properties: the ranking of the results by core journals (so-called Bradfordizing) and ranking by centrality of authors in co-authorship networks. Findings - The methods, which will be implemented, focus on the query and on the result side of a search and are designed to positively influence each other. Conceptually, they will improve the search quality and guarantee that the most relevant documents in result sets will be ranked higher. Originality/value - The central impact of the paper focuses on the integration of three structural value-adding methods, which aim at reducing the semantic complexity represented in distributed DLs at several stages in the information retrieval process: query construction, search and ranking and re-ranking.

Languages

  • e 47
  • d 6

Types

  • a 36
  • el 11
  • m 7
  • s 5
  • n 1
  • x 1
  • More… Less…