Search (448 results, page 22 of 23)

  • × theme_ss:"Wissensrepräsentation"
  1. OWL Web Ontology Language Guide (2004) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 4687) [ClassicSimilarity], result of:
          0.008582841 = score(doc=4687,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 4687, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4687)
      0.25 = coord(1/4)
    
    Abstract
    The World Wide Web as it is currently constituted resembles a poorly mapped geography. Our insight into the documents and capabilities available are based on keyword searches, abetted by clever use of document connectivity and usage patterns. The sheer mass of this data is unmanageable without powerful tool support. In order to map this terrain more precisely, computational agents require machine-readable descriptions of the content and capabilities of Web accessible resources. These descriptions must be in addition to the human-readable versions of that information. The OWL Web Ontology Language is intended to provide a language that can be used to describe the classes and relations between them that are inherent in Web documents and applications. This document demonstrates the use of the OWL language to - formalize a domain by defining classes and properties of those classes, - define individuals and assert properties about them, and - reason about these classes and individuals to the degree permitted by the formal semantics of the OWL language. The sections are organized to present an incremental definition of a set of classes, properties and individuals, beginning with the fundamentals and proceeding to more complex language components.
  2. ¬The Semantic Web - ISWC 2010 : 9th International Semantic Web Conference, ISWC 2010, Shanghai, China, November 7-11, 2010, Revised Selected Papers, Part 2. (2010) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 4706) [ClassicSimilarity], result of:
          0.008582841 = score(doc=4706,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 4706, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4706)
      0.25 = coord(1/4)
    
    Abstract
    The two-volume set LNCS 6496 and 6497 constitutes the refereed proceedings of the 9th International Semantic Web Conference, ISWC 2010, held in Shanghai, China, during November 7-11, 2010. Part I contains 51 papers out of 578 submissions to the research track. Part II contains 18 papers out of 66 submissions to the semantic Web in-use track, 6 papers out of 26 submissions to the doctoral consortium track, and also 4 invited talks. Each submitted paper were carefully reviewed. The International Semantic Web Conferences (ISWC) constitute the major international venue where the latest research results and technical innovations on all aspects of the Semantic Web are presented. ISWC brings together researchers, practitioners, and users from the areas of artificial intelligence, databases, social networks, distributed computing, Web engineering, information systems, natural language processing, soft computing, and human computer interaction to discuss the major challenges and proposed solutions, the success stories and failures, as well the visions that can advance research and drive innovation in the Semantic Web.
  3. Cregan, A.: ¬An OWL DL construction for the ISO Topic Map Data Model (2005) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 4718) [ClassicSimilarity], result of:
          0.008582841 = score(doc=4718,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 4718, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4718)
      0.25 = coord(1/4)
    
    Abstract
    Both Topic Maps and the W3C Semantic Web technologies are meta-level semantic maps describing relationships between information resources. Previous attempts at interoperability between XTM Topic Maps and RDF have proved problematic. The ISO's drafting of an explicit Topic Map Data Model [TMDM 05] combined with the advent of the W3C's XML and RDFbased Description Logic-equivalent Web Ontology Language [OWLDL 04] now provides the means for the construction of an unambiguous semantic model to represent Topic Maps, in a form that is equivalent to a Description Logic representation. This paper describes the construction of the proposed TMDM ISO Topic Map Standard in OWL DL (Description Logic equivalent) form. The construction is claimed to exactly match the features of the proposed TMDM. The intention is that the topic map constructs described herein, once officially published on the world-wide web, may be used by Topic Map authors to construct their Topic Maps in OWL DL. The advantage of OWL DL Topic Map construction over XTM, the existing XML-based DTD standard, is that OWL DL allows many constraints to be explicitly stated. OWL DL's suite of tools, although currently still somewhat immature, will provide the means for both querying and enforcing constraints. This goes a long way towards fulfilling the requirements for a Topic Map Query Language (TMQL) and Constraint Language (TMCL), which the Topic Map Community may choose to expend effort on extending. Additionally, OWL DL has a clearly defined formal semantics (Description Logic ref)
  4. Iorio, A. di; Peroni, S.; Vitali, F.: ¬A Semantic Web approach to everyday overlapping markup (2011) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 4749) [ClassicSimilarity], result of:
          0.008582841 = score(doc=4749,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 4749, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4749)
      0.25 = coord(1/4)
    
    Source
    Journal of the American Society for Information Science and Technology. 62(2011) no.9, S.1696-1716
  5. Allocca, C.; Aquin, M.d'; Motta, E.: Impact of using relationships between ontologies to enhance the ontology search results (2012) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 264) [ClassicSimilarity], result of:
          0.008582841 = score(doc=264,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 264, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=264)
      0.25 = coord(1/4)
    
    Abstract
    Using semantic web search engines, such as Watson, Swoogle or Sindice, to find ontologies is a complex exploratory activity. It generally requires formulating multiple queries, browsing pages of results, and assessing the returned ontologies against each other to obtain a relevant and adequate subset of ontologies for the intended use. Our hypothesis is that at least some of the difficulties related to searching ontologies stem from the lack of structure in the search results, where ontologies that are implicitly related to each other are presented as disconnected and shown on different result pages. In earlier publications we presented a software framework, Kannel, which is able to automatically detect and make explicit relationships between ontologies in large ontology repositories. In this paper, we present a study that compares the use of the Watson ontology search engine with an extension,Watson+Kannel, which provides information regarding the various relationships occurring between the result ontologies. We evaluate Watson+Kannel by demonstrating through various indicators that explicit relationships between ontologies improve users' efficiency in ontology search, thus validating our hypothesis.
  6. Román, J.H.; Hulin, K.J.; Collins, L.M.; Powell, J.E.: Entity disambiguation using semantic networks (2012) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 461) [ClassicSimilarity], result of:
          0.008582841 = score(doc=461,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 461, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=461)
      0.25 = coord(1/4)
    
    Source
    Journal of the American Society for Information Science and Technology. 63(2012) no.10, S.2087-2099
  7. Boteram, F.: Typisierung semantischer Relationen in integrierten Systemen der Wissensorganisation (2013) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 919) [ClassicSimilarity], result of:
          0.008582841 = score(doc=919,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 919, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=919)
      0.25 = coord(1/4)
    
    Abstract
    Die, differenzierte Typisierung semantischer Relationen hinsichtlich ihrer bedeutungstragenden inhaltlichen und formallogischen Eigenschaften in Systemen der Wissensorganisation ist eine Voraussetzung für leistungsstarke und benutzerfreundliche Modelle des information Retrieval und der Wissensexploration. Systeme, die mehrere Dokumentationssprachen miteinander verknüpfen und funktional integrieren, erfordern besondere Ansätze für die Typisierung der verwendeten oder benötigten Relationen. Aufbauend auf vorangegangenen Überlegungen zu Modellen der semantischen Interoperabilität in verteilten Systemen, welche durch ein zentrales Kernsystem miteinander verbunden und so in den übergeordneten Funktionszusammenhang der Wissensorganisation gestellt werden, werden differenzierte und funktionale Strategien zur Typisierung und stratifizierten Definition der unterschiedlichen Relationen in diesem System entwickelt. Um die von fortschrittlichen Retrievalparadigmen erforderten Funktionalitäten im Kontext vernetzter Systeme zur Wissensorganisation unterstützen zu können, werden die formallogischen, typologischen und strukturellen Eigenschaften sowie der eigentliche semantische Gehalt aller Relationstypen definiert, die zur Darstellung von Begriffsbeziehungen verwendet werden. Um die Vielzahl unterschiedlicher aber im Funktionszusammenhang des Gesamtsystems auf einander bezogenen Relationstypen präzise und effizient ordnen zu können, wird eine mehrfach gegliederte Struktur benötigt, welche die angestrebten Inventare in einer Ear den Nutzer übersichtlichen und intuitiv handhabbaren Form präsentieren und somit für eine Verwendung in explorativen Systemen vorhalten kann.
  8. Solskinnsbakk, G.; Gulla, J.A.; Haderlein, V.; Myrseth, P.; Cerrato, O.: Quality of hierarchies in ontologies and folksonomies (2012) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 1034) [ClassicSimilarity], result of:
          0.008582841 = score(doc=1034,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 1034, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1034)
      0.25 = coord(1/4)
    
    Abstract
    Ontologies have been a hot research topic for the recent decade and have been used for many applications such as information integration, semantic search, knowledge management, etc. Manual engineering of ontologies is a costly process and automatic ontology engineering lacks in precision. Folksonomies have recently emerged as another hot research topic and several research efforts have been made to extract lightweight ontologies automatically from folksonomy data. Due to the high cost of manual ontology engineering and the lack of precision in automatic ontology engineering it is important that we are able to evaluate the structure of the ontology. Detection of problems with the suggested ontology at an early stage can, especially for manually engineered ontologies, be cost saving. In this paper we present an approach to evaluate the quality of hierarchical relations in ontologies and folksonomy based structures. The approach is based on constructing shallow semantic representations of the ontology concepts and folksonomy tags. We specify four hypotheses regarding the semantic representations and different quality aspects of the hierarchical relations and perform an evaluation on two different data sets. The results of the evaluation confirm our hypotheses.
  9. Kless, D.; Milton, S.; Kazmierczak, E.; Lindenthal, J.: Thesaurus and ontology structure : formal and pragmatic differences and similarities (2015) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 2036) [ClassicSimilarity], result of:
          0.008582841 = score(doc=2036,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 2036, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2036)
      0.25 = coord(1/4)
    
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.7, S.1348-1366
  10. Cao, N.; Sun, J.; Lin, Y.-R.; Gotz, D.; Liu, S.; Qu, H.: FacetAtlas : Multifaceted visualization for rich text corpora (2010) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 3366) [ClassicSimilarity], result of:
          0.008582841 = score(doc=3366,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 3366, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3366)
      0.25 = coord(1/4)
    
    Abstract
    Documents in rich text corpora usually contain multiple facets of information. For example, an article about a specific disease often consists of different facets such as symptom, treatment, cause, diagnosis, prognosis, and prevention. Thus, documents may have different relations based on different facets. Powerful search tools have been developed to help users locate lists of individual documents that are most related to specific keywords. However, there is a lack of effective analysis tools that reveal the multifaceted relations of documents within or cross the document clusters. In this paper, we present FacetAtlas, a multifaceted visualization technique for visually analyzing rich text corpora. FacetAtlas combines search technology with advanced visual analytical tools to convey both global and local patterns simultaneously. We describe several unique aspects of FacetAtlas, including (1) node cliques and multifaceted edges, (2) an optimized density map, and (3) automated opacity pattern enhancement for highlighting visual patterns, (4) interactive context switch between facets. In addition, we demonstrate the power of FacetAtlas through a case study that targets patient education in the health care domain. Our evaluation shows the benefits of this work, especially in support of complex multifaceted data analysis.
  11. Mainzer, K.: ¬The emergence of self-conscious systems : from symbolic AI to embodied robotics (2014) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 3398) [ClassicSimilarity], result of:
          0.008582841 = score(doc=3398,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 3398, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3398)
      0.25 = coord(1/4)
    
    Source
    Philosophy, computing and information science. Eds.: R. Hagengruber u. U.V. Riss
  12. Jansen, L.: Four rules for classifying social entities (2014) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 3409) [ClassicSimilarity], result of:
          0.008582841 = score(doc=3409,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 3409, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3409)
      0.25 = coord(1/4)
    
    Source
    Philosophy, computing and information science. Eds.: R. Hagengruber u. U.V. Riss
  13. Wen, B.; Horlings, E.; Zouwen, M. van der; Besselaar, P. van den: Mapping science through bibliometric triangulation : an experimental approach applied to water research (2017) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 3437) [ClassicSimilarity], result of:
          0.008582841 = score(doc=3437,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 3437, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3437)
      0.25 = coord(1/4)
    
    Source
    Journal of the Association for Information Science and Technology. 68(2017) no.3, S.724-738
  14. Zhitomirsky-Geffet, M.; Erez, E.S.; Bar-Ilan, J.: Toward multiviewpoint ontology construction by collaboration of non-experts and crowdsourcing : the case of the effect of diet on health (2017) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 3439) [ClassicSimilarity], result of:
          0.008582841 = score(doc=3439,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 3439, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3439)
      0.25 = coord(1/4)
    
    Source
    Journal of the Association for Information Science and Technology. 68(2017) no.3, S.681-694
  15. Dietiker, S.: Cognitive Map einer Bibliothek : eine Überprüfung der Methodentauglichkeit im Bereich Bibliothekswissenschaft - am Beispiel der Kantonsbibliothek Graubünden (2016) 0.00
    0.0021457102 = product of:
      0.008582841 = sum of:
        0.008582841 = weight(_text_:information in 4570) [ClassicSimilarity], result of:
          0.008582841 = score(doc=4570,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 4570, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4570)
      0.25 = coord(1/4)
    
    Content
    Diese Publikation entstand im Rahmen einer Thesis zum Bachelor of Science FHO in Information Science. Vgl. unter: http://www.htwchur.ch/uploads/media/CSI_78_Dietiker.pdf.
  16. Semantic Media Wiki : Autoren sollen Wiki-Inhalte erschließen (2006) 0.00
    0.002124145 = product of:
      0.00849658 = sum of:
        0.00849658 = weight(_text_:information in 6027) [ClassicSimilarity], result of:
          0.00849658 = score(doc=6027,freq=4.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.0960027 = fieldWeight in 6027, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=6027)
      0.25 = coord(1/4)
    
    Content
    Aus den so festgelegten Beziehungen zwischen den verlinkten Begriffen sollen Computer automatisch sinnvolle Antworten auf komplexere Anfragen generieren können; z.B. eine Liste erzeugen, in der alle Länder von Afrika aufgeführt sind. Die Ländernamen führen als Link zurück zu dem Eintrag, in dem sie stehen - dem Artikel zum Land, für das man sich interessiert. Aus informationswissenschaftlicher Sicht ist das Informationsergebnis, das die neue Technologie produziert, relativ simpel. Aus sozialwissenschaftlicher Sicht steckt darin aber ein riesiges Potential zur Verbesserung der Bereitstellung von enzyklopädischer Information und Wissen für Menschen auf der ganzen Welt. Spannend ist auch die durch Semantic MediaWiki gegebene Möglichkeit der automatischen Zusammenführung von Informationen, die in den verschiedenen Wiki-Einträgen verteilt sind, bei einer hohen Konsistenz der Ergebnisse. Durch die feststehenden Beziehungen zwischen den Links enthält die automatisch erzeugte Liste nach Angaben der Karlsruher Forscher immer die gleichen Daten, egal, von welcher Seite aus man sie abruft. Die Suchmaschine holt sich die Bevölkerungszahl von Ägypten immer vom festgelegten Ägypten-Eintrag, so dass keine unterschiedlichen Zahlen in der Wiki-Landschaft kursieren können. Ein mit Semantic MediaWiki erstellter Testeintrag zu Deutschland kann unter http://ontoworld.org/index.php/Germany eingesehen werden. Die Faktenbox im unteren Teil des Eintrags zeigt an, was der "Eintrag" der Suchmaschine an Wissen über Deutschland anbieten kann. Diese Ergebnisse werden auch in dem Datenbeschreibungsstandard RDF angeboten. Mehr als das, was in der Faktenbox steht, kann der Eintrag nicht an die Suchmaschine abgeben."
    Source
    Information - Wissenschaft und Praxis. 57(2006) H.6/7, S.372
  17. Panyr, J.: Thesauri, Semantische Netze, Frames, Topic Maps, Taxonomien, Ontologien - begriffliche Verwirrung oder konzeptionelle Vielfalt? (2006) 0.00
    0.0017165683 = product of:
      0.006866273 = sum of:
        0.006866273 = weight(_text_:information in 6058) [ClassicSimilarity], result of:
          0.006866273 = score(doc=6058,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.0775819 = fieldWeight in 6058, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=6058)
      0.25 = coord(1/4)
    
    Source
    Information und Sprache: Beiträge zu Informationswissenschaft, Computerlinguistik, Bibliothekswesen und verwandten Fächern. Festschrift für Harald H. Zimmermann. Herausgegeben von Ilse Harms, Heinz-Dirk Luckhardt und Hans W. Giessen
  18. Botana Varela, J.: Unscharfe Wissensrepräsentationen bei der Implementation des Semantic Web (2004) 0.00
    0.0017165683 = product of:
      0.006866273 = sum of:
        0.006866273 = weight(_text_:information in 3359) [ClassicSimilarity], result of:
          0.006866273 = score(doc=3359,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.0775819 = fieldWeight in 3359, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=3359)
      0.25 = coord(1/4)
    
    Abstract
    In der vorliegenden Arbeit soll einen Ansatz zur Implementation einer Wissensrepräsentation mit den in Abschnitt 1.1. skizzierten Eigenschaften und dem Semantic Web als Anwendungsbereich vorgestellt werden. Die Arbeit ist im Wesentlichen in zwei Bereiche gegliedert: dem Untersuchungsbereich (Kapitel 2-5), in dem ich die in Abschnitt 1.1. eingeführte Terminologie definiert und ein umfassender Überblick über die zugrundeliegenden Konzepte gegeben werden soll, und dem Implementationsbereich (Kapitel 6), in dem aufbauend auf dem im Untersuchungsbereich erarbeiteten Wissen einen semantischen Suchdienst entwickeln werden soll. In Kapitel 2 soll zunächst das Konzept der semantischen Interpretation erläutert und in diesem Kontext hauptsächlich zwischen Daten, Information und Wissen unterschieden werden. In Kapitel 3 soll Wissensrepräsentation aus einer kognitiven Perspektive betrachtet und in diesem Zusammenhang das Konzept der Unschärfe beschrieben werden. In Kapitel 4 sollen sowohl aus historischer als auch aktueller Sicht die Ansätze zur Wissensrepräsentation und -auffindung beschrieben und in diesem Zusammenhang das Konzept der Unschärfe diskutiert werden. In Kapitel 5 sollen die aktuell im WWW eingesetzten Modelle und deren Einschränkungen erläutert werden. Anschließend sollen im Kontext der Entscheidungsfindung die Anforderungen beschrieben werden, die das WWW an eine adäquate Wissensrepräsentation stellt, und anhand der Technologien des Semantic Web die Repräsentationsparadigmen erläutert werden, die diese Anforderungen erfüllen. Schließlich soll das Topic Map-Paradigma erläutert werden. In Kapitel 6 soll aufbauend auf die im Untersuchtungsbereich gewonnenen Erkenntnisse ein Prototyp entwickelt werden. Dieser besteht im Wesentlichen aus Softwarewerkzeugen, die das automatisierte und computergestützte Extrahieren von Informationen, das unscharfe Modellieren, sowie das Auffinden von Wissen unterstützen. Die Implementation der Werkzeuge erfolgt in der Programmiersprache Java, und zur unscharfen Wissensrepräsentation werden Topic Maps eingesetzt. Die Implementation wird dabei schrittweise vorgestellt. Schließlich soll der Prototyp evaluiert und ein Ausblick auf zukünftige Erweiterungsmöglichkeiten gegeben werden. Und schließlich soll in Kapitel 7 eine Synthese formuliert werden.
  19. Veltman, K.H.: Towards a Semantic Web for culture 0.00
    0.0017165683 = product of:
      0.006866273 = sum of:
        0.006866273 = weight(_text_:information in 4040) [ClassicSimilarity], result of:
          0.006866273 = score(doc=4040,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.0775819 = fieldWeight in 4040, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=4040)
      0.25 = coord(1/4)
    
    Source
    Journal of digital information. 4(2004), no.4
  20. Reimer, U.; Brockhausen, P.; Lau, T.; Reich, J.R.: Ontology-based knowledge management at work : the Swiss life case studies (2004) 0.00
    0.0017165683 = product of:
      0.006866273 = sum of:
        0.006866273 = weight(_text_:information in 4411) [ClassicSimilarity], result of:
          0.006866273 = score(doc=4411,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.0775819 = fieldWeight in 4411, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=4411)
      0.25 = coord(1/4)
    
    Abstract
    This chapter describes two case studies conducted by the Swiss Life insurance group with the objective of proving the practical applicability and superiority of ontology-based knowledge management over classical approaches based on text retrieval technologies. The first case study in the domain of skills management uses manually constructed ontologies about skills, job functions and education. The purpose of the system is to give support for finding employees with certain skills. The ontologies are used to ensure that the user description of skills and the machine-held index of skills and people use the same vocabulary. The use of a shared vocabulary increases the performance of such a system significantly. The second case study aims at improving content-oriented access to passages of a 1000 page document about the International Accounting Standard on the corporate intranet. To this end, an ontology was automatically extracted from the document. It can be used to reformulate queries that turned out not to deliver the intended results. Since the ontology was automatically built, it is of a rather simple structure, consisting of weighted semantic associations between the relevant concepts in the document. We therefore call it a 'lightweight ontology'. The two case studies cover quite different aspects of using ontologies in knowledge management applications. Whereas in the second case study an ontology was automatically derived from a search space to improve information retrieval, in the first skills management case study the ontology itself introduces a structured search space. In one case study we gathered experience in building an ontology manually, while the challenge of the other case study was automatic ontology creation. A number of the novel Semantic Web-based tools described elsewhere in this book were used to build the two systems and both case studies described have led to projects to deploy live systems within Swiss Life.

Authors

Years

Languages

  • e 372
  • d 65
  • pt 4
  • f 1
  • sp 1
  • More… Less…

Types

  • a 328
  • el 100
  • m 32
  • x 25
  • s 13
  • n 9
  • r 6
  • p 3
  • A 1
  • EL 1
  • More… Less…

Subjects

Classifications