Search (306 results, page 1 of 16)

  • × theme_ss:"Wissensrepräsentation"
  1. Innovations and advanced techniques in systems, computing sciences and software engineering (2008) 0.08
    0.07983807 = product of:
      0.31935227 = sum of:
        0.07119748 = weight(_text_:230 in 4319) [ClassicSimilarity], result of:
          0.07119748 = score(doc=4319,freq=4.0), product of:
            0.13547163 = queryWeight, product of:
              6.727074 = idf(docFreq=143, maxDocs=44218)
              0.02013827 = queryNorm
            0.5255527 = fieldWeight in 4319, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              6.727074 = idf(docFreq=143, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4319)
        0.060652044 = weight(_text_:software in 4319) [ClassicSimilarity], result of:
          0.060652044 = score(doc=4319,freq=24.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.75917953 = fieldWeight in 4319, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4319)
        0.060652044 = weight(_text_:software in 4319) [ClassicSimilarity], result of:
          0.060652044 = score(doc=4319,freq=24.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.75917953 = fieldWeight in 4319, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4319)
        0.060652044 = weight(_text_:software in 4319) [ClassicSimilarity], result of:
          0.060652044 = score(doc=4319,freq=24.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.75917953 = fieldWeight in 4319, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4319)
        0.06619864 = product of:
          0.13239728 = sum of:
            0.13239728 = weight(_text_:engineering in 4319) [ClassicSimilarity], result of:
              0.13239728 = score(doc=4319,freq=34.0), product of:
                0.10819342 = queryWeight, product of:
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.02013827 = queryNorm
                1.2237091 = fieldWeight in 4319, product of:
                  5.8309517 = tf(freq=34.0), with freq of:
                    34.0 = termFreq=34.0
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4319)
          0.5 = coord(1/2)
      0.25 = coord(5/20)
    
    Abstract
    Innovations and Advanced Techniques in Systems, Computing Sciences and Software Engineering includes a set of rigorously reviewed world-class manuscripts addressing and detailing state-of-the-art research projects in the areas of Computer Science, Software Engineering, Computer Engineering, and Systems Engineering and Sciences. Innovations and Advanced Techniques in Systems, Computing Sciences and Software Engineering includes selected papers form the conference proceedings of the International Conference on Systems, Computing Sciences and Software Engineering (SCSS 2007) which was part of the International Joint Conferences on Computer, Information and Systems Sciences and Engineering (CISSE 2007).
    Classification
    ST 230
    Content
    Inhalt: Image and Pattern Recognition: Compression, Image processing, Signal Processing Architectures, Signal Processing for Communication, Signal Processing Implementation, Speech Compression, and Video Coding Architectures. Languages and Systems: Algorithms, Databases, Embedded Systems and Applications, File Systems and I/O, Geographical Information Systems, Kernel and OS Structures, Knowledge Based Systems, Modeling and Simulation, Object Based Software Engineering, Programming Languages, and Programming Models and tools. Parallel Processing: Distributed Scheduling, Multiprocessing, Real-time Systems, Simulation Modeling and Development, and Web Applications. New trends in computing: Computers for People of Special Needs, Fuzzy Inference, Human Computer Interaction, Incremental Learning, Internet-based Computing Models, Machine Intelligence, Natural Language Processing, Neural Networks, and Online Decision Support System
    LCSH
    Communications Engineering, Networks
    Software Engineering/Programming and Operating Systems
    Software engineering
    RSWK
    Computerarchitektur / Software Engineering / Telekommunikation / Online-Publikation
    RVK
    ST 230
    Subject
    Computerarchitektur / Software Engineering / Telekommunikation / Online-Publikation
    Communications Engineering, Networks
    Software Engineering/Programming and Operating Systems
    Software engineering
  2. Stock, W.G.: Wissensrepräsentation (2014) 0.05
    0.05405436 = product of:
      0.18018119 = sum of:
        0.028581016 = weight(_text_:23 in 5153) [ClassicSimilarity], result of:
          0.028581016 = score(doc=5153,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.3959864 = fieldWeight in 5153, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.078125 = fieldNorm(doc=5153)
        0.028581016 = weight(_text_:23 in 5153) [ClassicSimilarity], result of:
          0.028581016 = score(doc=5153,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.3959864 = fieldWeight in 5153, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.078125 = fieldNorm(doc=5153)
        0.015457011 = weight(_text_:und in 5153) [ClassicSimilarity], result of:
          0.015457011 = score(doc=5153,freq=4.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.34630734 = fieldWeight in 5153, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=5153)
        0.028581016 = weight(_text_:23 in 5153) [ClassicSimilarity], result of:
          0.028581016 = score(doc=5153,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.3959864 = fieldWeight in 5153, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.078125 = fieldNorm(doc=5153)
        0.05975184 = weight(_text_:methoden in 5153) [ClassicSimilarity], result of:
          0.05975184 = score(doc=5153,freq=2.0), product of:
            0.10436003 = queryWeight, product of:
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.02013827 = queryNorm
            0.5725548 = fieldWeight in 5153, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.078125 = fieldNorm(doc=5153)
        0.019229298 = weight(_text_:der in 5153) [ClassicSimilarity], result of:
          0.019229298 = score(doc=5153,freq=6.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.42746788 = fieldWeight in 5153, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.078125 = fieldNorm(doc=5153)
      0.3 = coord(6/20)
    
    Abstract
    Wissensrepräsentation thematisiert die Erstellung von Stellvertretern (Repräsentanten bzw. Surrogaten), die die Aboutness eines Dokuments abbilden. Man unterscheidet bei der Wissensrepräsentation zwischen Informationsfiltern (z.B. Wissensordnungen) und Methoden der Informationsverdichtung (z.B. Extracts oder Abstracts).
    Date
    3. 2.2015 19:23:02
    Source
    Lexikon der Bibliotheks- und Informationswissenschaft, Band 2, Lieferung 7
  3. Beierle, C.; Kern-Isberner, G.: Methoden wissensbasierter Systeme : Grundlagen, Algorithmen, Anwendungen (2008) 0.05
    0.0503535 = product of:
      0.14386715 = sum of:
        0.016167864 = weight(_text_:23 in 4622) [ClassicSimilarity], result of:
          0.016167864 = score(doc=4622,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.22400373 = fieldWeight in 4622, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.03125 = fieldNorm(doc=4622)
        0.016167864 = weight(_text_:23 in 4622) [ClassicSimilarity], result of:
          0.016167864 = score(doc=4622,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.22400373 = fieldWeight in 4622, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.03125 = fieldNorm(doc=4622)
        0.013825171 = weight(_text_:und in 4622) [ClassicSimilarity], result of:
          0.013825171 = score(doc=4622,freq=20.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.3097467 = fieldWeight in 4622, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.03125 = fieldNorm(doc=4622)
        0.016167864 = weight(_text_:23 in 4622) [ClassicSimilarity], result of:
          0.016167864 = score(doc=4622,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.22400373 = fieldWeight in 4622, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.03125 = fieldNorm(doc=4622)
        0.05344367 = weight(_text_:methoden in 4622) [ClassicSimilarity], result of:
          0.05344367 = score(doc=4622,freq=10.0), product of:
            0.10436003 = queryWeight, product of:
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.02013827 = queryNorm
            0.5121086 = fieldWeight in 4622, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.03125 = fieldNorm(doc=4622)
        0.009929967 = weight(_text_:der in 4622) [ClassicSimilarity], result of:
          0.009929967 = score(doc=4622,freq=10.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.22074346 = fieldWeight in 4622, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.03125 = fieldNorm(doc=4622)
        0.018164756 = product of:
          0.03632951 = sum of:
            0.03632951 = weight(_text_:engineering in 4622) [ClassicSimilarity], result of:
              0.03632951 = score(doc=4622,freq=4.0), product of:
                0.10819342 = queryWeight, product of:
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.02013827 = queryNorm
                0.335783 = fieldWeight in 4622, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4622)
          0.5 = coord(1/2)
      0.35 = coord(7/20)
    
    Abstract
    Dieses Buch präsentiert ein breites Spektrum aktueller Methoden zur Repräsentation und Verarbeitung (un)sicheren Wissens in maschinellen Systemen in didaktisch aufbereiteter Form. Neben symbolischen Ansätzen des nichtmonotonen Schließens (Default-Logik, hier konstruktiv und leicht verständlich mittels sog. Default-Bäume realisiert) werden auch ausführlich quantitative Methoden wie z.B. probabilistische Markov- und Bayes-Netze vorgestellt. Weitere Abschnitte beschäftigen sich mit Wissensdynamik (Truth Maintenance-Systeme), Aktionen und Planen, maschinellem Lernen, Data Mining und fallbasiertem Schließen.In einem vertieften Querschnitt werden zentrale alternative Ansätze einer logikbasierten Wissensmodellierung ausführlich behandelt. Detailliert beschriebene Algorithmen geben dem Praktiker nützliche Hinweise zur Anwendung der vorgestellten Ansätze an die Hand, während fundiertes Hintergrundwissen ein tieferes Verständnis für die Besonderheiten der einzelnen Methoden vermittelt . Mit einer weitgehend vollständigen Darstellung des Stoffes und zahlreichen, in den Text integrierten Aufgaben ist das Buch für ein Selbststudium konzipiert, eignet sich aber gleichermaßen für eine entsprechende Vorlesung. Im Online-Service zu diesem Buch werden u.a. ausführliche Lösungshinweise zu allen Aufgaben des Buches angeboten.Zahlreiche Beispiele mit medizinischem, biologischem, wirtschaftlichem und technischem Hintergrund illustrieren konkrete Anwendungsszenarien. Von namhaften Professoren empfohlen: State-of-the-Art bietet das Buch zu diesem klassischen Bereich der Informatik. Die wesentlichen Methoden wissensbasierter Systeme werden verständlich und anschaulich dargestellt. Repräsentation und Verarbeitung sicheren und unsicheren Wissens in maschinellen Systemen stehen dabei im Mittelpunkt. In der vierten, verbesserten Auflage wurde die Anzahl der motivierenden Selbsttestaufgaben mit aktuellem Praxisbezug nochmals erweitert. Ein Online-Service mit ausführlichen Musterlösungen erleichtert das Lernen.
    Date
    11. 2.2016 16:23:23
    LCSH
    Engineering
    Subject
    Engineering
  4. Reif, G.: Semantische Annotation (2006) 0.05
    0.04594432 = product of:
      0.15314773 = sum of:
        0.028013978 = weight(_text_:software in 5807) [ClassicSimilarity], result of:
          0.028013978 = score(doc=5807,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.35064998 = fieldWeight in 5807, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0625 = fieldNorm(doc=5807)
        0.008743806 = weight(_text_:und in 5807) [ClassicSimilarity], result of:
          0.008743806 = score(doc=5807,freq=2.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.19590102 = fieldWeight in 5807, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=5807)
        0.028013978 = weight(_text_:software in 5807) [ClassicSimilarity], result of:
          0.028013978 = score(doc=5807,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.35064998 = fieldWeight in 5807, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0625 = fieldNorm(doc=5807)
        0.04780147 = weight(_text_:methoden in 5807) [ClassicSimilarity], result of:
          0.04780147 = score(doc=5807,freq=2.0), product of:
            0.10436003 = queryWeight, product of:
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.02013827 = queryNorm
            0.45804384 = fieldWeight in 5807, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.0625 = fieldNorm(doc=5807)
        0.012560525 = weight(_text_:der in 5807) [ClassicSimilarity], result of:
          0.012560525 = score(doc=5807,freq=4.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.27922085 = fieldWeight in 5807, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0625 = fieldNorm(doc=5807)
        0.028013978 = weight(_text_:software in 5807) [ClassicSimilarity], result of:
          0.028013978 = score(doc=5807,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.35064998 = fieldWeight in 5807, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0625 = fieldNorm(doc=5807)
      0.3 = coord(6/20)
    
    Abstract
    In diesem Kapitel wird zuerst der Begriff Semantische Annotation eingeführt und es werden Techniken besprochen um die Annotationen mit dem ursprünglichen Dokument zu verknüpfen. Weiters wird auf Probleme eingegangen, die sich beim Erstellen der Annotationen ergeben. Im Anschluss daran werden Software Tools vorgestellt, die einen Benutzer beim Annotierungsprozess unterstützen. Zum Abschluss werden Methoden diskutiert, die den Annotierungsvorgang in den Entwicklungsprozess einer Web Applikation integrieren.
  5. Kienreich, W.; Strohmaier, M.: Wissensmodellierung - Basis für die Anwendung semantischer Technologien (2006) 0.05
    0.045178656 = product of:
      0.15059552 = sum of:
        0.02451223 = weight(_text_:software in 5804) [ClassicSimilarity], result of:
          0.02451223 = score(doc=5804,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.30681872 = fieldWeight in 5804, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5804)
        0.013251626 = weight(_text_:und in 5804) [ClassicSimilarity], result of:
          0.013251626 = score(doc=5804,freq=6.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.2968967 = fieldWeight in 5804, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5804)
        0.02451223 = weight(_text_:software in 5804) [ClassicSimilarity], result of:
          0.02451223 = score(doc=5804,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.30681872 = fieldWeight in 5804, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5804)
        0.041826285 = weight(_text_:methoden in 5804) [ClassicSimilarity], result of:
          0.041826285 = score(doc=5804,freq=2.0), product of:
            0.10436003 = queryWeight, product of:
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.02013827 = queryNorm
            0.40078837 = fieldWeight in 5804, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5804)
        0.021980919 = weight(_text_:der in 5804) [ClassicSimilarity], result of:
          0.021980919 = score(doc=5804,freq=16.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.4886365 = fieldWeight in 5804, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5804)
        0.02451223 = weight(_text_:software in 5804) [ClassicSimilarity], result of:
          0.02451223 = score(doc=5804,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.30681872 = fieldWeight in 5804, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5804)
      0.3 = coord(6/20)
    
    Abstract
    Der Transfer und die Nutzung von Wissen stellen ein zentrales Thema hei der Anwendung semantischer Technologien dar. In diesem Zusammenhang befasst sich das Gebiet der Wissensmodellierung mit der Explizierung von Wissen in formale, sowohl von Menschen als auch von Maschinen interpretierbare, Form. Ziel dieses Beitrags ist es aufzuzeigen. wie Methoden der Wissensmodellierung die Grundlage für die Gestaltung von Anwendungen auf Basis semantischer Technologien bilden. Der Beitrag liefert eine Definition eines Wissensbegriffs, erklärt eine Reihe von Formen der abstrakten Wissensrepräsentation und führt ein Kategorisierungsschema für aktuelle Ansätze zur Modellierung ein. Anschließend wird ein Überblick über agenten- und prozessorientierte Modellierungsansätze gegeben. die sowohl auf die Abbildung der realen Welt als auch auf die Abbildung von Software eingehen.
  6. Stuckenschmidt, H.; Harmelen, F. van: Information sharing on the semantic web (2005) 0.04
    0.03915523 = product of:
      0.15662092 = sum of:
        0.09888177 = weight(_text_:monographien in 2789) [ClassicSimilarity], result of:
          0.09888177 = score(doc=2789,freq=8.0), product of:
            0.13425075 = queryWeight, product of:
              6.666449 = idf(docFreq=152, maxDocs=44218)
              0.02013827 = queryNorm
            0.73654544 = fieldWeight in 2789, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              6.666449 = idf(docFreq=152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2789)
        0.005464879 = weight(_text_:und in 2789) [ClassicSimilarity], result of:
          0.005464879 = score(doc=2789,freq=2.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.12243814 = fieldWeight in 2789, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2789)
        0.021717992 = product of:
          0.043435983 = sum of:
            0.043435983 = weight(_text_:allgemein in 2789) [ClassicSimilarity], result of:
              0.043435983 = score(doc=2789,freq=4.0), product of:
                0.10581345 = queryWeight, product of:
                  5.254347 = idf(docFreq=627, maxDocs=44218)
                  0.02013827 = queryNorm
                0.41049585 = fieldWeight in 2789, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.254347 = idf(docFreq=627, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2789)
          0.5 = coord(1/2)
        0.007850328 = weight(_text_:der in 2789) [ClassicSimilarity], result of:
          0.007850328 = score(doc=2789,freq=4.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.17451303 = fieldWeight in 2789, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2789)
        0.022705944 = product of:
          0.04541189 = sum of:
            0.04541189 = weight(_text_:engineering in 2789) [ClassicSimilarity], result of:
              0.04541189 = score(doc=2789,freq=4.0), product of:
                0.10819342 = queryWeight, product of:
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.02013827 = queryNorm
                0.41972876 = fieldWeight in 2789, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2789)
          0.5 = coord(1/2)
      0.25 = coord(5/20)
    
    Abstract
    Das wachsende Informationsvolumen im WWW führt paradoxerweise zu einer immer schwierigeren Nutzung, das Finden und Verknüpfen von Informationen in einem unstrukturierten Umfeld wird zur Sisyphosarbeit. Hier versprechen Semantic-Web-Ansätze Abhilfe. Die Autoren beschreiben Technologien, wie eine semantische Integration verteilter Daten durch verteilte Ontologien erreicht werden kann. Diese Techniken sind sowohl für Forscher als auch für Professionals interessant, die z.B. die Integration von Produktdaten aus verteilten Datenbanken im WWW oder von lose miteinander verbunden Anwendungen in verteilten Organisationen implementieren sollen.
    Classification
    ST 205 Informatik / Monographien / Vernetzung, verteilte Systeme / Internet allgemein
    ST 515 Informatik / Monographien / Einzelne Anwendungen der Datenverarbeitung / Wirtschaftsinformatik / Wissensmanagement, Information engineering
    RVK
    ST 205 Informatik / Monographien / Vernetzung, verteilte Systeme / Internet allgemein
    ST 515 Informatik / Monographien / Einzelne Anwendungen der Datenverarbeitung / Wirtschaftsinformatik / Wissensmanagement, Information engineering
  7. Auer, S.; Sens, I.; Stocker, M.: Erschließung wissenschaftlicher Literatur mit dem Open Research Knowledge Graph (2020) 0.04
    0.035552423 = product of:
      0.11850807 = sum of:
        0.017148608 = weight(_text_:23 in 551) [ClassicSimilarity], result of:
          0.017148608 = score(doc=551,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.23759183 = fieldWeight in 551, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=551)
        0.017148608 = weight(_text_:23 in 551) [ClassicSimilarity], result of:
          0.017148608 = score(doc=551,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.23759183 = fieldWeight in 551, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=551)
        0.019673564 = weight(_text_:und in 551) [ClassicSimilarity], result of:
          0.019673564 = score(doc=551,freq=18.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.4407773 = fieldWeight in 551, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=551)
        0.017148608 = weight(_text_:23 in 551) [ClassicSimilarity], result of:
          0.017148608 = score(doc=551,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.23759183 = fieldWeight in 551, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=551)
        0.035851102 = weight(_text_:methoden in 551) [ClassicSimilarity], result of:
          0.035851102 = score(doc=551,freq=2.0), product of:
            0.10436003 = queryWeight, product of:
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.02013827 = queryNorm
            0.3435329 = fieldWeight in 551, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.046875 = fieldNorm(doc=551)
        0.011537581 = weight(_text_:der in 551) [ClassicSimilarity], result of:
          0.011537581 = score(doc=551,freq=6.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.25648075 = fieldWeight in 551, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.046875 = fieldNorm(doc=551)
      0.3 = coord(6/20)
    
    Abstract
    Die Weitergabe von Wissen hat sich seit vielen hundert Jahren nicht grundlegend verändert: Sie erfolgt in der Regel dokumentenbasiert - früher als klassischer Aufsatz auf Papier gedruckt, heute als PDF. Bei jährlich rund 2,5 Millionen neuen Forschungsbeiträgen ertrinken die Forschenden in einer Flut pseudodigitalisierter PDF-Publikationen. Die Folge: Die Forschung wird ernsthaft geschwächt. Denn viele Forschungsergebnisse können durch andere nicht reproduziert werden, es gibt mehr und mehr Redundanzen und das Meer von Publikationen ist unübersichtlich. Deshalb denkt die TIB - Leibniz-Informationszentrum Technik und Naturwissenschaften Wissenskommunikation neu: Statt auf statische PDF-Artikel setzt die TIB auf Wissensgraphen. Sie arbeitet daran, Wissen unterschiedlichster Form - Texte, Bilder, Grafiken, Audio- und Video-Dateien, 3D-Modelle und vieles mehr - intuitiv mithilfe dynamischer Wissensgraphen zu vernetzen. Der Wissensgraph soll verschiedene Forschungsideen, -ansätze, -methoden und -ergebnisse maschinenlesbar darstellen, sodass völlig neue Zusammenhänge von Wissen zutage treten und zur Lösung globaler Probleme beitragen könnten. Die großen gesellschaftlichen Herausforderungen verlangen Interdisziplinarität und das Zusammenfügen von Erkenntnis-Einzelteilen. Mit dem Wissensgraphen kann das gelingen und der Fluss wissenschaftlicher Erkenntnisse revolutioniert werden.
    Source
    B.I.T. Online. 23(2020) H.5, S.491-499
  8. Bittner, T.; Donnelly, M.; Winter, S.: Ontology and semantic interoperability (2006) 0.04
    0.035469428 = product of:
      0.14187771 = sum of:
        0.036391225 = weight(_text_:software in 4820) [ClassicSimilarity], result of:
          0.036391225 = score(doc=4820,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.4555077 = fieldWeight in 4820, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4820)
        0.036391225 = weight(_text_:software in 4820) [ClassicSimilarity], result of:
          0.036391225 = score(doc=4820,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.4555077 = fieldWeight in 4820, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4820)
        0.036391225 = weight(_text_:software in 4820) [ClassicSimilarity], result of:
          0.036391225 = score(doc=4820,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.4555077 = fieldWeight in 4820, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4820)
        0.027247135 = product of:
          0.05449427 = sum of:
            0.05449427 = weight(_text_:engineering in 4820) [ClassicSimilarity], result of:
              0.05449427 = score(doc=4820,freq=4.0), product of:
                0.10819342 = queryWeight, product of:
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.02013827 = queryNorm
                0.5036745 = fieldWeight in 4820, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4820)
          0.5 = coord(1/2)
        0.005456915 = product of:
          0.016370745 = sum of:
            0.016370745 = weight(_text_:22 in 4820) [ClassicSimilarity], result of:
              0.016370745 = score(doc=4820,freq=2.0), product of:
                0.07052079 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02013827 = queryNorm
                0.23214069 = fieldWeight in 4820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4820)
          0.33333334 = coord(1/3)
      0.25 = coord(5/20)
    
    Abstract
    One of the major problems facing systems for Computer Aided Design (CAD), Architecture Engineering and Construction (AEC) and Geographic Information Systems (GIS) applications today is the lack of interoperability among the various systems. When integrating software applications, substantial di culties can arise in translating information from one application to the other. In this paper, we focus on semantic di culties that arise in software integration. Applications may use di erent terminologies to describe the same domain. Even when appli-cations use the same terminology, they often associate di erent semantics with the terms. This obstructs information exchange among applications. To cir-cumvent this obstacle, we need some way of explicitly specifying the semantics for each terminology in an unambiguous fashion. Ontologies can provide such specification. It will be the task of this paper to explain what ontologies are and how they can be used to facilitate interoperability between software systems used in computer aided design, architecture engineering and construction, and geographic information processing.
    Date
    3.12.2016 18:39:22
  9. Giunchiglia, F.; Zaihrayeu, I.; Farazi, F.: Converting classifications into OWL ontologies (2009) 0.03
    0.034343183 = product of:
      0.11447728 = sum of:
        0.017148608 = weight(_text_:23 in 4690) [ClassicSimilarity], result of:
          0.017148608 = score(doc=4690,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.23759183 = fieldWeight in 4690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=4690)
        0.017148608 = weight(_text_:23 in 4690) [ClassicSimilarity], result of:
          0.017148608 = score(doc=4690,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.23759183 = fieldWeight in 4690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=4690)
        0.021010485 = weight(_text_:software in 4690) [ClassicSimilarity], result of:
          0.021010485 = score(doc=4690,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.2629875 = fieldWeight in 4690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4690)
        0.017148608 = weight(_text_:23 in 4690) [ClassicSimilarity], result of:
          0.017148608 = score(doc=4690,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.23759183 = fieldWeight in 4690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=4690)
        0.021010485 = weight(_text_:software in 4690) [ClassicSimilarity], result of:
          0.021010485 = score(doc=4690,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.2629875 = fieldWeight in 4690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4690)
        0.021010485 = weight(_text_:software in 4690) [ClassicSimilarity], result of:
          0.021010485 = score(doc=4690,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.2629875 = fieldWeight in 4690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4690)
      0.3 = coord(6/20)
    
    Abstract
    Classification schemes, such as the DMoZ web directory, provide a convenient and intuitive way for humans to access classified contents. While being easy to be dealt with for humans, classification schemes remain hard to be reasoned about by automated software agents. Among other things, this hardness is conditioned by the ambiguous na- ture of the natural language used to describe classification categories. In this paper we describe how classification schemes can be converted into OWL ontologies, thus enabling reasoning on them by Semantic Web applications. The proposed solution is based on a two phase approach in which category names are first encoded in a concept language and then, together with the structure of the classification scheme, are converted into an OWL ontology. We demonstrate the practical applicability of our approach by showing how the results of reasoning on these OWL ontologies can help improve the organization and use of web directories.
    Date
    23. 8.2011 19:43:39
  10. Knorz, G.; Rein, B.: Semantische Suche in einer Hochschulontologie (2005) 0.03
    0.034255974 = product of:
      0.11418658 = sum of:
        0.02451223 = weight(_text_:software in 1852) [ClassicSimilarity], result of:
          0.02451223 = score(doc=1852,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.30681872 = fieldWeight in 1852, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1852)
        0.01874063 = weight(_text_:und in 1852) [ClassicSimilarity], result of:
          0.01874063 = score(doc=1852,freq=12.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.41987535 = fieldWeight in 1852, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1852)
        0.02451223 = weight(_text_:software in 1852) [ClassicSimilarity], result of:
          0.02451223 = score(doc=1852,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.30681872 = fieldWeight in 1852, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1852)
        0.015542857 = weight(_text_:der in 1852) [ClassicSimilarity], result of:
          0.015542857 = score(doc=1852,freq=8.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.34551817 = fieldWeight in 1852, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1852)
        0.02451223 = weight(_text_:software in 1852) [ClassicSimilarity], result of:
          0.02451223 = score(doc=1852,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.30681872 = fieldWeight in 1852, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1852)
        0.006366401 = product of:
          0.019099202 = sum of:
            0.019099202 = weight(_text_:22 in 1852) [ClassicSimilarity], result of:
              0.019099202 = score(doc=1852,freq=2.0), product of:
                0.07052079 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02013827 = queryNorm
                0.2708308 = fieldWeight in 1852, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1852)
          0.33333334 = coord(1/3)
      0.3 = coord(6/20)
    
    Abstract
    Ontologien werden eingesetzt, um durch semantische Fundierung insbesondere für das Dokumentenretrieval eine grundlegend bessere Basis zu haben, als dies gegenwärtiger Stand der Technik ist. Vorgestellt wird eine an der FH Darmstadt entwickelte und eingesetzte Ontologie, die den Gegenstandsbereich Hochschule sowohl breit abdecken und gleichzeitig differenziert semantisch beschreiben soll. Das Problem der semantischen Suche besteht nun darin, dass sie für Informationssuchende so einfach wie bei gängigen Suchmaschinen zu nutzen sein soll, und gleichzeitig auf der Grundlage des aufwendigen Informationsmodells hochwertige Ergebnisse liefern muss. Es wird beschrieben, welche Möglichkeiten die verwendete Software K-Infinity bereitstellt und mit welchem Konzept diese Möglichkeiten für eine semantische Suche nach Dokumenten und anderen Informationseinheiten (Personen, Veranstaltungen, Projekte etc.) eingesetzt werden.
    Date
    11. 2.2011 18:22:58
    Source
    Information - Wissenschaft und Praxis. 56(2005) H.5/6, S.281-290
  11. Knorz, G.; Rein, B.: Semantische Suche in einer Hochschulontologie : Ontologie-basiertes Information-Filtering und -Retrieval mit relationalen Datenbanken (2005) 0.03
    0.034255974 = product of:
      0.11418658 = sum of:
        0.02451223 = weight(_text_:software in 4324) [ClassicSimilarity], result of:
          0.02451223 = score(doc=4324,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.30681872 = fieldWeight in 4324, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4324)
        0.01874063 = weight(_text_:und in 4324) [ClassicSimilarity], result of:
          0.01874063 = score(doc=4324,freq=12.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.41987535 = fieldWeight in 4324, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4324)
        0.02451223 = weight(_text_:software in 4324) [ClassicSimilarity], result of:
          0.02451223 = score(doc=4324,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.30681872 = fieldWeight in 4324, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4324)
        0.015542857 = weight(_text_:der in 4324) [ClassicSimilarity], result of:
          0.015542857 = score(doc=4324,freq=8.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.34551817 = fieldWeight in 4324, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4324)
        0.02451223 = weight(_text_:software in 4324) [ClassicSimilarity], result of:
          0.02451223 = score(doc=4324,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.30681872 = fieldWeight in 4324, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4324)
        0.006366401 = product of:
          0.019099202 = sum of:
            0.019099202 = weight(_text_:22 in 4324) [ClassicSimilarity], result of:
              0.019099202 = score(doc=4324,freq=2.0), product of:
                0.07052079 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02013827 = queryNorm
                0.2708308 = fieldWeight in 4324, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4324)
          0.33333334 = coord(1/3)
      0.3 = coord(6/20)
    
    Abstract
    Ontologien werden eingesetzt, um durch semantische Fundierung insbesondere für das Dokumentenretrieval eine grundlegend bessere Basis zu haben, als dies gegenwärtiger Stand der Technik ist. Vorgestellt wird eine an der FH Darmstadt entwickelte und eingesetzte Ontologie, die den Gegenstandsbereich Hochschule sowohl breit abdecken und gleichzeitig differenziert semantisch beschreiben soll. Das Problem der semantischen Suche besteht nun darin, dass sie für Informationssuchende so einfach wie bei gängigen Suchmaschinen zu nutzen sein soll, und gleichzeitig auf der Grundlage des aufwendigen Informationsmodells hochwertige Ergebnisse liefern muss. Es wird beschrieben, welche Möglichkeiten die verwendete Software K-Infinity bereitstellt und mit welchem Konzept diese Möglichkeiten für eine semantische Suche nach Dokumenten und anderen Informationseinheiten (Personen, Veranstaltungen, Projekte etc.) eingesetzt werden.
    Date
    11. 2.2011 18:22:25
  12. Fagundes, P.B.; Freund, G.P.; Vital, L.P.; Monteiro de Barros, C.; Macedo, D.D.J.de: Taxonomias, ontologias e tesauros : possibilidades de contribuição para o processo de Engenharia de Requisitos (2020) 0.03
    0.0339427 = product of:
      0.1357708 = sum of:
        0.030326022 = weight(_text_:software in 5828) [ClassicSimilarity], result of:
          0.030326022 = score(doc=5828,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.37958977 = fieldWeight in 5828, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5828)
        0.005464879 = weight(_text_:und in 5828) [ClassicSimilarity], result of:
          0.005464879 = score(doc=5828,freq=2.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.12243814 = fieldWeight in 5828, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5828)
        0.030326022 = weight(_text_:software in 5828) [ClassicSimilarity], result of:
          0.030326022 = score(doc=5828,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.37958977 = fieldWeight in 5828, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5828)
        0.030326022 = weight(_text_:software in 5828) [ClassicSimilarity], result of:
          0.030326022 = score(doc=5828,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.37958977 = fieldWeight in 5828, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5828)
        0.03932785 = product of:
          0.0786557 = sum of:
            0.0786557 = weight(_text_:engineering in 5828) [ClassicSimilarity], result of:
              0.0786557 = score(doc=5828,freq=12.0), product of:
                0.10819342 = queryWeight, product of:
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.02013827 = queryNorm
                0.72699153 = fieldWeight in 5828, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5828)
          0.5 = coord(1/2)
      0.25 = coord(5/20)
    
    Abstract
    Some of the fundamental activities of the software development process are related to the discipline of Requirements Engineering, whose objective is the discovery, analysis, documentation and verification of the requirements that will be part of the system. Requirements are the conditions or capabilities that software must have or perform to meet the users needs. The present study is being developed to propose a model of cooperation between Information Science and Requirements Engineering. Aims to present the analysis results on the possibilities of using the knowledge organization systems: taxonomies, thesauri and ontologies during the activities of Requirements Engineering: design, survey, elaboration, negotiation, specification, validation and requirements management. From the results obtained it was possible to identify in which stage of the Requirements Engineering process, each type of knowledge organization system could be used. We expect that this study put in evidence the need for new researchs and proposals to strengt the exchange between Information Science, as a science that has information as object of study, and the Requirements Engineering which has in the information the raw material to identify the informational needs of software users.
    Footnote
    Engl. Übers. des Titels: Taxonomies, ontologies and thesauri: possibilities of contribution to the process of Requirements Engineering.
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  13. Schubert, C.; Kinkeldey, C.; Reich, H.: Handbuch Datenbankanwendung zur Wissensrepräsentation im Verbundprojekt DeCOVER (2006) 0.03
    0.033549864 = product of:
      0.11183288 = sum of:
        0.022864813 = weight(_text_:23 in 4256) [ClassicSimilarity], result of:
          0.022864813 = score(doc=4256,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.31678912 = fieldWeight in 4256, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0625 = fieldNorm(doc=4256)
        0.022864813 = weight(_text_:23 in 4256) [ClassicSimilarity], result of:
          0.022864813 = score(doc=4256,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.31678912 = fieldWeight in 4256, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0625 = fieldNorm(doc=4256)
        0.0123656085 = weight(_text_:und in 4256) [ClassicSimilarity], result of:
          0.0123656085 = score(doc=4256,freq=4.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.27704588 = fieldWeight in 4256, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=4256)
        0.022864813 = weight(_text_:23 in 4256) [ClassicSimilarity], result of:
          0.022864813 = score(doc=4256,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.31678912 = fieldWeight in 4256, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0625 = fieldNorm(doc=4256)
        0.019859934 = weight(_text_:der in 4256) [ClassicSimilarity], result of:
          0.019859934 = score(doc=4256,freq=10.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.44148692 = fieldWeight in 4256, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0625 = fieldNorm(doc=4256)
        0.011012898 = product of:
          0.022025796 = sum of:
            0.022025796 = weight(_text_:29 in 4256) [ClassicSimilarity], result of:
              0.022025796 = score(doc=4256,freq=2.0), product of:
                0.070840135 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.02013827 = queryNorm
                0.31092256 = fieldWeight in 4256, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4256)
          0.5 = coord(1/2)
      0.3 = coord(6/20)
    
    Abstract
    Die Datenbank basierte Objektartenbeschreibung dient zur eigenschaftsbasierten Aufnahme aller Objektarten der Kataloge BNTK, CLC; GMES M 2.1, ATKIS und des DeCOVER Vorschlags. Das Ziel der Datenbankanwendung besteht in der 'manuellen' Beziehungsauswertung und Darstellung der gesamten Objektarten bezogen auf die erstellte Wissensrepräsentation. Anhand einer hierarchisch strukturierten Wissensrepräsentation lassen sich mit Ontologien Überführungen von Objektarten verwirklichen, die im Sinne der semantischen Interoperabilität als Zielstellung in dem Verbundprojekt DeCOVER besteht.
    Date
    29. 1.2011 18:45:23
  14. Beßler, S.: Wissensrepräsentation musealer Bestände mittels semantischer Netze : Analyse und Annotation eines Teilbestands des Haus der Geschichte der BRD in Bonn (2010) 0.03
    0.031548575 = product of:
      0.1261943 = sum of:
        0.029713312 = weight(_text_:software in 4024) [ClassicSimilarity], result of:
          0.029713312 = score(doc=4024,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3719205 = fieldWeight in 4024, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4024)
        0.020737756 = weight(_text_:und in 4024) [ClassicSimilarity], result of:
          0.020737756 = score(doc=4024,freq=20.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.46462005 = fieldWeight in 4024, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=4024)
        0.029713312 = weight(_text_:software in 4024) [ClassicSimilarity], result of:
          0.029713312 = score(doc=4024,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3719205 = fieldWeight in 4024, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4024)
        0.0163166 = weight(_text_:der in 4024) [ClassicSimilarity], result of:
          0.0163166 = score(doc=4024,freq=12.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.36271852 = fieldWeight in 4024, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.046875 = fieldNorm(doc=4024)
        0.029713312 = weight(_text_:software in 4024) [ClassicSimilarity], result of:
          0.029713312 = score(doc=4024,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3719205 = fieldWeight in 4024, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4024)
      0.25 = coord(5/20)
    
    Abstract
    Semantische Netze unterstützen den Suchvorgang im Information Retrieval. Über ihre vielfältigen Relationen und Inferenzen unterstützen sie den Anwender und helfen Daten im Kontext zu präsentieren und zu erfassen. Die Relationen ermöglichen Suchanfragen die große Treffermengen produzieren zu verfeinern und so Treffermengen zu erreichen die möglichst genau das enthalten was gesucht wurde. Es wird, anhand eines Ausschnitts des Datenbestands des Haus der Geschichte der Bundesrepublik Deutschland in Bonn, aufgezeigt wie bestehende Datenbestände in semantische Netze überführt werden können und wie diese anschließend für das Retrieval eingesetzt werden können. Für die Modellierung des semantischen Netz wird die Open Source Software Protégé in den Versionen 3.4.4. und 4.1_beta eingesetzt, die Möglichkeiten des Retrieval werden anhand der Abfragesprachen DL Query und SPARQL sowie anhand der Software Ontology Browser und OntoGraf erläutert.
    Imprint
    Köln : Fachhochschule / Fakultät für Informations- und Kommunikationswissenschaften
  15. ¬The Semantic Web : research and applications ; second European Semantic WebConference, ESWC 2005, Heraklion, Crete, Greece, May 29 - June 1, 2005 ; proceedings (2005) 0.03
    0.031161685 = product of:
      0.12464674 = sum of:
        0.029713312 = weight(_text_:software in 439) [ClassicSimilarity], result of:
          0.029713312 = score(doc=439,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3719205 = fieldWeight in 439, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=439)
        0.029713312 = weight(_text_:software in 439) [ClassicSimilarity], result of:
          0.029713312 = score(doc=439,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3719205 = fieldWeight in 439, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=439)
        0.008259674 = product of:
          0.016519347 = sum of:
            0.016519347 = weight(_text_:29 in 439) [ClassicSimilarity], result of:
              0.016519347 = score(doc=439,freq=2.0), product of:
                0.070840135 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.02013827 = queryNorm
                0.23319192 = fieldWeight in 439, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=439)
          0.5 = coord(1/2)
        0.029713312 = weight(_text_:software in 439) [ClassicSimilarity], result of:
          0.029713312 = score(doc=439,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3719205 = fieldWeight in 439, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=439)
        0.027247135 = product of:
          0.05449427 = sum of:
            0.05449427 = weight(_text_:engineering in 439) [ClassicSimilarity], result of:
              0.05449427 = score(doc=439,freq=4.0), product of:
                0.10819342 = queryWeight, product of:
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.02013827 = queryNorm
                0.5036745 = fieldWeight in 439, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.046875 = fieldNorm(doc=439)
          0.5 = coord(1/2)
      0.25 = coord(5/20)
    
    LCSH
    Software engineering
    Subject
    Software engineering
  16. Müller, T.: Wissensrepräsentation mit semantischen Netzen im Bereich Luftfahrt (2006) 0.03
    0.02973763 = product of:
      0.09912543 = sum of:
        0.02020983 = weight(_text_:23 in 1670) [ClassicSimilarity], result of:
          0.02020983 = score(doc=1670,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.28000468 = fieldWeight in 1670, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1670)
        0.02020983 = weight(_text_:23 in 1670) [ClassicSimilarity], result of:
          0.02020983 = score(doc=1670,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.28000468 = fieldWeight in 1670, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1670)
        0.016394636 = weight(_text_:und in 1670) [ClassicSimilarity], result of:
          0.016394636 = score(doc=1670,freq=18.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.3673144 = fieldWeight in 1670, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1670)
        0.02020983 = weight(_text_:23 in 1670) [ClassicSimilarity], result of:
          0.02020983 = score(doc=1670,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.28000468 = fieldWeight in 1670, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1670)
        0.017553868 = weight(_text_:der in 1670) [ClassicSimilarity], result of:
          0.017553868 = score(doc=1670,freq=20.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.390223 = fieldWeight in 1670, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1670)
        0.0045474293 = product of:
          0.013642288 = sum of:
            0.013642288 = weight(_text_:22 in 1670) [ClassicSimilarity], result of:
              0.013642288 = score(doc=1670,freq=2.0), product of:
                0.07052079 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02013827 = queryNorm
                0.19345059 = fieldWeight in 1670, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1670)
          0.33333334 = coord(1/3)
      0.3 = coord(6/20)
    
    Abstract
    Es ist ein semantisches Netz für den Gegenstandsbereich Luftfahrt modelliert worden, welches Unternehmensinformationen, Organisationen, Fluglinien, Flughäfen, etc. enthält, Diese sind 10 Hauptkategorien zugeordnet worden, die untergliedert nach Facetten sind. Die Begriffe des Gegenstandsbereiches sind mit 23 unterschiedlichen Relationen verknüpft worden (Z. B.: 'hat Standort in', bietet an, 'ist Homebase von', etc). Der Schwerpunkt der Betrachtung liegt auf dem Unterschied zwischen den drei klassischen Standardrelationen und den zusätzlich eingerichteten Relationen, bezüglich ihrem Nutzen für ein effizientes Retrieval. Die angelegten Kategorien und Relationen sind sowohl für eine kognitive als auch für eine maschinelle Verarbeitung geeignet.
    Das Ziel der vorliegenden Arbeit ist es, ein Modell für ein Informationssystems zu erstellen und die Voraussetzungen und Aspekte zu betrachten, die notwendig sind, um Einsichten in die begrifflichen Zusammenhänge des Gegenstandsbereiches Luftfahrt zu erlangen. Der Ansatz, der hier erläutert wird, plädiert für die Konstruktion einer begrifflichen Wissensstruktur in Form eines semantischen Netzes. Ausgangspunkt dieser Überlegungen ist die Auffassung, daß zwar das kontrollierte Vokabular eines Thesaurus mit seiner Verweisstruktur vielfältiges Wissen enthält, das aber aufgrund der drei klassischen Standardrelationen nur unzureichend repräsentiert und damit auch nur beschränkt zugänglich ist. Es wird erläutert, welche Vorteile eine Erweiterung der drei Thesaurusrelationen erbringen kann und in welcher Funktion die Relationen bei der Formulierung der Suchanfrage unterstützend sein können. Gezeigt wird, wie die Begriffstrukturen eines semantischen Netzes deutlicher hervortreten, wenn bei der Erstellung einer Wissensstruktur eines Gegenstandsbereiches Kategorien zugrunde gelegt werden und welche Gestaltungsprinzipien den Suchprozeß unterstützen können. Dazu werden die Voraussetzungen erörtert, die garantieren, daß komplexe Suchanfragen (erfolgreich) geleistet werden können und zu präzisen Treffermengen führen.
    Date
    23. 2.2005 10:27:09
    26. 9.2006 21:00:22
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  17. Burstein, M.; McDermott, D.V.: Ontology translation for interoperability among Semantic Web services (2005) 0.03
    0.028619321 = product of:
      0.09539773 = sum of:
        0.014290508 = weight(_text_:23 in 2661) [ClassicSimilarity], result of:
          0.014290508 = score(doc=2661,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.1979932 = fieldWeight in 2661, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2661)
        0.014290508 = weight(_text_:23 in 2661) [ClassicSimilarity], result of:
          0.014290508 = score(doc=2661,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.1979932 = fieldWeight in 2661, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2661)
        0.017508736 = weight(_text_:software in 2661) [ClassicSimilarity], result of:
          0.017508736 = score(doc=2661,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.21915624 = fieldWeight in 2661, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2661)
        0.014290508 = weight(_text_:23 in 2661) [ClassicSimilarity], result of:
          0.014290508 = score(doc=2661,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.1979932 = fieldWeight in 2661, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2661)
        0.017508736 = weight(_text_:software in 2661) [ClassicSimilarity], result of:
          0.017508736 = score(doc=2661,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.21915624 = fieldWeight in 2661, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2661)
        0.017508736 = weight(_text_:software in 2661) [ClassicSimilarity], result of:
          0.017508736 = score(doc=2661,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.21915624 = fieldWeight in 2661, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2661)
      0.3 = coord(6/20)
    
    Abstract
    Research on semantic web services promises greater interoperability among software agents and web services by enabling content-based automated service discovery and interaction and by utilizing. Although this is to be based on use of shared ontologies published on the semantic web, services produced and described by different developers may well use different, perhaps partly overlapping, sets of ontologies. Interoperability will depend on ontology mappings and architectures supporting the associated translation processes. The question we ask is, does the traditional approach of introducing mediator agents to translate messages between requestors and services work in such an open environment? This article reviews some of the processing assumptions that were made in the development of the semantic web service modeling ontology OWL-S and argues that, as a practical matter, the translation function cannot always be isolated in mediators. Ontology mappings need to be published on the semantic web just as ontologies themselves are. The translation for service discovery, service process model interpretation, task negotiation, service invocation, and response interpretation may then be distributed to various places in the architecture so that translation can be done in the specific goal-oriented informational contexts of the agents performing these processes. We present arguments for assigning translation responsibility to particular agents in the cases of service invocation, response translation, and match- making.
    Date
    23. 1.2016 19:07:35
  18. Schöndorf, P.: Nicht-konventionelle Thesaurusrelationen als Orientierungshilfen für Indexierung und Recherche: Analyse ausgewählter Beispiele (1988) 0.03
    0.028336016 = product of:
      0.14168008 = sum of:
        0.04001342 = weight(_text_:23 in 2311) [ClassicSimilarity], result of:
          0.04001342 = score(doc=2311,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.55438095 = fieldWeight in 2311, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.109375 = fieldNorm(doc=2311)
        0.04001342 = weight(_text_:23 in 2311) [ClassicSimilarity], result of:
          0.04001342 = score(doc=2311,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.55438095 = fieldWeight in 2311, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.109375 = fieldNorm(doc=2311)
        0.021639816 = weight(_text_:und in 2311) [ClassicSimilarity], result of:
          0.021639816 = score(doc=2311,freq=4.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.4848303 = fieldWeight in 2311, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.109375 = fieldNorm(doc=2311)
        0.04001342 = weight(_text_:23 in 2311) [ClassicSimilarity], result of:
          0.04001342 = score(doc=2311,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.55438095 = fieldWeight in 2311, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.109375 = fieldNorm(doc=2311)
      0.2 = coord(4/20)
    
    Date
    10.11.2015 10:23:01
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  19. Helbig, H.: ¬Die semantische Struktur natürlicher Sprache : Wissensrepräsentation mit MultiNet (2001) 0.03
    0.027042292 = product of:
      0.10816917 = sum of:
        0.028013978 = weight(_text_:software in 7072) [ClassicSimilarity], result of:
          0.028013978 = score(doc=7072,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.35064998 = fieldWeight in 7072, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0625 = fieldNorm(doc=7072)
        0.008743806 = weight(_text_:und in 7072) [ClassicSimilarity], result of:
          0.008743806 = score(doc=7072,freq=2.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.19590102 = fieldWeight in 7072, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=7072)
        0.028013978 = weight(_text_:software in 7072) [ClassicSimilarity], result of:
          0.028013978 = score(doc=7072,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.35064998 = fieldWeight in 7072, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0625 = fieldNorm(doc=7072)
        0.015383439 = weight(_text_:der in 7072) [ClassicSimilarity], result of:
          0.015383439 = score(doc=7072,freq=6.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.34197432 = fieldWeight in 7072, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0625 = fieldNorm(doc=7072)
        0.028013978 = weight(_text_:software in 7072) [ClassicSimilarity], result of:
          0.028013978 = score(doc=7072,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.35064998 = fieldWeight in 7072, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0625 = fieldNorm(doc=7072)
      0.25 = coord(5/20)
    
    Abstract
    Die Methodik der 'Mehrschichtigen Erweiterung Semantischer Netze' (MultiNet) ist sowohl für theoretische Untersuchungen als auch für die automatische Verarbeitung natürlicher Sprache auf dem Rechner geeignet. Die vorgestellten Ergebnisse sind eingebettet in ein System von Software-Werkzeugen, die eine praktische Nutzung der MultiNet-Darstellungsmittel als Formalismus zur Bedeutungsrepräsentation sichern
    Footnote
    2. Aufl. 2008 u.d.T.: Wissensverarbeitung und die Semantik der natürlichen Sprache
  20. Mustafa El Hadi, W.: Terminologies, ontologies and information access (2006) 0.03
    0.026324045 = product of:
      0.10529618 = sum of:
        0.022864813 = weight(_text_:23 in 1488) [ClassicSimilarity], result of:
          0.022864813 = score(doc=1488,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.31678912 = fieldWeight in 1488, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0625 = fieldNorm(doc=1488)
        0.022864813 = weight(_text_:23 in 1488) [ClassicSimilarity], result of:
          0.022864813 = score(doc=1488,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.31678912 = fieldWeight in 1488, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0625 = fieldNorm(doc=1488)
        0.022864813 = weight(_text_:23 in 1488) [ClassicSimilarity], result of:
          0.022864813 = score(doc=1488,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.31678912 = fieldWeight in 1488, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0625 = fieldNorm(doc=1488)
        0.011012898 = product of:
          0.022025796 = sum of:
            0.022025796 = weight(_text_:29 in 1488) [ClassicSimilarity], result of:
              0.022025796 = score(doc=1488,freq=2.0), product of:
                0.070840135 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.02013827 = queryNorm
                0.31092256 = fieldWeight in 1488, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1488)
          0.5 = coord(1/2)
        0.025688844 = product of:
          0.051377688 = sum of:
            0.051377688 = weight(_text_:engineering in 1488) [ClassicSimilarity], result of:
              0.051377688 = score(doc=1488,freq=2.0), product of:
                0.10819342 = queryWeight, product of:
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.02013827 = queryNorm
                0.47486886 = fieldWeight in 1488, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1488)
          0.5 = coord(1/2)
      0.25 = coord(5/20)
    
    Abstract
    Ontologies have become an important issue in research communities across several disciplines. This paper discusses some of the innovative techniques involving automatic terminology resources acquisition are briefly discussed. Suggests that NLP-based ontologies are useful in reducing the cost of ontology engineering. Emphasizes that linguistic ontologies covering both ontological and lexical information can offer solutions since they can be more easily updated by the resources of NLP products.
    Date
    29. 2.2008 16:25:23

Authors

Years

Languages

  • e 194
  • d 103
  • f 1
  • pt 1
  • sp 1
  • More… Less…

Types

  • a 210
  • el 72
  • m 27
  • x 22
  • s 12
  • n 5
  • r 5
  • p 1
  • More… Less…

Subjects

Classifications