Search (316 results, page 1 of 16)

  • × theme_ss:"Semantische Interoperabilität"
  1. Euzenat, J.; Bach, T.Le; Barrasa, J.; Bouquet, P.; Bo, J.De; Dieng, R.; Ehrig, M.; Hauswirth, M.; Jarrar, M.; Lara, R.; Maynard, D.; Napoli, A.; Stamou, G.; Stuckenschmidt, H.; Shvaiko, P.; Tessaris, S.; Acker, S. Van; Zaihrayeu, I.: State of the art on ontology alignment (2004) 0.08
    0.08146719 = product of:
      0.16293438 = sum of:
        0.015476737 = weight(_text_:for in 172) [ClassicSimilarity], result of:
          0.015476737 = score(doc=172,freq=16.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.23469281 = fieldWeight in 172, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.03125 = fieldNorm(doc=172)
        0.017707152 = weight(_text_:the in 172) [ClassicSimilarity], result of:
          0.017707152 = score(doc=172,freq=42.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.31953377 = fieldWeight in 172, product of:
              6.4807405 = tf(freq=42.0), with freq of:
                42.0 = termFreq=42.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.03125 = fieldNorm(doc=172)
        0.07501857 = weight(_text_:commission in 172) [ClassicSimilarity], result of:
          0.07501857 = score(doc=172,freq=2.0), product of:
            0.24417253 = queryWeight, product of:
              6.9519553 = idf(docFreq=114, maxDocs=44218)
              0.035122856 = queryNorm
            0.30723593 = fieldWeight in 172, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.9519553 = idf(docFreq=114, maxDocs=44218)
              0.03125 = fieldNorm(doc=172)
        0.015182858 = weight(_text_:of in 172) [ClassicSimilarity], result of:
          0.015182858 = score(doc=172,freq=32.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.27643585 = fieldWeight in 172, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=172)
        0.017707152 = weight(_text_:the in 172) [ClassicSimilarity], result of:
          0.017707152 = score(doc=172,freq=42.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.31953377 = fieldWeight in 172, product of:
              6.4807405 = tf(freq=42.0), with freq of:
                42.0 = termFreq=42.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.03125 = fieldNorm(doc=172)
        0.021841917 = product of:
          0.043683834 = sum of:
            0.043683834 = weight(_text_:communities in 172) [ClassicSimilarity], result of:
              0.043683834 = score(doc=172,freq=2.0), product of:
                0.18632571 = queryWeight, product of:
                  5.3049703 = idf(docFreq=596, maxDocs=44218)
                  0.035122856 = queryNorm
                0.23444878 = fieldWeight in 172, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.3049703 = idf(docFreq=596, maxDocs=44218)
                  0.03125 = fieldNorm(doc=172)
          0.5 = coord(1/2)
      0.5 = coord(6/12)
    
    Abstract
    In this document we provide an overall view of the state of the art in ontology alignment. It is organised as a description of the need for ontology alignment, a presentation of the techniques currently in use for ontology alignment and a presentation of existing systems. The state of the art is not restricted to any discipline and consider as some form of ontology alignment the work made on schema matching within the database area for instance. Heterogeneity problems on the semantic web can be solved, for some of them, by aligning heterogeneous ontologies. This is illustrated through a number of use cases of ontology alignment. Aligning ontologies consists of providing the corresponding entities in these ontologies. This process is precisely defined in deliverable D2.2.1. The current deliverable presents the many techniques currently used for implementing this process. These techniques are classified along the many features that can be found in ontologies (labels, structures, instances, semantics). They resort to many different disciplines such as statistics, machine learning or data analysis. The alignment itself is obtained by combining these techniques towards a particular goal (obtaining an alignment with particular features, optimising some criterion). Several combination techniques are also presented. Finally, these techniques have been experimented in various systems for ontology alignment or schema matching. Several such systems are presented briefly in the last section and characterized by the above techniques they rely on. The conclusion is that many techniques are available for achieving ontology alignment and many systems have been developed based on these techniques. However, few comparisons and few integration is actually provided by these implementations. This deliverable serves as a basis for considering further action along these two lines. It provide a first inventory of what should be evaluated and suggests what evaluation criterion can be used.
    Content
    This document is part of a research project funded by the IST Programme of the Commission of the European Communities as project number IST-2004-507482.
  2. Veltman, K.H.: Syntactic and semantic interoperability : new approaches to knowledge and the Semantic Web (2001) 0.07
    0.071124226 = product of:
      0.14224845 = sum of:
        0.004783498 = weight(_text_:information in 3883) [ClassicSimilarity], result of:
          0.004783498 = score(doc=3883,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.0775819 = fieldWeight in 3883, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=3883)
        0.012235435 = weight(_text_:for in 3883) [ClassicSimilarity], result of:
          0.012235435 = score(doc=3883,freq=10.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.18554096 = fieldWeight in 3883, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.03125 = fieldNorm(doc=3883)
        0.017280413 = weight(_text_:the in 3883) [ClassicSimilarity], result of:
          0.017280413 = score(doc=3883,freq=40.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.31183305 = fieldWeight in 3883, product of:
              6.3245554 = tf(freq=40.0), with freq of:
                40.0 = termFreq=40.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.03125 = fieldNorm(doc=3883)
        0.07501857 = weight(_text_:commission in 3883) [ClassicSimilarity], result of:
          0.07501857 = score(doc=3883,freq=2.0), product of:
            0.24417253 = queryWeight, product of:
              6.9519553 = idf(docFreq=114, maxDocs=44218)
              0.035122856 = queryNorm
            0.30723593 = fieldWeight in 3883, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.9519553 = idf(docFreq=114, maxDocs=44218)
              0.03125 = fieldNorm(doc=3883)
        0.015650133 = weight(_text_:of in 3883) [ClassicSimilarity], result of:
          0.015650133 = score(doc=3883,freq=34.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.28494355 = fieldWeight in 3883, product of:
              5.8309517 = tf(freq=34.0), with freq of:
                34.0 = termFreq=34.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=3883)
        0.017280413 = weight(_text_:the in 3883) [ClassicSimilarity], result of:
          0.017280413 = score(doc=3883,freq=40.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.31183305 = fieldWeight in 3883, product of:
              6.3245554 = tf(freq=40.0), with freq of:
                40.0 = termFreq=40.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.03125 = fieldNorm(doc=3883)
      0.5 = coord(6/12)
    
    Abstract
    At VVWW-7 (Brisbane, 1997), Tim Berners-Lee outlined his vision of a global reasoning web. At VVWW- 8 (Toronto, May 1998), he developed this into a vision of a semantic web, where one Gould search not just for isolated words, but for meaning in the form of logically provable claims. In the past four years this vision has spread with amazing speed. The semantic web has been adopted by the European Commission as one of the important goals of the Sixth Framework Programme. In the United States it has become linked with the Defense Advanced Research Projects Agency (DARPA). While this quest to achieve a semantic web is new, the quest for meaning in language has a history that is almost as old as language itself. Accordingly this paper opens with a survey of the historical background. The contributions of the Dublin Core are reviewed briefly. To achieve a semantic web requires both syntactic and semantic interoperability. These challenges are outlined. A basic contention of this paper is that semantic interoperability requires much more than a simple agreement concerning the static meaning of a term. Different levels of agreement (local, regional, national and international) are involved and these levels have their own history. Hence, one of the larger challenges is to create new systems of knowledge organization, which identify and connect these different levels. With respect to meaning or semantics, early twentieth century pioneers such as Wüster were hopeful that it might be sufficient to limit oneself to isolated terms and words without reference to the larger grammatical context: to concept systems rather than to propositional logic. While a fascination with concept systems implicitly dominates many contemporary discussions, this paper suggests why this approach is not sufficient. The final section of this paper explores how an approach using propositional logic could lead to a new approach to universals and particulars. This points to a re-organization of knowledge, and opens the way for a vision of a semantic web with all the historical and cultural richness and complexity of language itself.
    Footnote
    Initially written for Dublin Core Meeting in 2000 which rejected the article.
    Source
    New review of information networking. 7(2001) no.xx, S.xx-xx
  3. Garcia Marco, F.J.: Compatibility & heterogeneity in knowledge organization : some reflections around a case study in the field of consumer information (2008) 0.07
    0.06677649 = product of:
      0.100164734 = sum of:
        0.010274758 = product of:
          0.030824272 = sum of:
            0.030824272 = weight(_text_:f in 1678) [ClassicSimilarity], result of:
              0.030824272 = score(doc=1678,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.22018565 = fieldWeight in 1678, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1678)
          0.33333334 = coord(1/3)
        0.010356578 = weight(_text_:information in 1678) [ClassicSimilarity], result of:
          0.010356578 = score(doc=1678,freq=6.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.16796975 = fieldWeight in 1678, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1678)
        0.009531214 = weight(_text_:und in 1678) [ClassicSimilarity], result of:
          0.009531214 = score(doc=1678,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.12243814 = fieldWeight in 1678, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1678)
        0.006839816 = weight(_text_:for in 1678) [ClassicSimilarity], result of:
          0.006839816 = score(doc=1678,freq=2.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.103720546 = fieldWeight in 1678, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1678)
        0.017414892 = weight(_text_:the in 1678) [ClassicSimilarity], result of:
          0.017414892 = score(doc=1678,freq=26.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3142598 = fieldWeight in 1678, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1678)
        0.016435927 = weight(_text_:of in 1678) [ClassicSimilarity], result of:
          0.016435927 = score(doc=1678,freq=24.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.2992506 = fieldWeight in 1678, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1678)
        0.017414892 = weight(_text_:the in 1678) [ClassicSimilarity], result of:
          0.017414892 = score(doc=1678,freq=26.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3142598 = fieldWeight in 1678, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1678)
        0.011896656 = product of:
          0.023793312 = sum of:
            0.023793312 = weight(_text_:22 in 1678) [ClassicSimilarity], result of:
              0.023793312 = score(doc=1678,freq=2.0), product of:
                0.12299426 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035122856 = queryNorm
                0.19345059 = fieldWeight in 1678, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1678)
          0.5 = coord(1/2)
      0.6666667 = coord(8/12)
    
    Abstract
    A case study in compatibility and heterogeneity of knowledge organization (KO) systems and processes is presented. It is based in the experience of the author in the field of information for consumer protection, a good example of the emerging transdisciplinary applied social sciences. The activities and knowledge organization problems and solutions of the Aragonian Consumers' Information and Documentation Centre are described and analyzed. Six assertions can be concluded: a) heterogeneity and compatibility are certainly an inherent problem in knowledge organization and also in practical domains; b) knowledge organization is also a social task, not only a lögical one; c) knowledge organization is affected by economical and efficiency considerations; d) knowledge organization is at the heart of Knowledge Management; e) identifying and maintaining the focus in interdisciplinary fields is a must; f the different knowledge organization tools of a institution must be considered as an integrated system, pursuing a unifying model.
    Date
    16. 3.2008 18:22:50
    Source
    Kompatibilität, Medien und Ethik in der Wissensorganisation - Compatibility, Media and Ethics in Knowledge Organization: Proceedings der 10. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation Wien, 3.-5. Juli 2006 - Proceedings of the 10th Conference of the German Section of the International Society of Knowledge Organization Vienna, 3-5 July 2006. Ed.: H.P. Ohly, S. Netscher u. K. Mitgutsch
  4. Jacobs, J.-H.; Mengel, T.; Müller, K.: Insights and Outlooks : a retrospective view on the CrissCross project (2011) 0.06
    0.059802245 = product of:
      0.10251814 = sum of:
        0.01438466 = product of:
          0.04315398 = sum of:
            0.04315398 = weight(_text_:f in 4785) [ClassicSimilarity], result of:
              0.04315398 = score(doc=4785,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.3082599 = fieldWeight in 4785, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4785)
          0.33333334 = coord(1/3)
        0.008371122 = weight(_text_:information in 4785) [ClassicSimilarity], result of:
          0.008371122 = score(doc=4785,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.13576832 = fieldWeight in 4785, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4785)
        0.013343699 = weight(_text_:und in 4785) [ClassicSimilarity], result of:
          0.013343699 = score(doc=4785,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.17141339 = fieldWeight in 4785, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4785)
        0.009575742 = weight(_text_:for in 4785) [ClassicSimilarity], result of:
          0.009575742 = score(doc=4785,freq=2.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.14520876 = fieldWeight in 4785, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4785)
        0.020286093 = weight(_text_:the in 4785) [ClassicSimilarity], result of:
          0.020286093 = score(doc=4785,freq=18.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36607194 = fieldWeight in 4785, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4785)
        0.016270736 = weight(_text_:of in 4785) [ClassicSimilarity], result of:
          0.016270736 = score(doc=4785,freq=12.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.29624295 = fieldWeight in 4785, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4785)
        0.020286093 = weight(_text_:the in 4785) [ClassicSimilarity], result of:
          0.020286093 = score(doc=4785,freq=18.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36607194 = fieldWeight in 4785, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4785)
      0.5833333 = coord(7/12)
    
    Abstract
    This paper discusses goals, methods and benefits of the conceptual mapping approach developed by the CrissCross project, in the framework of which the topical headings of the German subject headings authority file Schlagwortnormdatei (SWD) have been mapped to notations of the Dewey Decimal Classification (DDC). Projectspecific retrieval concepts for improving thematic access in heterogeneous information spaces are outlined and explained on the basis of significant examples.
    Series
    Bibliotheca Academica - Reihe Informations- und Bibliothekswissenschaften; Bd. 1
    Source
    Concepts in context: Proceedings of the Cologne Conference on Interoperability and Semantics in Knowledge Organization July 19th - 20th, 2010. Eds.: F. Boteram, W. Gödert u. J. Hubrich
  5. Krause, J.: Shell Model, Semantic Web and Web Information Retrieval (2006) 0.06
    0.05951997 = product of:
      0.11903994 = sum of:
        0.016912218 = weight(_text_:information in 6061) [ClassicSimilarity], result of:
          0.016912218 = score(doc=6061,freq=16.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.27429342 = fieldWeight in 6061, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6061)
        0.016508548 = weight(_text_:und in 6061) [ClassicSimilarity], result of:
          0.016508548 = score(doc=6061,freq=6.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.21206908 = fieldWeight in 6061, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6061)
        0.015294294 = weight(_text_:for in 6061) [ClassicSimilarity], result of:
          0.015294294 = score(doc=6061,freq=10.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.2319262 = fieldWeight in 6061, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6061)
        0.025097532 = weight(_text_:the in 6061) [ClassicSimilarity], result of:
          0.025097532 = score(doc=6061,freq=54.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.4528966 = fieldWeight in 6061, product of:
              7.3484693 = tf(freq=54.0), with freq of:
                54.0 = termFreq=54.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6061)
        0.020129817 = weight(_text_:of in 6061) [ClassicSimilarity], result of:
          0.020129817 = score(doc=6061,freq=36.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.36650562 = fieldWeight in 6061, product of:
              6.0 = tf(freq=36.0), with freq of:
                36.0 = termFreq=36.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6061)
        0.025097532 = weight(_text_:the in 6061) [ClassicSimilarity], result of:
          0.025097532 = score(doc=6061,freq=54.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.4528966 = fieldWeight in 6061, product of:
              7.3484693 = tf(freq=54.0), with freq of:
                54.0 = termFreq=54.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6061)
      0.5 = coord(6/12)
    
    Abstract
    The middle of the 1990s are coined by the increased enthusiasm for the possibilities of the WWW, which has only recently deviated - at least in relation to scientific information - for the differentiated measuring of its advantages and disadvantages. Web Information Retrieval originated as a specialized discipline with great commercial significance (for an overview see Lewandowski 2005). Besides the new technological structure that enables the indexing and searching (in seconds) of unimaginable amounts of data worldwide, new assessment processes for the ranking of search results are being developed, which use the link structures of the Web. They are the main innovation with respect to the traditional "mother discipline" of Information Retrieval. From the beginning, link structures of Web pages are applied to commercial search engines in a wide array of variations. From the perspective of scientific information, link topology based approaches were in essence trying to solve a self-created problem: on the one hand, it quickly became clear that the openness of the Web led to an up-tonow unknown increase in available information, but this also caused the quality of the Web pages searched to become a problem - and with it the relevance of the results. The gatekeeper function of traditional information providers, which narrows down every user query to focus on high-quality sources was lacking. Therefore, the recognition of the "authoritativeness" of the Web pages by general search engines such as Google was one of the most important factors for their success.
    Source
    Information und Sprache: Beiträge zu Informationswissenschaft, Computerlinguistik, Bibliothekswesen und verwandten Fächern. Festschrift für Harald H. Zimmermann. Herausgegeben von Ilse Harms, Heinz-Dirk Luckhardt und Hans W. Giessen
  6. Concepts in Context : Proceedings of the Cologne Conference on Interoperability and Semantics in Knowledge Organization July 19th - 20th, 2010 (2011) 0.06
    0.056686457 = product of:
      0.085029684 = sum of:
        0.014530702 = product of:
          0.043592103 = sum of:
            0.043592103 = weight(_text_:f in 628) [ClassicSimilarity], result of:
              0.043592103 = score(doc=628,freq=4.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.31138954 = fieldWeight in 628, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=628)
          0.33333334 = coord(1/3)
        0.008456109 = weight(_text_:information in 628) [ClassicSimilarity], result of:
          0.008456109 = score(doc=628,freq=4.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.13714671 = fieldWeight in 628, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=628)
        0.009531214 = weight(_text_:und in 628) [ClassicSimilarity], result of:
          0.009531214 = score(doc=628,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.12243814 = fieldWeight in 628, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=628)
        0.00967296 = weight(_text_:for in 628) [ClassicSimilarity], result of:
          0.00967296 = score(doc=628,freq=4.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.14668301 = fieldWeight in 628, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0390625 = fieldNorm(doc=628)
        0.009660044 = weight(_text_:the in 628) [ClassicSimilarity], result of:
          0.009660044 = score(doc=628,freq=8.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.17431997 = fieldWeight in 628, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=628)
        0.011621956 = weight(_text_:of in 628) [ClassicSimilarity], result of:
          0.011621956 = score(doc=628,freq=12.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.21160212 = fieldWeight in 628, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=628)
        0.009660044 = weight(_text_:the in 628) [ClassicSimilarity], result of:
          0.009660044 = score(doc=628,freq=8.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.17431997 = fieldWeight in 628, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=628)
        0.011896656 = product of:
          0.023793312 = sum of:
            0.023793312 = weight(_text_:22 in 628) [ClassicSimilarity], result of:
              0.023793312 = score(doc=628,freq=2.0), product of:
                0.12299426 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035122856 = queryNorm
                0.19345059 = fieldWeight in 628, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=628)
          0.5 = coord(1/2)
      0.6666667 = coord(8/12)
    
    Content
    Winfried Gödert: Programmatic Issues and Introduction - Dagobert Soergel: Conceptual Foundations for Semantic Mapping and Semantic Search - Jan-Helge Jacobs, Tina Mengel, Katrin Müller: Insights and Outlooks: A Retrospective View on the CrissCross Project - Yvonne Jahns, Helga Karg: Translingual Retrieval: Moving between Vocabularies - MACS 2010 - Jessica Hubrich: Intersystem Relations: Characteristics and Functionalities - Stella G Dextre Clarke: In Pursuit of Interoperability: Can We Standardize Mapping Types? - Philipp Mayr, Philipp Schaer, Peter Mutschke: A Science Model Driven Retrieval Prototype - Claudia Effenberger, Julia Hauser: Would an Explicit Versioning of the DDC Bring Advantages for Retrieval? - Gordon Dunsire: Interoperability and Semantics in RDF Representations of FRBR, FRAD and FRSAD - Maja Zumer: FRSAD: Challenges of Modeling the Aboutness - Michael Panzer: Two Tales of a Concept: Aligning FRSAD with SKOS - Felix Boteram: Integrating Semantic Interoperability into FRSAD
    Date
    22. 2.2013 11:34:18
    Editor
    Boteram, F.
    Footnote
    Rez. in: ZfBB 59(2012) H.2, S.119-120 (F. Förster); KO 39(2012) no.6, S.461-463 (I. Dahlberg)
    RSWK
    Wissensorganisation / Information Retrieval / Kongress / Köln <2010>
    Series
    Bibliotheca Academica - Reihe Informations- und Bibliothekswissenschaften; Bd. 1
    Subject
    Wissensorganisation / Information Retrieval / Kongress / Köln <2010>
  7. Mayr, P.; Schaer, P.; Mutschke, P.: ¬A science model driven retrieval prototype (2011) 0.05
    0.053400096 = product of:
      0.091543026 = sum of:
        0.012329709 = product of:
          0.036989126 = sum of:
            0.036989126 = weight(_text_:f in 649) [ClassicSimilarity], result of:
              0.036989126 = score(doc=649,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.26422277 = fieldWeight in 649, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.046875 = fieldNorm(doc=649)
          0.33333334 = coord(1/3)
        0.007175247 = weight(_text_:information in 649) [ClassicSimilarity], result of:
          0.007175247 = score(doc=649,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.116372846 = fieldWeight in 649, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=649)
        0.011437457 = weight(_text_:und in 649) [ClassicSimilarity], result of:
          0.011437457 = score(doc=649,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.14692576 = fieldWeight in 649, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=649)
        0.008207779 = weight(_text_:for in 649) [ClassicSimilarity], result of:
          0.008207779 = score(doc=649,freq=2.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.12446466 = fieldWeight in 649, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=649)
        0.019223245 = weight(_text_:the in 649) [ClassicSimilarity], result of:
          0.019223245 = score(doc=649,freq=22.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.34689236 = fieldWeight in 649, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=649)
        0.013946345 = weight(_text_:of in 649) [ClassicSimilarity], result of:
          0.013946345 = score(doc=649,freq=12.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.25392252 = fieldWeight in 649, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=649)
        0.019223245 = weight(_text_:the in 649) [ClassicSimilarity], result of:
          0.019223245 = score(doc=649,freq=22.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.34689236 = fieldWeight in 649, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=649)
      0.5833333 = coord(7/12)
    
    Abstract
    This paper is about a better understanding of the structure and dynamics of science and the usage of these insights for compensating the typical problems that arises in metadata-driven Digital Libraries. Three science model driven retrieval services are presented: co-word analysis based query expansion, re-ranking via Bradfordizing and author centrality. The services are evaluated with relevance assessments from which two important implications emerge: (1) precision values of the retrieval services are the same or better than the tf-idf retrieval baseline and (2) each service retrieved a disjoint set of documents. The different services each favor quite other - but still relevant - documents than pure term-frequency based rankings. The proposed models and derived retrieval services therefore open up new viewpoints on the scientific knowledge space and provide an alternative framework to structure scholarly information systems.
    Series
    Bibliotheca Academica - Reihe Informations- und Bibliothekswissenschaften; Bd. 1
    Source
    Concepts in context: Proceedings of the Cologne Conference on Interoperability and Semantics in Knowledge Organization July 19th - 20th, 2010. Eds.: F. Boteram, W. Gödert u. J. Hubrich
  8. Boteram, F.: Integrating semantic interoperability into FRSAD (2011) 0.05
    0.053051956 = product of:
      0.10610391 = sum of:
        0.017436842 = product of:
          0.052310523 = sum of:
            0.052310523 = weight(_text_:f in 4791) [ClassicSimilarity], result of:
              0.052310523 = score(doc=4791,freq=4.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.37366742 = fieldWeight in 4791, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4791)
          0.33333334 = coord(1/3)
        0.011437457 = weight(_text_:und in 4791) [ClassicSimilarity], result of:
          0.011437457 = score(doc=4791,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.14692576 = fieldWeight in 4791, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=4791)
        0.018353151 = weight(_text_:for in 4791) [ClassicSimilarity], result of:
          0.018353151 = score(doc=4791,freq=10.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.27831143 = fieldWeight in 4791, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=4791)
        0.020897869 = weight(_text_:the in 4791) [ClassicSimilarity], result of:
          0.020897869 = score(doc=4791,freq=26.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.37711173 = fieldWeight in 4791, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=4791)
        0.017080715 = weight(_text_:of in 4791) [ClassicSimilarity], result of:
          0.017080715 = score(doc=4791,freq=18.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.3109903 = fieldWeight in 4791, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4791)
        0.020897869 = weight(_text_:the in 4791) [ClassicSimilarity], result of:
          0.020897869 = score(doc=4791,freq=26.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.37711173 = fieldWeight in 4791, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=4791)
      0.5 = coord(6/12)
    
    Abstract
    Since its launch in 1998, the initiative introducing the Functional Requirements of Bibliographic Records (FRBR) has provided and established a new standar for the representation of bibliographic data. Following the success of the Functional Requirements for Bibliographic Records, a number of related modules have been developed, complementing the initial standard. One of the most recent and promising efforts in this field is the initiative surrounding the development and approval of the Functional Requirements for Subject Authority Data (FRSAD). FRSAD is a specification of the requirements for modelling subject authority data, based on a conceptual model, which combines various theoretical aspects and technical modelling strategies. This article analyses the model's adaptability and appropriateness in an environment of heterogeneous and distributed systems and makes some suggestions for further improvements.
    Series
    Bibliotheca Academica - Reihe Informations- und Bibliothekswissenschaften; Bd. 1
    Source
    Concepts in context: Proceedings of the Cologne Conference on Interoperability and Semantics in Knowledge Organization July 19th - 20th, 2010. Eds.: F. Boteram, W. Gödert u. J. Hubrich
  9. García-Marco, F.-J.: Enhancing the visibility and relevance of thesauri in the Web : searching for a hub in the linked data environment (2016) 0.05
    0.05302265 = product of:
      0.09089597 = sum of:
        0.010274758 = product of:
          0.030824272 = sum of:
            0.030824272 = weight(_text_:f in 2916) [ClassicSimilarity], result of:
              0.030824272 = score(doc=2916,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.22018565 = fieldWeight in 2916, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2916)
          0.33333334 = coord(1/3)
        0.010356578 = weight(_text_:information in 2916) [ClassicSimilarity], result of:
          0.010356578 = score(doc=2916,freq=6.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.16796975 = fieldWeight in 2916, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2916)
        0.009531214 = weight(_text_:und in 2916) [ClassicSimilarity], result of:
          0.009531214 = score(doc=2916,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.12243814 = fieldWeight in 2916, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2916)
        0.015294294 = weight(_text_:for in 2916) [ClassicSimilarity], result of:
          0.015294294 = score(doc=2916,freq=10.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.2319262 = fieldWeight in 2916, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2916)
        0.017414892 = weight(_text_:the in 2916) [ClassicSimilarity], result of:
          0.017414892 = score(doc=2916,freq=26.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3142598 = fieldWeight in 2916, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2916)
        0.010609345 = weight(_text_:of in 2916) [ClassicSimilarity], result of:
          0.010609345 = score(doc=2916,freq=10.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.19316542 = fieldWeight in 2916, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2916)
        0.017414892 = weight(_text_:the in 2916) [ClassicSimilarity], result of:
          0.017414892 = score(doc=2916,freq=26.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3142598 = fieldWeight in 2916, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2916)
      0.5833333 = coord(7/12)
    
    Abstract
    Thesauri have triumphed in many domains that require precise and exhaustive information because of their representational power, their capability to integrate the concept-based and alphabetical approaches to organizing information, and their standardization and, more recently, formalization. Nevertheless, there is room to improve their relevance in the digital age by embracing the open linked data initiatives and by taking advantage of their structural and functional proximity to some of the big collaborative knowledge repositories in the Internet, notably the Wikipedia environment. With a focus on its implications for enhanced interoperability, this structural proximity is analysed, and the benefits of such collaboration for the different potential stakeholders are considered. It is proposed that better devices for ensuring semantic browsing are provided when necessary, and that an open hub for thesauri interconnection is developed, perhaps using existing big open Internet semantic facilities, such as Wikipedia.
    Content
    Beitrag in einem Special issue: The Great Debate: "This House Believes that the Traditional Thesaurus has no Place in Modern Information Retrieval." [19 February 2015, 14:00-17:30 preceded by ISKO UK AGM and followed by networking, wine and nibbles; vgl.: http://www.iskouk.org/content/great-debate].
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  10. Vetere, G.; Lenzerini, M.: Models for semantic interoperability in service-oriented architectures (2005) 0.05
    0.052682947 = product of:
      0.12643908 = sum of:
        0.06508181 = product of:
          0.19524543 = sum of:
            0.19524543 = weight(_text_:3a in 306) [ClassicSimilarity], result of:
              0.19524543 = score(doc=306,freq=2.0), product of:
                0.29777196 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.035122856 = queryNorm
                0.65568775 = fieldWeight in 306, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=306)
          0.33333334 = coord(1/3)
        0.013542145 = weight(_text_:for in 306) [ClassicSimilarity], result of:
          0.013542145 = score(doc=306,freq=4.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.20535621 = fieldWeight in 306, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
        0.01512036 = weight(_text_:the in 306) [ClassicSimilarity], result of:
          0.01512036 = score(doc=306,freq=10.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.2728539 = fieldWeight in 306, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
        0.017574405 = weight(_text_:of in 306) [ClassicSimilarity], result of:
          0.017574405 = score(doc=306,freq=14.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.31997898 = fieldWeight in 306, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
        0.01512036 = weight(_text_:the in 306) [ClassicSimilarity], result of:
          0.01512036 = score(doc=306,freq=10.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.2728539 = fieldWeight in 306, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
      0.41666666 = coord(5/12)
    
    Abstract
    Although service-oriented architectures go a long way toward providing interoperability in distributed, heterogeneous environments, managing semantic differences in such environments remains a challenge. We give an overview of the issue of semantic interoperability (integration), provide a semantic characterization of services, and discuss the role of ontologies. Then we analyze four basic models of semantic interoperability that differ in respect to their mapping between service descriptions and ontologies and in respect to where the evaluation of the integration logic is performed. We also provide some guidelines for selecting one of the possible interoperability models.
    Content
    Vgl.: http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=5386707&url=http%3A%2F%2Fieeexplore.ieee.org%2Fxpls%2Fabs_all.jsp%3Farnumber%3D5386707.
  11. Golub, K.; Tudhope, D.; Zeng, M.L.; Zumer, M.: Terminology registries for knowledge organization systems : functionality, use, and attributes (2014) 0.05
    0.052520268 = product of:
      0.105040535 = sum of:
        0.010147331 = weight(_text_:information in 1347) [ClassicSimilarity], result of:
          0.010147331 = score(doc=1347,freq=4.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.16457605 = fieldWeight in 1347, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1347)
        0.025955278 = weight(_text_:for in 1347) [ClassicSimilarity], result of:
          0.025955278 = score(doc=1347,freq=20.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.39359182 = fieldWeight in 1347, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=1347)
        0.018328644 = weight(_text_:the in 1347) [ClassicSimilarity], result of:
          0.018328644 = score(doc=1347,freq=20.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3307489 = fieldWeight in 1347, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=1347)
        0.018004656 = weight(_text_:of in 1347) [ClassicSimilarity], result of:
          0.018004656 = score(doc=1347,freq=20.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.32781258 = fieldWeight in 1347, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=1347)
        0.018328644 = weight(_text_:the in 1347) [ClassicSimilarity], result of:
          0.018328644 = score(doc=1347,freq=20.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3307489 = fieldWeight in 1347, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=1347)
        0.014275986 = product of:
          0.028551972 = sum of:
            0.028551972 = weight(_text_:22 in 1347) [ClassicSimilarity], result of:
              0.028551972 = score(doc=1347,freq=2.0), product of:
                0.12299426 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035122856 = queryNorm
                0.23214069 = fieldWeight in 1347, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1347)
          0.5 = coord(1/2)
      0.5 = coord(6/12)
    
    Abstract
    Terminology registries (TRs) are a crucial element of the infrastructure required for resource discovery services, digital libraries, Linked Data, and semantic interoperability generally. They can make the content of knowledge organization systems (KOS) available both for human and machine access. The paper describes the attributes and functionality for a TR, based on a review of published literature, existing TRs, and a survey of experts. A domain model based on user tasks is constructed and a set of core metadata elements for use in TRs is proposed. Ideally, the TR should allow searching as well as browsing for a KOS, matching a user's search while also providing information about existing terminology services, accessible to both humans and machines. The issues surrounding metadata for KOS are also discussed, together with the rationale for different aspects and the importance of a core set of KOS metadata for future machine-based access; a possible core set of metadata elements is proposed. This is dealt with in terms of practical experience and in relation to the Dublin Core Application Profile.
    Date
    22. 8.2014 17:12:54
    Source
    Journal of the Association for Information Science and Technology. 65(2014) no.9, S.1901-1916
  12. Sieglerschmidt, J.: Convergence of internet services in the cultural heritage sector : the long way to common vocabularies, metadata formats, ontologies (2008) 0.05
    0.05160397 = product of:
      0.10320794 = sum of:
        0.016044341 = weight(_text_:information in 1686) [ClassicSimilarity], result of:
          0.016044341 = score(doc=1686,freq=10.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.2602176 = fieldWeight in 1686, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1686)
        0.011437457 = weight(_text_:und in 1686) [ClassicSimilarity], result of:
          0.011437457 = score(doc=1686,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.14692576 = fieldWeight in 1686, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=1686)
        0.008207779 = weight(_text_:for in 1686) [ClassicSimilarity], result of:
          0.008207779 = score(doc=1686,freq=2.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.12446466 = fieldWeight in 1686, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=1686)
        0.023897627 = weight(_text_:the in 1686) [ClassicSimilarity], result of:
          0.023897627 = score(doc=1686,freq=34.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.43124378 = fieldWeight in 1686, product of:
              5.8309517 = tf(freq=34.0), with freq of:
                34.0 = termFreq=34.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=1686)
        0.019723112 = weight(_text_:of in 1686) [ClassicSimilarity], result of:
          0.019723112 = score(doc=1686,freq=24.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.3591007 = fieldWeight in 1686, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=1686)
        0.023897627 = weight(_text_:the in 1686) [ClassicSimilarity], result of:
          0.023897627 = score(doc=1686,freq=34.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.43124378 = fieldWeight in 1686, product of:
              5.8309517 = tf(freq=34.0), with freq of:
                34.0 = termFreq=34.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=1686)
      0.5 = coord(6/12)
    
    Abstract
    Since several years it has been observed that information offered by different knowledge producing institutions on the internet is more and more interlinked. This tendency will increase, because the fragmented information offers on the internet make the retrieval of information difficult as even impossible. At the same time the quantity of information offered on the internet grows exponentially in Europe - and elsewhere - due to many digitization projects. Insofar as funding institutions base the acceptance of projects on the observation of certain documentation standards the knowledge created will be retrievable and will remain so for a longer time. Otherwise the retrieval of information will become a matter of chance due to the limits of fragmented, knowledge producing social groups.
    Source
    Kompatibilität, Medien und Ethik in der Wissensorganisation - Compatibility, Media and Ethics in Knowledge Organization: Proceedings der 10. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation Wien, 3.-5. Juli 2006 - Proceedings of the 10th Conference of the German Section of the International Society of Knowledge Organization Vienna, 3-5 July 2006. Ed.: H.P. Ohly, S. Netscher u. K. Mitgutsch
  13. Wake, S.; Nicholson, D.: HILT: High-Level Thesaurus Project : building consensus for interoperable subject access across communities (2001) 0.05
    0.05150512 = product of:
      0.10301024 = sum of:
        0.008285262 = weight(_text_:information in 1224) [ClassicSimilarity], result of:
          0.008285262 = score(doc=1224,freq=6.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.1343758 = fieldWeight in 1224, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=1224)
        0.009477528 = weight(_text_:for in 1224) [ClassicSimilarity], result of:
          0.009477528 = score(doc=1224,freq=6.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.14371942 = fieldWeight in 1224, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.03125 = fieldNorm(doc=1224)
        0.020078024 = weight(_text_:the in 1224) [ClassicSimilarity], result of:
          0.020078024 = score(doc=1224,freq=54.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36231726 = fieldWeight in 1224, product of:
              7.3484693 = tf(freq=54.0), with freq of:
                54.0 = termFreq=54.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.03125 = fieldNorm(doc=1224)
        0.014202263 = weight(_text_:of in 1224) [ClassicSimilarity], result of:
          0.014202263 = score(doc=1224,freq=28.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.25858206 = fieldWeight in 1224, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=1224)
        0.020078024 = weight(_text_:the in 1224) [ClassicSimilarity], result of:
          0.020078024 = score(doc=1224,freq=54.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36231726 = fieldWeight in 1224, product of:
              7.3484693 = tf(freq=54.0), with freq of:
                54.0 = termFreq=54.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.03125 = fieldNorm(doc=1224)
        0.030889137 = product of:
          0.061778273 = sum of:
            0.061778273 = weight(_text_:communities in 1224) [ClassicSimilarity], result of:
              0.061778273 = score(doc=1224,freq=4.0), product of:
                0.18632571 = queryWeight, product of:
                  5.3049703 = idf(docFreq=596, maxDocs=44218)
                  0.035122856 = queryNorm
                0.33156064 = fieldWeight in 1224, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.3049703 = idf(docFreq=596, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1224)
          0.5 = coord(1/2)
      0.5 = coord(6/12)
    
    Abstract
    This article provides an overview of the work carried out by the HILT Project <http://hilt.cdlr.strath.ac.uk> in making recommendations towards interoperable subject access, or cross-searching and browsing distributed services amongst the archives, libraries, museums and electronic services sectors. The article details consensus achieved at the 19 June 2001 HILT Workshop and discusses the HILT Stakeholder Survey. In 1999 Péter Jascó wrote that "savvy searchers" are asking for direction. Three years later the scenario he describes, that of searchers cross-searching databases where the subject vocabulary used in each case is different, still rings true. Jascó states that, in many cases, databases do not offer the necessary aids required to use the "preferred terms of the subject-controlled vocabulary". The databases to which Jascó refers are Dialog and DataStar. However, the situation he describes applies as well to the area that HILT is researching: that of cross-searching and browsing by subject across databases and catalogues in archives, libraries, museums and online information services. So how does a user access information on a particular subject when it is indexed across a multitude of services under different, but quite often similar, subject terms? Also, if experienced searchers are having problems, what about novice searchers? As information professionals, it is our role to investigate such problems and recommend solutions. Although there is no hard empirical evidence one way or another, HILT participants agree that the problem for users attempting to search across databases is real. There is a strong likelihood that users are disadvantaged by the use of different subject terminology combined with a multitude of different practices taking place within the archive, library, museums and online communities. Arguably, failure to address this problem of interoperability undermines the value of cross-searching and browsing facilities, and wastes public money because relevant resources are 'hidden' from searchers. HILT is charged with analysing this broad problem through qualitative methods, with the main aim of presenting a set of recommendations on how to make it easier to cross-search and browse distributed services. Because this is a very large problem composed of many strands, HILT recognizes that any proposed solutions must address a host of issues. Recommended solutions must be affordable, sustainable, politically acceptable, useful, future-proof and international in scope. It also became clear to the HILT team that progress toward finding solutions to the interoperability problem could only be achieved through direct dialogue with other parties keen to solve this problem, and that the problem was as much about consensus building as it was about finding a solution. This article describes how HILT approached the cross-searching problem; how it investigated the nature of the problem, detailing results from the HILT Stakeholder Survey; and how it achieved consensus through the recent HILT Workshop.
  14. Kalfoglou, Y.; Schorlemmer, M.: Ontology mapping : the state of the art 0.05
    0.050692808 = product of:
      0.101385616 = sum of:
        0.007175247 = weight(_text_:information in 171) [ClassicSimilarity], result of:
          0.007175247 = score(doc=171,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.116372846 = fieldWeight in 171, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=171)
        0.011607553 = weight(_text_:for in 171) [ClassicSimilarity], result of:
          0.011607553 = score(doc=171,freq=4.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.17601961 = fieldWeight in 171, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=171)
        0.01738808 = weight(_text_:the in 171) [ClassicSimilarity], result of:
          0.01738808 = score(doc=171,freq=18.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.31377596 = fieldWeight in 171, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=171)
        0.015063776 = weight(_text_:of in 171) [ClassicSimilarity], result of:
          0.015063776 = score(doc=171,freq=14.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.2742677 = fieldWeight in 171, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=171)
        0.01738808 = weight(_text_:the in 171) [ClassicSimilarity], result of:
          0.01738808 = score(doc=171,freq=18.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.31377596 = fieldWeight in 171, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=171)
        0.032762878 = product of:
          0.065525755 = sum of:
            0.065525755 = weight(_text_:communities in 171) [ClassicSimilarity], result of:
              0.065525755 = score(doc=171,freq=2.0), product of:
                0.18632571 = queryWeight, product of:
                  5.3049703 = idf(docFreq=596, maxDocs=44218)
                  0.035122856 = queryNorm
                0.35167316 = fieldWeight in 171, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.3049703 = idf(docFreq=596, maxDocs=44218)
                  0.046875 = fieldNorm(doc=171)
          0.5 = coord(1/2)
      0.5 = coord(6/12)
    
    Abstract
    Ontology mapping is seen as a solution provider in today's landscape of ontology research. As the number of ontologies that are made publicly available and accessible on the Web increases steadily, so does the need for applications to use them. A single ontology is no longer enough to support the tasks envisaged by a distributed environment like the Semantic Web. Multiple ontologies need to be accessed from several applications. Mapping could provide a common layer from which several ontologies could be accessed and hence could exchange information in semantically sound manners. Developing such mappings has been the focus of a variety of works originating from diverse communities over a number of years. In this article we comprehensively review and present these works. We also provide insights on the pragmatics of ontology mapping and elaborate on a theoretical approach for defining ontology mapping.
  15. Stempfhuber, M.; Zapilko, B.: Modelling text-fact-integration in digital libraries (2009) 0.05
    0.05066234 = product of:
      0.10132468 = sum of:
        0.016044341 = weight(_text_:information in 3393) [ClassicSimilarity], result of:
          0.016044341 = score(doc=3393,freq=10.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.2602176 = fieldWeight in 3393, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3393)
        0.011437457 = weight(_text_:und in 3393) [ClassicSimilarity], result of:
          0.011437457 = score(doc=3393,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.14692576 = fieldWeight in 3393, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=3393)
        0.02010487 = weight(_text_:for in 3393) [ClassicSimilarity], result of:
          0.02010487 = score(doc=3393,freq=12.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.3048749 = fieldWeight in 3393, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=3393)
        0.018328644 = weight(_text_:the in 3393) [ClassicSimilarity], result of:
          0.018328644 = score(doc=3393,freq=20.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3307489 = fieldWeight in 3393, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=3393)
        0.017080715 = weight(_text_:of in 3393) [ClassicSimilarity], result of:
          0.017080715 = score(doc=3393,freq=18.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.3109903 = fieldWeight in 3393, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=3393)
        0.018328644 = weight(_text_:the in 3393) [ClassicSimilarity], result of:
          0.018328644 = score(doc=3393,freq=20.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3307489 = fieldWeight in 3393, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=3393)
      0.5 = coord(6/12)
    
    Abstract
    Digital Libraries currently face the challenge of integrating many different types of research information (e.g. publications, primary data, expert's profiles, institutional profiles, project information etc.) according to their scientific users' needs. To date no general, integrated model for knowledge organization and retrieval in Digital Libraries exists. This causes the problem of structural and semantic heterogeneity due to the wide range of metadata standards, indexing vocabularies and indexing approaches used for different types of information. The research presented in this paper focuses on areas in which activities are being undertaken in the field of Digital Libraries in order to treat semantic interoperability problems. We present a model for the integrated retrieval of factual and textual data which combines multiple approaches to semantic interoperability und sets them into context. Embedded in the research cycle, traditional content indexing methods for publications meet the newer, but rarely used ontology-based approaches which seem to be better suited for representing complex information like the one contained in survey data. The benefits of our model are (1) easy re-use of available knowledge organisation systems and (2) reduced efforts for domain modelling with ontologies.
    Theme
    Information Gateway
  16. Dunsire, G.; Willer, M.: Initiatives to make standard library metadata models and structures available to the Semantic Web (2010) 0.05
    0.048973136 = product of:
      0.09794627 = sum of:
        0.009566996 = weight(_text_:information in 3965) [ClassicSimilarity], result of:
          0.009566996 = score(doc=3965,freq=8.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.1551638 = fieldWeight in 3965, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=3965)
        0.010943705 = weight(_text_:for in 3965) [ClassicSimilarity], result of:
          0.010943705 = score(doc=3965,freq=8.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.16595288 = fieldWeight in 3965, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.03125 = fieldNorm(doc=3965)
        0.020446459 = weight(_text_:the in 3965) [ClassicSimilarity], result of:
          0.020446459 = score(doc=3965,freq=56.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36896583 = fieldWeight in 3965, product of:
              7.483315 = tf(freq=56.0), with freq of:
                56.0 = termFreq=56.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.03125 = fieldNorm(doc=3965)
        0.014700741 = weight(_text_:of in 3965) [ClassicSimilarity], result of:
          0.014700741 = score(doc=3965,freq=30.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.26765788 = fieldWeight in 3965, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=3965)
        0.020446459 = weight(_text_:the in 3965) [ClassicSimilarity], result of:
          0.020446459 = score(doc=3965,freq=56.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36896583 = fieldWeight in 3965, product of:
              7.483315 = tf(freq=56.0), with freq of:
                56.0 = termFreq=56.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.03125 = fieldNorm(doc=3965)
        0.021841917 = product of:
          0.043683834 = sum of:
            0.043683834 = weight(_text_:communities in 3965) [ClassicSimilarity], result of:
              0.043683834 = score(doc=3965,freq=2.0), product of:
                0.18632571 = queryWeight, product of:
                  5.3049703 = idf(docFreq=596, maxDocs=44218)
                  0.035122856 = queryNorm
                0.23444878 = fieldWeight in 3965, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.3049703 = idf(docFreq=596, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3965)
          0.5 = coord(1/2)
      0.5 = coord(6/12)
    
    Abstract
    This paper describes recent initiatives to make standard library metadata models and structures available to the Semantic Web, including IFLA standards such as Functional Requirements for Bibliographic Records (FRBR), Functional Requirements for Authority Data (FRAD), and International Standard Bibliographic Description (ISBD) along with the infrastructure that supports them. The FRBR Review Group is currently developing representations of FRAD and the entityrelationship model of FRBR in resource description framework (RDF) applications, using a combination of RDF, RDF Schema (RDFS), Simple Knowledge Organisation System (SKOS) and Web Ontology Language (OWL), cross-relating both models where appropriate. The ISBD/XML Task Group is investigating the representation of ISBD in RDF. The IFLA Namespaces project is developing an administrative and technical infrastructure to support such initiatives and encourage uptake of standards by other agencies. The paper describes similar initiatives with related external standards such as RDA - resource description and access, REICAT (the new Italian cataloguing rules) and CIDOC Conceptual Reference Model (CRM). The DCMI RDA Task Group is working with the Joint Steering Committee for RDA to develop Semantic Web representations of RDA structural elements, which are aligned with FRBR and FRAD, and controlled metadata content vocabularies. REICAT is also based on FRBR, and an object-oriented version of FRBR has been integrated with CRM, which itself has an RDF representation. CRM was initially based on the metadata needs of the museum community, and is now seeking extension to the archives community with the eventual aim of developing a model common to the main cultural information domains of archives, libraries and museums. The Vocabulary Mapping Framework (VMF) project has developed a Semantic Web tool to automatically generate mappings between metadata models from the information communities, including publishers. The tool is based on several standards, including CRM, FRAD, FRBR, MARC21 and RDA.
    The paper discusses the importance of these initiatives in releasing as linked data the very large quantities of rich, professionally-generated metadata stored in formats based on these standards, such as UNIMARC and MARC21, addressing such issues as critical mass for semantic and statistical inferencing, integration with user- and machine-generated metadata, and authenticity, veracity and trust. The paper also discusses related initiatives to release controlled vocabularies, including the Dewey Decimal Classification (DDC), ISBD, Library of Congress Name Authority File (LCNAF), Library of Congress Subject Headings (LCSH), Rameau (French subject headings), Universal Decimal Classification (UDC), and the Virtual International Authority File (VIAF) as linked data. Finally, the paper discusses the potential collective impact of these initiatives on metadata workflows and management systems.
    Content
    Vortrag im Rahmen der Session 93. Cataloguing der WORLD LIBRARY AND INFORMATION CONGRESS: 76TH IFLA GENERAL CONFERENCE AND ASSEMBLY, 10-15 August 2010, Gothenburg, Sweden - 149. Information Technology, Cataloguing, Classification and Indexing with Knowledge Management
  17. Dahlberg, I.: Kompatibilität und Integration : Probleme und Lösungen in der Wissensorganisation (2008) 0.05
    0.04879064 = product of:
      0.09758128 = sum of:
        0.005979372 = weight(_text_:information in 1677) [ClassicSimilarity], result of:
          0.005979372 = score(doc=1677,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.09697737 = fieldWeight in 1677, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1677)
        0.016508548 = weight(_text_:und in 1677) [ClassicSimilarity], result of:
          0.016508548 = score(doc=1677,freq=6.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.21206908 = fieldWeight in 1677, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1677)
        0.01184691 = weight(_text_:for in 1677) [ClassicSimilarity], result of:
          0.01184691 = score(doc=1677,freq=6.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.17964928 = fieldWeight in 1677, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1677)
        0.02213394 = weight(_text_:the in 1677) [ClassicSimilarity], result of:
          0.02213394 = score(doc=1677,freq=42.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.39941722 = fieldWeight in 1677, product of:
              6.4807405 = tf(freq=42.0), with freq of:
                42.0 = termFreq=42.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1677)
        0.018978573 = weight(_text_:of in 1677) [ClassicSimilarity], result of:
          0.018978573 = score(doc=1677,freq=32.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.34554482 = fieldWeight in 1677, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1677)
        0.02213394 = weight(_text_:the in 1677) [ClassicSimilarity], result of:
          0.02213394 = score(doc=1677,freq=42.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.39941722 = fieldWeight in 1677, product of:
              6.4807405 = tf(freq=42.0), with freq of:
                42.0 = termFreq=42.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1677)
      0.5 = coord(6/12)
    
    Abstract
    The trend in the fifties and sixties of the past century away from the use of universal classification systems such as the UDC towards establishing thesauri in special subject fields for the description of the conceptual contents of documents lead documentalists soon to realize that the necessary common tool for a collaboration among centers of similar subject fields was lacking. Therefore compatibility and integration studies began between the different thesauri of such fields, leading often to more comprehensive thesauri, such as macrothesauri. The paper describes this historic development and also the solutions found at the 1995 ISKO-Conference in Warsaw/Poland on Compatibility and Integration as given in its papers, its recommendations and also in the conceptual frame of its comprehensive bibliography on this topic. In conclusion a new solution is presented oriented toward combining the use of a universal classification system with the new developments of ontologies and their problem of interoperability and heterogeneity.
    Content
    Enthält im Anhang (S.48-49) die: "Recommendations of the Research Seminar on Compatibility and Integration of Order Systems organized by the International Society fpr Knowledge Organization (ISKO) and the Society for Professional Information (TIP), Warsaw, Poland, September 13-15, 1995".
    Source
    Kompatibilität, Medien und Ethik in der Wissensorganisation - Compatibility, Media and Ethics in Knowledge Organization: Proceedings der 10. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation Wien, 3.-5. Juli 2006 - Proceedings of the 10th Conference of the German Section of the International Society of Knowledge Organization Vienna, 3-5 July 2006. Ed.: H.P. Ohly, S. Netscher u. K. Mitgutsch
  18. Soergel, D.: Conceptual foundations for semantic mapping and semantic search (2011) 0.05
    0.047603056 = product of:
      0.08160524 = sum of:
        0.012329709 = product of:
          0.036989126 = sum of:
            0.036989126 = weight(_text_:f in 3939) [ClassicSimilarity], result of:
              0.036989126 = score(doc=3939,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.26422277 = fieldWeight in 3939, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3939)
          0.33333334 = coord(1/3)
        0.007175247 = weight(_text_:information in 3939) [ClassicSimilarity], result of:
          0.007175247 = score(doc=3939,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.116372846 = fieldWeight in 3939, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3939)
        0.011437457 = weight(_text_:und in 3939) [ClassicSimilarity], result of:
          0.011437457 = score(doc=3939,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.14692576 = fieldWeight in 3939, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=3939)
        0.014216291 = weight(_text_:for in 3939) [ClassicSimilarity], result of:
          0.014216291 = score(doc=3939,freq=6.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.21557912 = fieldWeight in 3939, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=3939)
        0.014197307 = weight(_text_:the in 3939) [ClassicSimilarity], result of:
          0.014197307 = score(doc=3939,freq=12.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.25619698 = fieldWeight in 3939, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=3939)
        0.008051926 = weight(_text_:of in 3939) [ClassicSimilarity], result of:
          0.008051926 = score(doc=3939,freq=4.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.14660224 = fieldWeight in 3939, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=3939)
        0.014197307 = weight(_text_:the in 3939) [ClassicSimilarity], result of:
          0.014197307 = score(doc=3939,freq=12.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.25619698 = fieldWeight in 3939, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=3939)
      0.5833333 = coord(7/12)
    
    Abstract
    This article proposes an approach to mapping between Knowledge Organization Systems (KOS), including ontologies, classifications, taxonomies, and thesauri and even natural languages, that is based on deep semantics. In this approach, concepts in each KOS are expressed through canonical expressions, such as description logic formulas, that combine atomic (or elemental) concepts drawn from a core classification. Relationships between concepts within or across KOS can then be derived by reasoning over the canonical expressions. The canonical expressions can also be used to provide a facet-based query formulation front-end for free-text search. The article illustrates this approach through many examples. It presents methods for the efficient construction of canonical expressions (linguistic analysis, exploiting information in the KOS and their hierarchies, and crowdsourcing) that make this approach feasible.
    Series
    Bibliotheca Academica - Reihe Informations- und Bibliothekswissenschaften; Bd. 1
    Source
    Concepts in context: Proceedings of the Cologne Conference on Interoperability and Semantics in Knowledge Organization July 19th - 20th, 2010. Eds.: F. Boteram, W. Gödert u. J. Hubrich
  19. Boteram, F.; Hubrich, J.: Specifying intersystem relations : requirements, strategies, and issues (2010) 0.05
    0.04648709 = product of:
      0.09297418 = sum of:
        0.012329709 = product of:
          0.036989126 = sum of:
            0.036989126 = weight(_text_:f in 3691) [ClassicSimilarity], result of:
              0.036989126 = score(doc=3691,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.26422277 = fieldWeight in 3691, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3691)
          0.33333334 = coord(1/3)
        0.008207779 = weight(_text_:for in 3691) [ClassicSimilarity], result of:
          0.008207779 = score(doc=3691,freq=2.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.12446466 = fieldWeight in 3691, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=3691)
        0.020078024 = weight(_text_:the in 3691) [ClassicSimilarity], result of:
          0.020078024 = score(doc=3691,freq=24.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36231726 = fieldWeight in 3691, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=3691)
        0.018004656 = weight(_text_:of in 3691) [ClassicSimilarity], result of:
          0.018004656 = score(doc=3691,freq=20.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.32781258 = fieldWeight in 3691, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=3691)
        0.020078024 = weight(_text_:the in 3691) [ClassicSimilarity], result of:
          0.020078024 = score(doc=3691,freq=24.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36231726 = fieldWeight in 3691, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=3691)
        0.014275986 = product of:
          0.028551972 = sum of:
            0.028551972 = weight(_text_:22 in 3691) [ClassicSimilarity], result of:
              0.028551972 = score(doc=3691,freq=2.0), product of:
                0.12299426 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035122856 = queryNorm
                0.23214069 = fieldWeight in 3691, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3691)
          0.5 = coord(1/2)
      0.5 = coord(6/12)
    
    Abstract
    Ideally, intersystem relations complement highly expressive and thoroughly structured relational indexing languages. The relational structures of the participating systems contribute to the meaning of the individual terms or classes. When conceptualizing mapping relations the structural and functional design of the respective systems must be fully taken into account. As intersystem relations may differ considerably from familiar interconcept relations, the creation of an adequate inventory that is general in coverage and specific in depth demands a deep understanding of the requirements and properties of mapping relations. The characteristics of specific mapping relations largely rely on the characteristics of the systems they are intended to connect. The detailed declaration of differences and peculiarities of specific mapping relations is an important prerequisite for modelling these relations. First approaches towards specifying
    Date
    22. 7.2010 17:11:51
  20. Doerr, M.: Semantic problems of thesaurus mapping (2001) 0.05
    0.046088874 = product of:
      0.09217775 = sum of:
        0.008456109 = weight(_text_:information in 5902) [ClassicSimilarity], result of:
          0.008456109 = score(doc=5902,freq=4.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.13714671 = fieldWeight in 5902, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5902)
        0.009531214 = weight(_text_:und in 5902) [ClassicSimilarity], result of:
          0.009531214 = score(doc=5902,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.12243814 = fieldWeight in 5902, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5902)
        0.013679632 = weight(_text_:for in 5902) [ClassicSimilarity], result of:
          0.013679632 = score(doc=5902,freq=8.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.20744109 = fieldWeight in 5902, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5902)
        0.01991469 = weight(_text_:the in 5902) [ClassicSimilarity], result of:
          0.01991469 = score(doc=5902,freq=34.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3593698 = fieldWeight in 5902, product of:
              5.8309517 = tf(freq=34.0), with freq of:
                34.0 = termFreq=34.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5902)
        0.02068142 = weight(_text_:of in 5902) [ClassicSimilarity], result of:
          0.02068142 = score(doc=5902,freq=38.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.37654874 = fieldWeight in 5902, product of:
              6.164414 = tf(freq=38.0), with freq of:
                38.0 = termFreq=38.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5902)
        0.01991469 = weight(_text_:the in 5902) [ClassicSimilarity], result of:
          0.01991469 = score(doc=5902,freq=34.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3593698 = fieldWeight in 5902, product of:
              5.8309517 = tf(freq=34.0), with freq of:
                34.0 = termFreq=34.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5902)
      0.5 = coord(6/12)
    
    Abstract
    With networked information access to heterogeneous data sources, the problem of terminology provision and interoperability of controlled vocabulary schemes such as thesauri becomes increasingly urgent. Solutions are needed to improve the performance of full-text retrieval systems and to guide the design of controlled terminology schemes for use in structured data, including metadata. Thesauri are created in different languages, with different scope and points of view and at different levels of abstraction and detail, to accomodate access to a specific group of collections. In any wider search accessing distributed collections, the user would like to start with familiar terminology and let the system find out the correspondences to other terminologies in order to retrieve equivalent results from all addressed collections. This paper investigates possible semantic differences that may hinder the unambiguous mapping and transition from one thesaurus to another. It focusses on the differences of meaning of terms and their relations as intended by their creators for indexing and querying a specific collection, in contrast to methods investigating the statistical relevance of terms for objects in a collection. It develops a notion of optimal mapping, paying particular attention to the intellectual quality of mappings between terms from different vocabularies and to problems of polysemy. Proposals are made to limit the vagueness introduced by the transition from one vocabulary to another. The paper shows ways in which thesaurus creators can improve their methodology to meet the challenges of networked access of distributed collections created under varying conditions. For system implementers, the discussion will lead to a better understanding of the complexity of the problem
    Source
    Journal of digital information. 1(2001) no.8,
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus

Years

Languages

  • e 229
  • d 81
  • no 1
  • pt 1
  • More… Less…

Types

  • a 215
  • el 96
  • m 18
  • x 11
  • r 9
  • s 7
  • n 2
  • p 2
  • More… Less…

Subjects