Search (7 results, page 1 of 1)

  • × author_ss:"Eckert, K."
  1. Eckert, K.; Schulz, A.: SABINE: OPAC oder opak? : kein Durchblick beim neuen Online Public Access Catalogue der Universität des Saarlandes (1995) 0.04
    0.041568942 = product of:
      0.06235341 = sum of:
        0.045045823 = weight(_text_:im in 2824) [ClassicSimilarity], result of:
          0.045045823 = score(doc=2824,freq=2.0), product of:
            0.1442303 = queryWeight, product of:
              2.8267863 = idf(docFreq=7115, maxDocs=44218)
              0.051022716 = queryNorm
            0.3123187 = fieldWeight in 2824, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.8267863 = idf(docFreq=7115, maxDocs=44218)
              0.078125 = fieldNorm(doc=2824)
        0.017307585 = product of:
          0.051922753 = sum of:
            0.051922753 = weight(_text_:online in 2824) [ClassicSimilarity], result of:
              0.051922753 = score(doc=2824,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.33531237 = fieldWeight in 2824, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2824)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Footnote
    Vgl. Erwiderung von: A. Heinen u. M. Müller: SABINE: eine Schönheit im Werden
  2. Eckert, K.: SKOS: eine Sprache für die Übertragung von Thesauri ins Semantic Web (2011) 0.04
    0.036313996 = product of:
      0.05447099 = sum of:
        0.03603666 = weight(_text_:im in 4331) [ClassicSimilarity], result of:
          0.03603666 = score(doc=4331,freq=2.0), product of:
            0.1442303 = queryWeight, product of:
              2.8267863 = idf(docFreq=7115, maxDocs=44218)
              0.051022716 = queryNorm
            0.24985497 = fieldWeight in 4331, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.8267863 = idf(docFreq=7115, maxDocs=44218)
              0.0625 = fieldNorm(doc=4331)
        0.01843433 = product of:
          0.055302992 = sum of:
            0.055302992 = weight(_text_:22 in 4331) [ClassicSimilarity], result of:
              0.055302992 = score(doc=4331,freq=2.0), product of:
                0.17867287 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051022716 = queryNorm
                0.30952093 = fieldWeight in 4331, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4331)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Abstract
    Das Semantic Web - bzw. Linked Data - hat das Potenzial, die Verfügbarkeit von Daten und Wissen, sowie den Zugriff darauf zu revolutionieren. Einen großen Beitrag dazu können Wissensorganisationssysteme wie Thesauri leisten, die die Daten inhaltlich erschließen und strukturieren. Leider sind immer noch viele dieser Systeme lediglich in Buchform oder in speziellen Anwendungen verfügbar. Wie also lassen sie sich für das Semantic Web nutzen? Das Simple Knowledge Organization System (SKOS) bietet eine Möglichkeit, die Wissensorganisationssysteme in eine Form zu "übersetzen", die im Web zitiert und mit anderen Resourcen verknüpft werden kann.
    Date
    15. 3.2011 19:21:22
  3. Eckert, K.; Albers, C.: Neue Dienstleistungsangebote wissenschaftlicher Bibliotheken in Europa (1995) 0.02
    0.021021385 = product of:
      0.06306415 = sum of:
        0.06306415 = weight(_text_:im in 3046) [ClassicSimilarity], result of:
          0.06306415 = score(doc=3046,freq=8.0), product of:
            0.1442303 = queryWeight, product of:
              2.8267863 = idf(docFreq=7115, maxDocs=44218)
              0.051022716 = queryNorm
            0.4372462 = fieldWeight in 3046, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.8267863 = idf(docFreq=7115, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3046)
      0.33333334 = coord(1/3)
    
    Abstract
    Die als LIB/2-Studie der EG-Kommission bekanntgewordene Untersuchung stellt eine umfassende Bestandsaufnahme des EDV-Einsatzes in den Öffentlichen und wissenschaftlichen Bibliotheken der Mitgliedsstaaten der Gemeinschaft dar. Sie wurde zweistufig in den Jahren 1986 bis 1993 durchgeführt und in Länderübersichten veröffentlicht. Im Frühjahr 1994 wurde im Auftrag des DBI eine Auswertung der LIB/2-Studie und ihres Updates im Hinblick auf das Angebot computerbasierter Dienstleistungen in den untersuchten wissenschaftlichen Bibliotheken vorgenommen. Die vorliegende Veröffentlichung ist eine Beschreibung des Entwicklungsstandes in Großbritannien, Irland, Dänemark, Niederlande, Deutschland, Portugal und Spanien. Ziel ist es, auf Grund der exemplarisch aufgezeigten Unterschiede, Stärken und Schwächen zu identifizieren, die ihrerseits wiederum für die Ableitung von Empfehlungen für die weitere Entwicklung im deutschen wissenschaftlichen Bibliothekswesen als Grundlage dienen können
  4. Eckert, K.: Linked Open Projects : Nachnutzung von Projektergebnissen als Linked Data (2010) 0.01
    0.010510692 = product of:
      0.031532075 = sum of:
        0.031532075 = weight(_text_:im in 4278) [ClassicSimilarity], result of:
          0.031532075 = score(doc=4278,freq=2.0), product of:
            0.1442303 = queryWeight, product of:
              2.8267863 = idf(docFreq=7115, maxDocs=44218)
              0.051022716 = queryNorm
            0.2186231 = fieldWeight in 4278, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.8267863 = idf(docFreq=7115, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4278)
      0.33333334 = coord(1/3)
    
    Abstract
    In vielen wissenschaftlichen Projekten - nicht nur im Bibliotheksbereich - geht es um die Erzeugung von Daten, häufig mit Hilfe automatischer Verfahren. Die Nachnutzung dieser Daten gestaltet sich häufig schwierig. In diesem Artikel werden wissenschaftliche Projekte beschrieben, die an der Universitätsbibliothek Mannheim durchgeführt wurden und werden. Anhand einfacher Beispiele wird gezeigt, wie durch Linked Data die Daten, die in diesen Projekten generiert werden, leicht und flexibel nachgenutzt werden können.
  5. Eckert, K.: Thesaurus analysis and visualization in semantic search applications (2007) 0.00
    0.004052635 = product of:
      0.012157904 = sum of:
        0.012157904 = product of:
          0.03647371 = sum of:
            0.03647371 = weight(_text_:retrieval in 3222) [ClassicSimilarity], result of:
              0.03647371 = score(doc=3222,freq=4.0), product of:
                0.15433937 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.051022716 = queryNorm
                0.23632148 = fieldWeight in 3222, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3222)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    The use of thesaurus-based indexing is a common approach for increasing the performance of information retrieval. In this thesis, we examine the suitability of a thesaurus for a given set of information and evaluate improvements of existing thesauri to get better search results. On this area, we focus on two aspects: 1. We demonstrate an analysis of the indexing results achieved by an automatic document indexer and the involved thesaurus. 2. We propose a method for thesaurus evaluation which is based on a combination of statistical measures and appropriate visualization techniques that support the detection of potential problems in a thesaurus. In this chapter, we give an overview of the context of our work. Next, we briefly outline the basics of thesaurus-based information retrieval and describe the Collexis Engine that was used for our experiments. In Chapter 3, we describe two experiments in automatically indexing documents in the areas of medicine and economics with corresponding thesauri and compare the results to available manual annotations. Chapter 4 describes methods for assessing thesauri and visualizing the result in terms of a treemap. We depict examples of interesting observations supported by the method and show that we actually find critical problems. We conclude with a discussion of open questions and future research in Chapter 5.
  6. Kempf, A.O.; Ritze, D.; Eckert, K.; Zapilko, B.: New ways of mapping knowledge organization systems : using a semi­automatic matching­procedure for building up vocabulary crosswalks (2013) 0.00
    0.0034615172 = product of:
      0.010384551 = sum of:
        0.010384551 = product of:
          0.031153653 = sum of:
            0.031153653 = weight(_text_:online in 989) [ClassicSimilarity], result of:
              0.031153653 = score(doc=989,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.20118743 = fieldWeight in 989, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.046875 = fieldNorm(doc=989)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    Crosswalks between different vocabularies are an indispensable prerequisite for integrated and high quality search scenarios in distributed data environments. Offered through the web and linked with each other they act as a central link so that users can move back and forth between different data sources available online. In the past, crosswalks between different thesauri have primarily been developed manually. In the long run the intellectual updating of such crosswalks requires huge personnel expenses. Therefore, an integration of automatic matching procedures, as for example Ontology Matching Tools, seems an obvious need. On the basis of computer generated correspondences between the Thesaurus for Economics (STW) and the Thesaurus for the Social Sciences (TheSoz) our contribution will explore cross-border approaches between IT-assisted tools and procedures on the one hand and external quality measurements via domain experts on the other hand. The techniques that emerge enable semi-automatically performed vocabulary crosswalks.
  7. Kempf, A.O.; Ritze, D.; Eckert, K.; Zapilko, B.: New ways of mapping knowledge organization systems : using a semi-automatic matching procedure for building up vocabulary crosswalks (2014) 0.00
    0.0028845975 = product of:
      0.008653793 = sum of:
        0.008653793 = product of:
          0.025961377 = sum of:
            0.025961377 = weight(_text_:online in 1371) [ClassicSimilarity], result of:
              0.025961377 = score(doc=1371,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.16765618 = fieldWeight in 1371, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1371)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    Crosswalks between different vocabularies are an indispensable prerequisite for integrated, high-quality search scenarios in distributed data environments where more than one controlled vocabulary is in use. Offered through the web and linked with each other they act as a central link so that users can move back and forth between different online data sources. In the past, crosswalks between different thesauri have usually been developed manually. In the long run the intellectual updating of such crosswalks is expensive. An obvious solution would be to apply automatic matching procedures, such as the so-called ontology matching tools. On the basis of computer-generated correspondences between the Thesaurus for the Social Sciences (TSS) and the Thesaurus for Economics (STW), our contribution explores the trade-off between IT-assisted tools and procedures on the one hand and external quality evaluation by domain experts on the other hand. This paper presents techniques for semi-automatic development and maintenance of vocabulary crosswalks. The performance of multiple matching tools was first evaluated against a reference set of correct mappings, then the tools were used to generate new mappings. It was concluded that the ontology matching tools can be used effectively to speed up the work of domain experts. By optimizing the workflow, the method promises to facilitate sustained updating of high-quality vocabulary crosswalks.