Search (14 results, page 1 of 1)

  • × author_ss:"Kempf, A.O."
  1. Kempf, A.O.: Thesauri und Interoperabilität mit anderen Vokabularen : Die neue Thesaurusnorm ISO 25964 (2013) 0.36
    0.3612946 = product of:
      0.43355352 = sum of:
        0.044417795 = weight(_text_:und in 1144) [ClassicSimilarity], result of:
          0.044417795 = score(doc=1144,freq=6.0), product of:
            0.104724824 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.04725067 = queryNorm
            0.42413816 = fieldWeight in 1144, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=1144)
        0.12236831 = weight(_text_:anwendung in 1144) [ClassicSimilarity], result of:
          0.12236831 = score(doc=1144,freq=2.0), product of:
            0.22876309 = queryWeight, product of:
              4.8414783 = idf(docFreq=948, maxDocs=44218)
              0.04725067 = queryNorm
            0.5349128 = fieldWeight in 1144, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.8414783 = idf(docFreq=948, maxDocs=44218)
              0.078125 = fieldNorm(doc=1144)
        0.040036436 = weight(_text_:des in 1144) [ClassicSimilarity], result of:
          0.040036436 = score(doc=1144,freq=2.0), product of:
            0.13085164 = queryWeight, product of:
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.04725067 = queryNorm
            0.30596817 = fieldWeight in 1144, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.078125 = fieldNorm(doc=1144)
        0.17099062 = weight(_text_:prinzips in 1144) [ClassicSimilarity], result of:
          0.17099062 = score(doc=1144,freq=2.0), product of:
            0.27041927 = queryWeight, product of:
              5.723078 = idf(docFreq=392, maxDocs=44218)
              0.04725067 = queryNorm
            0.6323167 = fieldWeight in 1144, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.723078 = idf(docFreq=392, maxDocs=44218)
              0.078125 = fieldNorm(doc=1144)
        0.055740345 = product of:
          0.11148069 = sum of:
            0.11148069 = weight(_text_:thesaurus in 1144) [ClassicSimilarity], result of:
              0.11148069 = score(doc=1144,freq=2.0), product of:
                0.21834905 = queryWeight, product of:
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.04725067 = queryNorm
                0.5105618 = fieldWeight in 1144, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1144)
          0.5 = coord(1/2)
      0.8333333 = coord(5/6)
    
    Source
    Information - Wissenschaft und Praxis. 64(2013) H.6, S.365-368
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  2. Kempf, A.O.; Baum, K.: Thesaurus-based indexing of research data in the social sciences : opportunities and difficulties of internationalization efforts (2013) 0.24
    0.2418396 = product of:
      0.29020754 = sum of:
        0.015386774 = weight(_text_:und in 1656) [ClassicSimilarity], result of:
          0.015386774 = score(doc=1656,freq=2.0), product of:
            0.104724824 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.04725067 = queryNorm
            0.14692576 = fieldWeight in 1656, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=1656)
        0.07342099 = weight(_text_:anwendung in 1656) [ClassicSimilarity], result of:
          0.07342099 = score(doc=1656,freq=2.0), product of:
            0.22876309 = queryWeight, product of:
              4.8414783 = idf(docFreq=948, maxDocs=44218)
              0.04725067 = queryNorm
            0.3209477 = fieldWeight in 1656, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.8414783 = idf(docFreq=948, maxDocs=44218)
              0.046875 = fieldNorm(doc=1656)
        0.02402186 = weight(_text_:des in 1656) [ClassicSimilarity], result of:
          0.02402186 = score(doc=1656,freq=2.0), product of:
            0.13085164 = queryWeight, product of:
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.04725067 = queryNorm
            0.18358089 = fieldWeight in 1656, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.046875 = fieldNorm(doc=1656)
        0.10259437 = weight(_text_:prinzips in 1656) [ClassicSimilarity], result of:
          0.10259437 = score(doc=1656,freq=2.0), product of:
            0.27041927 = queryWeight, product of:
              5.723078 = idf(docFreq=392, maxDocs=44218)
              0.04725067 = queryNorm
            0.37939 = fieldWeight in 1656, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.723078 = idf(docFreq=392, maxDocs=44218)
              0.046875 = fieldNorm(doc=1656)
        0.07478353 = product of:
          0.14956705 = sum of:
            0.14956705 = weight(_text_:thesaurus in 1656) [ClassicSimilarity], result of:
              0.14956705 = score(doc=1656,freq=10.0), product of:
                0.21834905 = queryWeight, product of:
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.04725067 = queryNorm
                0.68499064 = fieldWeight in 1656, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1656)
          0.5 = coord(1/2)
      0.8333333 = coord(5/6)
    
    Abstract
    Efforts towards internationalization have become increasingly important in scientific environments. As for content-based indexing of scientific research data, however, standards leading to internationally coherent indexing which is vital for retrieval purposes are not yet sufficiently developed. Even concerning the concrete use of indexing instruments, launched by initiatives on an international scale, there are still no binding policies and guidelines. Against this backdrop, essential criteria which internationally applicable indexing systems should meet will be outlined. These will be illustrated through the multilingual European Language Social Science Thesaurus (ELSST), originally based on the UK Data Archive's (UKDA) Humanities and Social Science Electronic Thesaurus (HASSET) and ultimately developed by the Council of European Social Science Data Archives (CESSDA). Additionally, the general pros and cons of using international versus national indexing languages will be weighed using the ELSST and the Thesaurus for the Social Sciences (TSS) developed by GESIS - Leibniz-Institute for the Social Sciences. In this light, the benefit of vocabulary crosswalks for supporting a combined use of international and national indexing systems will be discussed.
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  3. Kempf, A.O.; Neubert, J.: ¬The role of thesauri in an Open Web : a case study of the STW Thesaurus for economics (2016) 0.24
    0.2418396 = product of:
      0.29020754 = sum of:
        0.015386774 = weight(_text_:und in 2912) [ClassicSimilarity], result of:
          0.015386774 = score(doc=2912,freq=2.0), product of:
            0.104724824 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.04725067 = queryNorm
            0.14692576 = fieldWeight in 2912, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=2912)
        0.07342099 = weight(_text_:anwendung in 2912) [ClassicSimilarity], result of:
          0.07342099 = score(doc=2912,freq=2.0), product of:
            0.22876309 = queryWeight, product of:
              4.8414783 = idf(docFreq=948, maxDocs=44218)
              0.04725067 = queryNorm
            0.3209477 = fieldWeight in 2912, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.8414783 = idf(docFreq=948, maxDocs=44218)
              0.046875 = fieldNorm(doc=2912)
        0.02402186 = weight(_text_:des in 2912) [ClassicSimilarity], result of:
          0.02402186 = score(doc=2912,freq=2.0), product of:
            0.13085164 = queryWeight, product of:
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.04725067 = queryNorm
            0.18358089 = fieldWeight in 2912, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.046875 = fieldNorm(doc=2912)
        0.10259437 = weight(_text_:prinzips in 2912) [ClassicSimilarity], result of:
          0.10259437 = score(doc=2912,freq=2.0), product of:
            0.27041927 = queryWeight, product of:
              5.723078 = idf(docFreq=392, maxDocs=44218)
              0.04725067 = queryNorm
            0.37939 = fieldWeight in 2912, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.723078 = idf(docFreq=392, maxDocs=44218)
              0.046875 = fieldNorm(doc=2912)
        0.07478353 = product of:
          0.14956705 = sum of:
            0.14956705 = weight(_text_:thesaurus in 2912) [ClassicSimilarity], result of:
              0.14956705 = score(doc=2912,freq=10.0), product of:
                0.21834905 = queryWeight, product of:
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.04725067 = queryNorm
                0.68499064 = fieldWeight in 2912, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2912)
          0.5 = coord(1/2)
      0.8333333 = coord(5/6)
    
    Abstract
    This paper illustrates the changing role of thesauri interlinked with overall changes of modern information infrastructure services, referring to "STW Thesaurus for Economics" as a case study. It starts with an overview of the history and development of the STW and describes the far-reaching changes brought about by its publication on the Web, with regard to subject indexing, retrieval and new uses for Linked Open Data. It argues that only the most recent technological developments help thesauri to exploit their full potential which is why they more than ever have a place in current information retrieval and infrastructure.
    Content
    Beitrag in einem Special issue: The Great Debate: "This House Believes that the Traditional Thesaurus has no Place in Modern Information Retrieval." [19 February 2015, 14:00-17:30 preceded by ISKO UK AGM and followed by networking, wine and nibbles; vgl.: http://www.iskouk.org/content/great-debate].
    Object
    STW Thesaurus
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  4. Kempf, A.O.: Thesauri (2023) 0.23
    0.22597596 = product of:
      0.27117115 = sum of:
        0.037689745 = weight(_text_:und in 782) [ClassicSimilarity], result of:
          0.037689745 = score(doc=782,freq=12.0), product of:
            0.104724824 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.04725067 = queryNorm
            0.35989314 = fieldWeight in 782, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=782)
        0.07342099 = weight(_text_:anwendung in 782) [ClassicSimilarity], result of:
          0.07342099 = score(doc=782,freq=2.0), product of:
            0.22876309 = queryWeight, product of:
              4.8414783 = idf(docFreq=948, maxDocs=44218)
              0.04725067 = queryNorm
            0.3209477 = fieldWeight in 782, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.8414783 = idf(docFreq=948, maxDocs=44218)
              0.046875 = fieldNorm(doc=782)
        0.02402186 = weight(_text_:des in 782) [ClassicSimilarity], result of:
          0.02402186 = score(doc=782,freq=2.0), product of:
            0.13085164 = queryWeight, product of:
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.04725067 = queryNorm
            0.18358089 = fieldWeight in 782, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.046875 = fieldNorm(doc=782)
        0.10259437 = weight(_text_:prinzips in 782) [ClassicSimilarity], result of:
          0.10259437 = score(doc=782,freq=2.0), product of:
            0.27041927 = queryWeight, product of:
              5.723078 = idf(docFreq=392, maxDocs=44218)
              0.04725067 = queryNorm
            0.37939 = fieldWeight in 782, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.723078 = idf(docFreq=392, maxDocs=44218)
              0.046875 = fieldNorm(doc=782)
        0.03344421 = product of:
          0.06688842 = sum of:
            0.06688842 = weight(_text_:thesaurus in 782) [ClassicSimilarity], result of:
              0.06688842 = score(doc=782,freq=2.0), product of:
                0.21834905 = queryWeight, product of:
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.04725067 = queryNorm
                0.30633712 = fieldWeight in 782, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.046875 = fieldNorm(doc=782)
          0.5 = coord(1/2)
      0.8333333 = coord(5/6)
    
    Abstract
    In der Informationswissenschaft stehen Thesauri für kontrollierte und strukturierte Vokabulare, in denen Begriffe, verstanden als geistige Einheiten, durch Bezeichnungen der natürlichen Sprache repräsentiert werden. Jedem Begriff werden eine Vorzugsbenennung, ein sog. Deskriptor, der bei der Inhaltserschließung als Schlagwort vergeben wird, und weitere bedeutungsgleiche oder -ähnliche Bezeichnungen, sog. Nicht-Deskriptoren, als Zugangsvokabular bzw. alternative Sucheinstiege zugewiesen. Dieser Art werden der Variabilität und der Mehrdeutigkeit natürlicher Sprache Rechnung getragen. Darüber hinaus werden zwischen Begriffen bzw. ihren Bezeichnungen spezifische, reziproke Relationen kenntlich gemacht (s. Abschnitt 1), die die Bedeutungsbeziehungen bzw. das "semantische Gefüge" zwischen den Begriffen aufzeigen. Diese Kernprinzipien dieser Wissensorganisationsmethode dienen sowohl auf Seiten der Inhaltserschließenden als auch auf Seiten der in einem Informationssystem Recherchierenden in unterschiedlicher Weise der Benutzerführung und Suchunterstützung. Als Grundlage für semantisches Retrieval etwa sorgen sie bei automatischer Erweiterung der Suchanfrage um die hinterlegten Nicht-Deskriptoren für eine erfolgreiche Suche weitgehend unabhängig vom konkreten Suchterm.
    Source
    Grundlagen der Informationswissenschaft. Hrsg.: Rainer Kuhlen, Dirk Lewandowski, Wolfgang Semar und Christa Womser-Hacker. 7., völlig neu gefasste Ausg
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  5. Kempf, A.O.; Zapilko, B.: Normdatenpflege in Zeiten der Automatisierung : Erstellung und Evaluation automatisch aufgebauter Thesaurus-Crosskonkordanzen (2013) 0.14
    0.1350882 = product of:
      0.2701764 = sum of:
        0.030773548 = weight(_text_:und in 1021) [ClassicSimilarity], result of:
          0.030773548 = score(doc=1021,freq=8.0), product of:
            0.104724824 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.04725067 = queryNorm
            0.29385152 = fieldWeight in 1021, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=1021)
        0.02402186 = weight(_text_:des in 1021) [ClassicSimilarity], result of:
          0.02402186 = score(doc=1021,freq=2.0), product of:
            0.13085164 = queryWeight, product of:
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.04725067 = queryNorm
            0.18358089 = fieldWeight in 1021, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.046875 = fieldNorm(doc=1021)
        0.215381 = sum of:
          0.17697012 = weight(_text_:thesaurus in 1021) [ClassicSimilarity], result of:
            0.17697012 = score(doc=1021,freq=14.0), product of:
              0.21834905 = queryWeight, product of:
                4.6210785 = idf(docFreq=1182, maxDocs=44218)
                0.04725067 = queryNorm
              0.81049186 = fieldWeight in 1021, product of:
                3.7416575 = tf(freq=14.0), with freq of:
                  14.0 = termFreq=14.0
                4.6210785 = idf(docFreq=1182, maxDocs=44218)
                0.046875 = fieldNorm(doc=1021)
          0.03841088 = weight(_text_:22 in 1021) [ClassicSimilarity], result of:
            0.03841088 = score(doc=1021,freq=2.0), product of:
              0.16546379 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04725067 = queryNorm
              0.23214069 = fieldWeight in 1021, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=1021)
      0.5 = coord(3/6)
    
    Abstract
    Thesaurus-Crosskonkordanzen bilden eine wichtige Voraussetzung für die integrierte Suche in einer verteilten Datenstruktur. Ihr Aufbau erfordert allerdings erhebliche personelle Ressourcen. Der vorliegende Beitrag liefert Evaluationsergebnisse des Library Track 2012 der Ontology Alignment Evaluation Initiative (OAEI), in dem Crosskonkordanzen zwischen dem Thesaurus Sozialwissenschaften (TheSoz) und dem Standard Thesaurus Wirtschaft (STW) erstmals automatisch erstellt wurden. Die Evaluation weist auf deutliche Unterschiede in den getesteten Matching- Tools hin und stellt die qualitativen Unterschiede einer automatisch im Vergleich zu einer intellektuell erstellten Crosskonkordanz heraus. Die Ergebnisse sprechen für einen Einsatz automatisch generierter Thesaurus-Crosskonkordanzen, um Domänenexperten eine maschinell erzeugte Vorselektion von möglichen Äquivalenzrelationen anzubieten.
    Date
    18. 8.2013 12:53:22
    Object
    Thesaurus Sozialwissenschaften
    Standard Thesaurus Wirtschaft
    Source
    Information - Wissenschaft und Praxis. 64(2013) H.4, S.199-208
  6. Kempf, A.O.: Automatische Indexierung in der sozialwissenschaftlichen Fachinformation : eine Evaluationsstudie zur maschinellen Erschließung für die Datenbank SOLIS (2012) 0.05
    0.049087826 = product of:
      0.09817565 = sum of:
        0.04070958 = weight(_text_:und in 903) [ClassicSimilarity], result of:
          0.04070958 = score(doc=903,freq=14.0), product of:
            0.104724824 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.04725067 = queryNorm
            0.38872904 = fieldWeight in 903, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=903)
        0.02402186 = weight(_text_:des in 903) [ClassicSimilarity], result of:
          0.02402186 = score(doc=903,freq=2.0), product of:
            0.13085164 = queryWeight, product of:
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.04725067 = queryNorm
            0.18358089 = fieldWeight in 903, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.046875 = fieldNorm(doc=903)
        0.03344421 = product of:
          0.06688842 = sum of:
            0.06688842 = weight(_text_:thesaurus in 903) [ClassicSimilarity], result of:
              0.06688842 = score(doc=903,freq=2.0), product of:
                0.21834905 = queryWeight, product of:
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.04725067 = queryNorm
                0.30633712 = fieldWeight in 903, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.046875 = fieldNorm(doc=903)
          0.5 = coord(1/2)
      0.5 = coord(3/6)
    
    Abstract
    Automatische Indexierungsverfahren werden mit Zunahme der digitalen Verfügbarkeit von Metadaten und Volltexten mehr und mehr als eine mögliche Antwort auf das Management unstrukturierter Daten diskutiert. In der sozialwissenschaftlichen Fachinformation existiert in diesem Zusammenhang seit einiger Zeit der Vorschlag eines sogenannten Schalenmodells (vgl. Krause, 1996) mit unterschiedlichen Qualitätsstufen bei der inhaltlichen Erschließung. Vor diesem Hintergrund beschreibt die Arbeit zunächst Methoden und Verfahren der inhaltlichen und automatischen Indexierung, bevor vier Testläufe eines automatischen Indexierungssystems (MindServer) zur automatischen Erschließung von Datensätzen der bibliographischen Literaturdatenbank SOLIS mit Deskriptoren des Thesaurus Sozialwissenschaften sowie der Klassifikation Sozialwissenschaften beschrieben und analysiert werden. Es erfolgt eine ausführliche Fehleranalyse mit Beispielen sowie eine abschließende Diskussion, inwieweit die automatische Erschließung in dieser Form für die Randbereiche der Datenbank SOLIS für die Zukunft einen gangbaren Weg darstellt.
    Imprint
    Berlin : Humboldt-Universität zu Berlin, Philosophische Fakultät I, Institut für Bibliotheks- und Informationswissenschaft
    Series
    Berliner Handreichungen zur Bibliotheks- und Informationswissenschaft; 329
  7. Neubert, J.; Kempf, A.O.: Standard-Thesaurus Wirtschaft : nach Komplettüberarbeitung in Version 9.0 verfügbar (2015) 0.05
    0.04754608 = product of:
      0.14263824 = sum of:
        0.04804372 = weight(_text_:des in 2048) [ClassicSimilarity], result of:
          0.04804372 = score(doc=2048,freq=2.0), product of:
            0.13085164 = queryWeight, product of:
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.04725067 = queryNorm
            0.36716178 = fieldWeight in 2048, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.09375 = fieldNorm(doc=2048)
        0.09459451 = product of:
          0.18918902 = sum of:
            0.18918902 = weight(_text_:thesaurus in 2048) [ClassicSimilarity], result of:
              0.18918902 = score(doc=2048,freq=4.0), product of:
                0.21834905 = queryWeight, product of:
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.04725067 = queryNorm
                0.8664522 = fieldWeight in 2048, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2048)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Mitteilung an verschiedene Diskussionslisten zur Veröffentlichung der Version 9.0 des Standard-Thesaurus Wirtschaft (STW) durch die Deutsche Zentralbibliothek für Wirtschaftswissenschaften (ZBW) - Leibniz-Informationszentrum Wirtschaft.
  8. Toepfer, M.; Kempf, A.O.: Automatische Indexierung auf Basis von Titeln und Autoren-Keywords : ein Werkstattbericht (2016) 0.03
    0.02522712 = product of:
      0.07568136 = sum of:
        0.036266975 = weight(_text_:und in 3209) [ClassicSimilarity], result of:
          0.036266975 = score(doc=3209,freq=16.0), product of:
            0.104724824 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.04725067 = queryNorm
            0.34630734 = fieldWeight in 3209, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3209)
        0.03941438 = product of:
          0.07882876 = sum of:
            0.07882876 = weight(_text_:thesaurus in 3209) [ClassicSimilarity], result of:
              0.07882876 = score(doc=3209,freq=4.0), product of:
                0.21834905 = queryWeight, product of:
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.04725067 = queryNorm
                0.36102176 = fieldWeight in 3209, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3209)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Automatische Verfahren sind für Bibliotheken essentiell, um die Erschliessung stetig wachsender Datenmengen zu stemmen. Die Deutsche Zentralbibliothek für Wirtschaftswissenschaften - Leibniz-Informationszentrum Wirtschaft sammelt seit Längerem Erfahrungen im Bereich automatischer Indexierung und baut hier eigene Kompetenzen auf. Aufgrund rechtlicher Restriktionen werden unter anderem Ansätze untersucht, die ohne Volltextnutzung arbeiten. Dieser Beitrag gibt einen Einblick in ein laufendes Teilprojekt, das unter Verwendung von Titeln und Autoren-Keywords auf eine Nachnormierung der inhaltsbeschreibenden Metadaten auf den Standard-Thesaurus Wirtschaft (STW) abzielt. Wir erläutern den Hintergrund der Arbeit, betrachten die Systemarchitektur und stellen erste vielversprechende Ergebnisse eines dokumentenorientierten Verfahrens vor.
    Im Folgenden erläutern wir zunächst den Hintergrund der aktuellen Arbeit. Wir beziehen uns auf Erfahrungen mit maschinellen Verfahren allgemein und an der Deutschen Zentralbibliothek für Wirtschaftswissenschaften (ZBW) - Leibniz-Informationszentrum Wirtschaft im Speziellen. Im Anschluss geben wir einen konkreten Einblick in ein laufendes Teilprojekt, bei dem die Systemarchitektur der Automatik gegenüber früheren Arbeiten Titel und Autoren-Keywords gemeinsam verwendet, um eine Nachnormierung auf den Standard-Thesaurus Wirtschaft (STW) zu erzielen. Im Gegenssatz zu einer statischen Verknüpfung im Sinne einer Crosskonkordanz bzw. Vokabularabbildung ist das jetzt verfolgte Vorgehen dokumentenorientiert und damit in der Lage, kontextbezogene Zuordnungen vorzunehmen. Der Artikel stellt neben der Systemarchitektur auch erste experimentelle Ergebnisse vor, die im Vergleich zu titelbasierten Vorhersagen bereits deutliche Verbesserungen aufzeigen.
    Content
    Beitrag in einem Themenschwerpunkt 'Computerlinguistik und Bibliotheken'. Vgl.: http://0277.ch/ojs/index.php/cdrs_0277/article/view/156/354.
  9. Kempf, A.O.: Neue Verfahrenswege der Wissensorganisation : eine Evaluation automatischer Indexierung in der sozialwissenschaftlichen Fachinformation (2017) 0.03
    0.02517884 = product of:
      0.07553652 = sum of:
        0.035902474 = weight(_text_:und in 3497) [ClassicSimilarity], result of:
          0.035902474 = score(doc=3497,freq=8.0), product of:
            0.104724824 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.04725067 = queryNorm
            0.34282678 = fieldWeight in 3497, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3497)
        0.03963405 = weight(_text_:des in 3497) [ClassicSimilarity], result of:
          0.03963405 = score(doc=3497,freq=4.0), product of:
            0.13085164 = queryWeight, product of:
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.04725067 = queryNorm
            0.302893 = fieldWeight in 3497, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3497)
      0.33333334 = coord(2/6)
    
    Source
    Theorie, Semantik und Organisation von Wissen: Proceedings der 13. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) und dem 13. Internationalen Symposium der Informationswissenschaft der Higher Education Association for Information Science (HI) Potsdam (19.-20.03.2013): 'Theory, Information and Organization of Knowledge' / Proceedings der 14. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) und Natural Language & Information Systems (NLDB) Passau (16.06.2015): 'Lexical Resources for Knowledge Organization' / Proceedings des Workshops der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) auf der SEMANTICS Leipzig (1.09.2014): 'Knowledge Organization and Semantic Web' / Proceedings des Workshops der Polnischen und Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) Cottbus (29.-30.09.2011): 'Economics of Knowledge Production and Organization'. Hrsg. von W. Babik, H.P. Ohly u. K. Weber
  10. Kempf, A.O.: Automatische Inhaltserschließung in der Fachinformation (2013) 0.02
    0.017984888 = product of:
      0.05395466 = sum of:
        0.025644625 = weight(_text_:und in 905) [ClassicSimilarity], result of:
          0.025644625 = score(doc=905,freq=8.0), product of:
            0.104724824 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.04725067 = queryNorm
            0.24487628 = fieldWeight in 905, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=905)
        0.028310036 = weight(_text_:des in 905) [ClassicSimilarity], result of:
          0.028310036 = score(doc=905,freq=4.0), product of:
            0.13085164 = queryWeight, product of:
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.04725067 = queryNorm
            0.21635216 = fieldWeight in 905, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.7693076 = idf(docFreq=7536, maxDocs=44218)
              0.0390625 = fieldNorm(doc=905)
      0.33333334 = coord(2/6)
    
    Abstract
    Der Artikel basiert auf einer Masterarbeit mit dem Titel "Automatische Indexierung in der sozialwissenschaftlichen Fachinformation. Eine Evaluationsstudie zur maschinellen Erschließung für die Datenbank SOLIS" (Kempf 2012), die im Rahmen des Aufbaustudiengangs Bibliotheks- und Informationswissenschaft an der Humboldt- Universität zu Berlin am Lehrstuhl Information Retrieval verfasst wurde. Auf der Grundlage des Schalenmodells zur Inhaltserschließung in der Fachinformation stellt der Artikel Evaluationsergebnisse eines automatischen Erschließungsverfahrens für den Einsatz in der sozialwissenschaftlichen Fachinformation vor. Ausgehend von dem von Krause beschriebenen Anwendungsszenario, wonach SOLIS-Datenbestände (Sozialwissenschaftliches Literaturinformationssystem) von geringerer Relevanz automatisch erschlossen werden sollten, wurden auf dieser Dokumentgrundlage zwei Testreihen mit der Indexierungssoftware MindServer der Firma Recommind durchgeführt. Neben den Auswirkungen allgemeiner Systemeinstellungen in der ersten Testreihe wurde in der zweiten Testreihe die Indexierungsleistung der Software für die Rand- und die Kernbereiche der Literaturdatenbank miteinander verglichen. Für letztere Testreihe wurden für beide Bereiche der Datenbank spezifische Versionen der Indexierungssoftware aufgebaut, die anhand von Dokumentkorpora aus den entsprechenden Bereichen trainiert wurden. Die Ergebnisse der Evaluation, die auf der Grundlage intellektuell generierter Vergleichsdaten erfolgt, weisen auf Unterschiede in der Indexierungsleistung zwischen Rand- und Kernbereichen hin, die einerseits gegen den Einsatz automatischer Indexierungsverfahren in den Randbereichen sprechen. Andererseits deutet sich an, dass sich die Indexierungsresultate durch den Aufbau fachteilgebietsspezifischer Trainingsmengen verbessern lassen.
    Source
    Information - Wissenschaft und Praxis. 64(2013) H.2/3, S.96-106
  11. Kempf, A.O.; Ritze, D.; Eckert, K.; Zapilko, B.: New ways of mapping knowledge organization systems : using a semi­automatic matching­procedure for building up vocabulary crosswalks (2013) 0.01
    0.011148071 = product of:
      0.06688842 = sum of:
        0.06688842 = product of:
          0.13377684 = sum of:
            0.13377684 = weight(_text_:thesaurus in 989) [ClassicSimilarity], result of:
              0.13377684 = score(doc=989,freq=8.0), product of:
                0.21834905 = queryWeight, product of:
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.04725067 = queryNorm
                0.61267424 = fieldWeight in 989, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.046875 = fieldNorm(doc=989)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Crosswalks between different vocabularies are an indispensable prerequisite for integrated and high quality search scenarios in distributed data environments. Offered through the web and linked with each other they act as a central link so that users can move back and forth between different data sources available online. In the past, crosswalks between different thesauri have primarily been developed manually. In the long run the intellectual updating of such crosswalks requires huge personnel expenses. Therefore, an integration of automatic matching procedures, as for example Ontology Matching Tools, seems an obvious need. On the basis of computer generated correspondences between the Thesaurus for Economics (STW) and the Thesaurus for the Social Sciences (TheSoz) our contribution will explore cross-border approaches between IT-assisted tools and procedures on the one hand and external quality measurements via domain experts on the other hand. The techniques that emerge enable semi-automatically performed vocabulary crosswalks.
    Object
    Thesaurus Sozialwissenschaften
    Standard Thesaurus Wirtschaft
  12. Kempf, A.O.; Ritze, D.; Eckert, K.; Zapilko, B.: New ways of mapping knowledge organization systems : using a semi-automatic matching procedure for building up vocabulary crosswalks (2014) 0.01
    0.009290058 = product of:
      0.055740345 = sum of:
        0.055740345 = product of:
          0.11148069 = sum of:
            0.11148069 = weight(_text_:thesaurus in 1371) [ClassicSimilarity], result of:
              0.11148069 = score(doc=1371,freq=8.0), product of:
                0.21834905 = queryWeight, product of:
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.04725067 = queryNorm
                0.5105618 = fieldWeight in 1371, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1371)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Crosswalks between different vocabularies are an indispensable prerequisite for integrated, high-quality search scenarios in distributed data environments where more than one controlled vocabulary is in use. Offered through the web and linked with each other they act as a central link so that users can move back and forth between different online data sources. In the past, crosswalks between different thesauri have usually been developed manually. In the long run the intellectual updating of such crosswalks is expensive. An obvious solution would be to apply automatic matching procedures, such as the so-called ontology matching tools. On the basis of computer-generated correspondences between the Thesaurus for the Social Sciences (TSS) and the Thesaurus for Economics (STW), our contribution explores the trade-off between IT-assisted tools and procedures on the one hand and external quality evaluation by domain experts on the other hand. This paper presents techniques for semi-automatic development and maintenance of vocabulary crosswalks. The performance of multiple matching tools was first evaluated against a reference set of correct mappings, then the tools were used to generate new mappings. It was concluded that the ontology matching tools can be used effectively to speed up the work of domain experts. By optimizing the workflow, the method promises to facilitate sustained updating of high-quality vocabulary crosswalks.
    Object
    Thesaurus Wirtschaft
    Thesaurus Sozialwissenschaften
  13. Kempf, A.O.; Neubert, J.; Faden, M.: ¬The missing link : a vocabulary mapping effort in economics (2015) 0.00
    0.004645029 = product of:
      0.027870173 = sum of:
        0.027870173 = product of:
          0.055740345 = sum of:
            0.055740345 = weight(_text_:thesaurus in 2251) [ClassicSimilarity], result of:
              0.055740345 = score(doc=2251,freq=2.0), product of:
                0.21834905 = queryWeight, product of:
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.04725067 = queryNorm
                0.2552809 = fieldWeight in 2251, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.6210785 = idf(docFreq=1182, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2251)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    In economics there exists an internationally established classification system. Research literature is usually classified according to the JEL classification codes, a classification system originated by the Journal of Economic Literature and published by the American Economic Association (AEA). Complementarily to keywords which are usually assigned freely, economists widely use the JEL codes when classifying their publications. In cooperation with KU Leuven, ZBW - Leibniz Information Centre for Economics has published an unofficial multilingual version of JEL in SKOS format. In addition to this, exists the STW Thesaurus for Economics a bilingual domain-specific controlled vocabulary maintained by the German National Library of Economics (ZBW). Developed in the mid-1990s and since then constantly updated according to the current terminology usage in the latest international research literature in economics it covers all sub-fields both in the economics as well as in business economics and business practice containing subject headings which are clearly delimited from each other. It has been published on the web as Linked Open Data in the year 2009.
  14. Kempf, A.O.; Baum, K.: Von der Ein-Datenbank-Suche zum verteilten Suchszenario : Zum Aufbau von Crosskonkordanzen zwischen der Fachklassifikation Sozialwissenschaften und der Dewey-Dezimalklassifikation (2013) 0.00
    0.0042741043 = product of:
      0.025644625 = sum of:
        0.025644625 = weight(_text_:und in 1654) [ClassicSimilarity], result of:
          0.025644625 = score(doc=1654,freq=2.0), product of:
            0.104724824 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.04725067 = queryNorm
            0.24487628 = fieldWeight in 1654, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.078125 = fieldNorm(doc=1654)
      0.16666667 = coord(1/6)