Search (50 results, page 1 of 3)

  • × theme_ss:"Automatisches Indexieren"
  1. Plaunt, C.; Norgard, B.A.: ¬An association-based method for automatic indexing with a controlled vocabulary (1998) 0.10
    0.098847225 = product of:
      0.19769445 = sum of:
        0.19769445 = sum of:
          0.16300207 = weight(_text_:headings in 1794) [ClassicSimilarity], result of:
            0.16300207 = score(doc=1794,freq=12.0), product of:
              0.24837378 = queryWeight, product of:
                4.849944 = idf(docFreq=940, maxDocs=44218)
                0.051211677 = queryNorm
              0.6562773 = fieldWeight in 1794, product of:
                3.4641016 = tf(freq=12.0), with freq of:
                  12.0 = termFreq=12.0
                4.849944 = idf(docFreq=940, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1794)
          0.034692377 = weight(_text_:22 in 1794) [ClassicSimilarity], result of:
            0.034692377 = score(doc=1794,freq=2.0), product of:
              0.17933457 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051211677 = queryNorm
              0.19345059 = fieldWeight in 1794, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1794)
      0.5 = coord(1/2)
    
    Abstract
    In this article, we describe and test a two-stage algorithm based on a lexical collocation technique which maps from the lexical clues contained in a document representation into a controlled vocabulary list of subject headings. Using a collection of 4.626 INSPEC documents, we create a 'dictionary' of associations between the lexical items contained in the titles, authors, and abstracts, and controlled vocabulary subject headings assigned to those records by human indexers using a likelihood ratio statistic as the measure of association. In the deployment stage, we use the dictiony to predict which of the controlled vocabulary subject headings best describe new documents when they are presented to the system. Our evaluation of this algorithm, in which we compare the automatically assigned subject headings to the subject headings assigned to the test documents by human catalogers, shows that we can obtain results comparable to, and consistent with, human cataloging. In effect we have cast this as a classic partial match information retrieval problem. We consider the problem to be one of 'retrieving' (or assigning) the most probably 'relevant' (or correct) controlled vocabulary subject headings to a document based on the clues contained in that document
    Date
    11. 9.2000 19:53:22
  2. Olsgaard, J.N.; Evans, E.J.: Improving keyword indexing (1981) 0.03
    0.03327266 = product of:
      0.06654532 = sum of:
        0.06654532 = product of:
          0.13309065 = sum of:
            0.13309065 = weight(_text_:headings in 4996) [ClassicSimilarity], result of:
              0.13309065 = score(doc=4996,freq=2.0), product of:
                0.24837378 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.051211677 = queryNorm
                0.5358482 = fieldWeight in 4996, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4996)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This communication examines some of the most frequently cited critisms of keyword indexing. These critisms include (1) absence of general subject headings, (2) limited entry points, and (3) irrelevant indexing. Some solutions are suggested to meet these critisms.
  3. Junger, U.: Can indexing be automated? : the example of the Deutsche Nationalbibliothek (2012) 0.03
    0.03293825 = product of:
      0.0658765 = sum of:
        0.0658765 = product of:
          0.131753 = sum of:
            0.131753 = weight(_text_:headings in 1717) [ClassicSimilarity], result of:
              0.131753 = score(doc=1717,freq=4.0), product of:
                0.24837378 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.051211677 = queryNorm
                0.5304626 = fieldWeight in 1717, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1717)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The German subject headings authority file (Schlagwortnormdatei/SWD) provides a broad controlled vocabulary for indexing documents of all subjects. Traditionally used for intellectual subject cataloguing primarily of books the Deutsche Nationalbibliothek (DNB, German National Library) has been working on developping and implementing procedures for automated assignment of subject headings for online publications. This project, its results and problems are sketched in the paper.
  4. Short, M.: Text mining and subject analysis for fiction; or, using machine learning and information extraction to assign subject headings to dime novels (2019) 0.03
    0.03293825 = product of:
      0.0658765 = sum of:
        0.0658765 = product of:
          0.131753 = sum of:
            0.131753 = weight(_text_:headings in 5481) [ClassicSimilarity], result of:
              0.131753 = score(doc=5481,freq=4.0), product of:
                0.24837378 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.051211677 = queryNorm
                0.5304626 = fieldWeight in 5481, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5481)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This article describes multiple experiments in text mining at Northern Illinois University that were undertaken to improve the efficiency and accuracy of cataloging. It focuses narrowly on subject analysis of dime novels, a format of inexpensive fiction that was popular in the United States between 1860 and 1915. NIU holds more than 55,000 dime novels in its collections, which it is in the process of comprehensively digitizing. Classification, keyword extraction, named-entity recognition, clustering, and topic modeling are discussed as means of assigning subject headings to improve their discoverability by researchers and to increase the productivity of digitization workflows.
  5. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.03
    0.0277539 = product of:
      0.0555078 = sum of:
        0.0555078 = product of:
          0.1110156 = sum of:
            0.1110156 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.1110156 = score(doc=402,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  6. Abdul, H.; Khoo, C.: Automatic indexing of medical literature using phrase matching : an exploratory study 0.03
    0.026618127 = product of:
      0.053236254 = sum of:
        0.053236254 = product of:
          0.10647251 = sum of:
            0.10647251 = weight(_text_:headings in 3601) [ClassicSimilarity], result of:
              0.10647251 = score(doc=3601,freq=2.0), product of:
                0.24837378 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.051211677 = queryNorm
                0.42867854 = fieldWeight in 3601, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3601)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Reports the 1st part of a study to apply the technique of phrase matching to the automatic assignment of MeSH subject headings and subheadings to abstracts of periodical articles.
  7. Fuhr, N.; Niewelt, B.: ¬Ein Retrievaltest mit automatisch indexierten Dokumenten (1984) 0.02
    0.024284663 = product of:
      0.048569325 = sum of:
        0.048569325 = product of:
          0.09713865 = sum of:
            0.09713865 = weight(_text_:22 in 262) [ClassicSimilarity], result of:
              0.09713865 = score(doc=262,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.5416616 = fieldWeight in 262, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=262)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    20.10.2000 12:22:23
  8. Hlava, M.M.K.: Automatic indexing : comparing rule-based and statistics-based indexing systems (2005) 0.02
    0.024284663 = product of:
      0.048569325 = sum of:
        0.048569325 = product of:
          0.09713865 = sum of:
            0.09713865 = weight(_text_:22 in 6265) [ClassicSimilarity], result of:
              0.09713865 = score(doc=6265,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.5416616 = fieldWeight in 6265, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6265)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Information outlook. 9(2005) no.8, S.22-23
  9. Losee, R.M.: ¬A Gray code based ordering for documents on shelves : classification for browsing and retrieval (1992) 0.02
    0.023290861 = product of:
      0.046581723 = sum of:
        0.046581723 = product of:
          0.093163446 = sum of:
            0.093163446 = weight(_text_:headings in 2335) [ClassicSimilarity], result of:
              0.093163446 = score(doc=2335,freq=2.0), product of:
                0.24837378 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.051211677 = queryNorm
                0.37509373 = fieldWeight in 2335, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2335)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    A document classifier places documents together in a linear arrangement for browsing or high-speed access by human or computerised information retrieval systems. Requirements for document classification and browsing systems are developed from similarity measures, distance measures, and the notion of subject aboutness. A requirement that documents be arranged in decreasing order of similarity as the distance from a given document increases can often not be met. Based on these requirements, information-theoretic considerations, and the Gray code, a classification system is proposed that can classifiy documents without human intervention. A measure of classifier performance is developed, and used to evaluate experimental results comparing the distance between subject headings assigned to documents given classifications from the proposed system and the Library of Congress Classification (LCC) system
  10. Shafer, K.: Scorpion Project explores using Dewey to organize the Web (1996) 0.02
    0.023290861 = product of:
      0.046581723 = sum of:
        0.046581723 = product of:
          0.093163446 = sum of:
            0.093163446 = weight(_text_:headings in 6750) [ClassicSimilarity], result of:
              0.093163446 = score(doc=6750,freq=2.0), product of:
                0.24837378 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.051211677 = queryNorm
                0.37509373 = fieldWeight in 6750, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=6750)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    As the amount of accessible information on the WWW increases, so will the cost of accessing it, even if search servcies remain free, due to the increasing amount of time users will have to spend to find needed items. Considers what the seemingly unorganized Web and the organized world of libraries can offer each other. The OCLC Scorpion Project is attempting to combine indexing and cataloguing, specifically focusing on building tools for automatic subject recognition using the technqiues of library science and information retrieval. If subject headings or concept domains can be automatically assigned to electronic items, improved filtering tools for searching can be produced
  11. Junger, U.: Can indexing be automated? : the example of the Deutsche Nationalbibliothek (2014) 0.02
    0.023290861 = product of:
      0.046581723 = sum of:
        0.046581723 = product of:
          0.093163446 = sum of:
            0.093163446 = weight(_text_:headings in 1969) [ClassicSimilarity], result of:
              0.093163446 = score(doc=1969,freq=2.0), product of:
                0.24837378 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.051211677 = queryNorm
                0.37509373 = fieldWeight in 1969, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1969)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The German Integrated Authority File (Gemeinsame Normdatei, GND), provides a broad controlled vocabulary for indexing documents on all subjects. Traditionally used for intellectual subject cataloging primarily for books, the Deutsche Nationalbibliothek (DNB, German National Library) has been working on developing and implementing procedures for automated assignment of subject headings for online publications. This project, its results, and problems are outlined in this article.
  12. Moulaison-Sandy, H.; Adkins, D.; Bossaller, J.; Cho, H.: ¬An automated approach to describing fiction : a methodology to use book reviews to identify affect (2021) 0.02
    0.023290861 = product of:
      0.046581723 = sum of:
        0.046581723 = product of:
          0.093163446 = sum of:
            0.093163446 = weight(_text_:headings in 710) [ClassicSimilarity], result of:
              0.093163446 = score(doc=710,freq=2.0), product of:
                0.24837378 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.051211677 = queryNorm
                0.37509373 = fieldWeight in 710, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=710)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Subject headings and genre terms are notoriously difficult to apply, yet are important for fiction. The current project functions as a proof of concept, using a text-mining methodology to identify affective information (emotion and tone) about fiction titles from professional book reviews as a potential first step in automating the subject analysis process. Findings are presented and discussed, comparing results to the range of aboutness and isness information in library cataloging records. The methodology is likewise presented, and how future work might expand on the current project to enhance catalog records through text-mining is explored.
  13. Chou, C.; Chu, T.: ¬An analysis of BERT (NLP) for assisted subject indexing for Project Gutenberg (2022) 0.02
    0.023290861 = product of:
      0.046581723 = sum of:
        0.046581723 = product of:
          0.093163446 = sum of:
            0.093163446 = weight(_text_:headings in 1139) [ClassicSimilarity], result of:
              0.093163446 = score(doc=1139,freq=2.0), product of:
                0.24837378 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.051211677 = queryNorm
                0.37509373 = fieldWeight in 1139, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1139)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In light of AI (Artificial Intelligence) and NLP (Natural language processing) technologies, this article examines the feasibility of using AI/NLP models to enhance the subject indexing of digital resources. While BERT (Bidirectional Encoder Representations from Transformers) models are widely used in scholarly communities, the authors assess whether BERT models can be used in machine-assisted indexing in the Project Gutenberg collection, through suggesting Library of Congress subject headings filtered by certain Library of Congress Classification subclass labels. The findings of this study are informative for further research on BERT models to assist with automatic subject indexing for digital library collections.
  14. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.02
    0.020815425 = product of:
      0.04163085 = sum of:
        0.04163085 = product of:
          0.0832617 = sum of:
            0.0832617 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
              0.0832617 = score(doc=58,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.46428138 = fieldWeight in 58, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=58)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:44
  15. Hauer, M.: Automatische Indexierung (2000) 0.02
    0.020815425 = product of:
      0.04163085 = sum of:
        0.04163085 = product of:
          0.0832617 = sum of:
            0.0832617 = weight(_text_:22 in 5887) [ClassicSimilarity], result of:
              0.0832617 = score(doc=5887,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.46428138 = fieldWeight in 5887, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=5887)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Wissen in Aktion: Wege des Knowledge Managements. 22. Online-Tagung der DGI, Frankfurt am Main, 2.-4.5.2000. Proceedings. Hrsg.: R. Schmidt
  16. Fuhr, N.: Rankingexperimente mit gewichteter Indexierung (1986) 0.02
    0.020815425 = product of:
      0.04163085 = sum of:
        0.04163085 = product of:
          0.0832617 = sum of:
            0.0832617 = weight(_text_:22 in 2051) [ClassicSimilarity], result of:
              0.0832617 = score(doc=2051,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.46428138 = fieldWeight in 2051, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2051)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    14. 6.2015 22:12:56
  17. Hauer, M.: Tiefenindexierung im Bibliothekskatalog : 17 Jahre intelligentCAPTURE (2019) 0.02
    0.020815425 = product of:
      0.04163085 = sum of:
        0.04163085 = product of:
          0.0832617 = sum of:
            0.0832617 = weight(_text_:22 in 5629) [ClassicSimilarity], result of:
              0.0832617 = score(doc=5629,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.46428138 = fieldWeight in 5629, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=5629)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    B.I.T.online. 22(2019) H.2, S.163-166
  18. Gil-Leiva, I.: SISA-automatic indexing system for scientific articles : experiments with location heuristics rules versus TF-IDF rules (2017) 0.02
    0.019963596 = product of:
      0.039927192 = sum of:
        0.039927192 = product of:
          0.079854384 = sum of:
            0.079854384 = weight(_text_:headings in 3622) [ClassicSimilarity], result of:
              0.079854384 = score(doc=3622,freq=2.0), product of:
                0.24837378 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.051211677 = queryNorm
                0.3215089 = fieldWeight in 3622, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3622)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Indexing is contextualized and a brief description is provided of some of the most used automatic indexing systems. We describe SISA, a system which uses location heuristics rules, statistical rules like term frequency (TF) or TF-IDF to obtain automatic or semi-automatic indexing, depending on the user's preference. The aim of this research is to ascertain which rules (location heuristics rules or TF-IDF rules) provide the best indexing terms. SISA is used to obtain the automatic indexing of 200 scientific articles on fruit growing written in Portuguese. It uses, on the one hand, location heuristics rules founded on the value of certain parts of the articles for indexing such as titles, abstracts, keywords, headings, first paragraph, conclusions and references and, on the other, TF-IDF rules. The indexing is then evaluated to ascertain retrieval performance through recall, precision and f-measure. Automatic indexing of the articles with location heuristics rules provided the best results with the evaluation measures.
  19. Biebricher, N.; Fuhr, N.; Lustig, G.; Schwantner, M.; Knorz, G.: ¬The automatic indexing system AIR/PHYS : from research to application (1988) 0.02
    0.017346188 = product of:
      0.034692377 = sum of:
        0.034692377 = product of:
          0.06938475 = sum of:
            0.06938475 = weight(_text_:22 in 1952) [ClassicSimilarity], result of:
              0.06938475 = score(doc=1952,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.38690117 = fieldWeight in 1952, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1952)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    16. 8.1998 12:51:22
  20. Kutschekmanesch, S.; Lutes, B.; Moelle, K.; Thiel, U.; Tzeras, K.: Automated multilingual indexing : a synthesis of rule-based and thesaurus-based methods (1998) 0.02
    0.017346188 = product of:
      0.034692377 = sum of:
        0.034692377 = product of:
          0.06938475 = sum of:
            0.06938475 = weight(_text_:22 in 4157) [ClassicSimilarity], result of:
              0.06938475 = score(doc=4157,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.38690117 = fieldWeight in 4157, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4157)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Information und Märkte: 50. Deutscher Dokumentartag 1998, Kongreß der Deutschen Gesellschaft für Dokumentation e.V. (DGD), Rheinische Friedrich-Wilhelms-Universität Bonn, 22.-24. September 1998. Hrsg. von Marlies Ockenfeld u. Gerhard J. Mantwill

Years

Languages

  • e 34
  • d 15
  • ru 1
  • More… Less…

Types

  • a 46
  • el 3
  • x 3
  • m 1
  • More… Less…