Search (32 results, page 1 of 2)

  • × year_i:[1990 TO 2000}
  • × theme_ss:"Indexierungsstudien"
  1. Gretz, M.; Thomas, M.: Indexierungen in biomedizinischen Literaturdatenbanken : eine vergleichende Analyse (1991) 0.02
    0.018251318 = product of:
      0.05475395 = sum of:
        0.017669076 = weight(_text_:in in 5104) [ClassicSimilarity], result of:
          0.017669076 = score(doc=5104,freq=16.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.29755569 = fieldWeight in 5104, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5104)
        0.037084877 = weight(_text_:und in 5104) [ClassicSimilarity], result of:
          0.037084877 = score(doc=5104,freq=10.0), product of:
            0.09675359 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.043654136 = queryNorm
            0.38329202 = fieldWeight in 5104, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5104)
      0.33333334 = coord(2/6)
    
    Abstract
    Auf der Grundlage von vier Originaldokumenten, d.h. dokumentarischen Bezugseinheiten (DBEs), wird die Indexierung in vier biomedizinischen Online-Datenbanken (MEDLINE, EMBASE, BIOSIS PREVIEWS, SCISEARCH) analysiert. Anhand von Beispielen werden inahltliche Erschließung, Indexierungstiefe, Indexierungsbreite, Indexierungskonsistenz, Präzision (durch syntaktisches Indexieren, Gewichtung, Proximity Operatoren) und Wiederauffindbarkeit (Recall) der in den Datenbanken gespeicherten Dokumentationseinheien (DBEs) untersucht. Die zeitaufwendigere intellektuelle Indexierung bei MEDLINE und EMBASE erweist sich als wesentlich präziser als die schneller verfügbare maschinelle Zuteilung von Deskriptoren in BIOSIS PREVIEWS und SCISEARCH. In Teil 1 der Untersuchung werden die Indexierungen in MEDLINE und EMBASE, in Teil 2 die Deskriptorenzuteilungen in BIOSIS PREVIEWS und SCISEARCH verglichen
  2. Ladewig, C.; Rieger, M.: Ähnlichkeitsmessung mit und ohne aspektische Indexierung (1998) 0.02
    0.016507369 = product of:
      0.049522102 = sum of:
        0.0071393843 = weight(_text_:in in 2526) [ClassicSimilarity], result of:
          0.0071393843 = score(doc=2526,freq=2.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.120230645 = fieldWeight in 2526, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=2526)
        0.042382717 = weight(_text_:und in 2526) [ClassicSimilarity], result of:
          0.042382717 = score(doc=2526,freq=10.0), product of:
            0.09675359 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.043654136 = queryNorm
            0.438048 = fieldWeight in 2526, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=2526)
      0.33333334 = coord(2/6)
    
    Abstract
    Für eine fiktive Dokumentmenge wird eine Dokument-Wort-Matrix erstellt und mittels zweier Suchanfragen, ebenfalls als Matrix dargestellt, die Retrievalergebnisse ermittelt. Den Wörtern der Dokumentmenge werden in einem zweiten Schritt Aspekte zugeordnet und die Untersuchung erneut durchgeführt. Ein Vergleich bestätigt die schon früher gefundenen Vorteile des aspektischen Indexierung gegenüber anderen Methoden der Retrievalverbesserung, wie Trunkierung und Controlled Terms
    Source
    nfd Information - Wissenschaft und Praxis. 49(1998) H.8, S.459-462
  3. Veenema, F.: To index or not to index (1996) 0.01
    0.012007969 = product of:
      0.036023907 = sum of:
        0.012365777 = weight(_text_:in in 7247) [ClassicSimilarity], result of:
          0.012365777 = score(doc=7247,freq=6.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.2082456 = fieldWeight in 7247, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=7247)
        0.02365813 = product of:
          0.04731626 = sum of:
            0.04731626 = weight(_text_:22 in 7247) [ClassicSimilarity], result of:
              0.04731626 = score(doc=7247,freq=2.0), product of:
                0.15286934 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043654136 = queryNorm
                0.30952093 = fieldWeight in 7247, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=7247)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Describes an experiment comparing the performance of automatic full-text indexing software for personal computers with the human intellectual assignment of indexing terms in each document in a collection. Considers the times required to index the document, to retrieve documents satisfying 5 typical foreseen information needs, and the recall and precision ratios of searching. The software used is QuickFinder facility in WordPerfect 6.1 for Windows
    Source
    Canadian journal of information and library science. 21(1996) no.2, S.1-22
  4. Booth, A.: How consistent is MEDLINE indexing? (1990) 0.01
    0.008982609 = product of:
      0.026947826 = sum of:
        0.006246961 = weight(_text_:in in 3510) [ClassicSimilarity], result of:
          0.006246961 = score(doc=3510,freq=2.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.10520181 = fieldWeight in 3510, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3510)
        0.020700864 = product of:
          0.04140173 = sum of:
            0.04140173 = weight(_text_:22 in 3510) [ClassicSimilarity], result of:
              0.04140173 = score(doc=3510,freq=2.0), product of:
                0.15286934 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043654136 = queryNorm
                0.2708308 = fieldWeight in 3510, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3510)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    A known-item search for abstracts to previously retrieved references revealed that 2 documents from the same annual volume had been indexed twice. Working from the premise that the whole volume may have been double-indexed, a search strategy was devised that limited the journal code to the year in question. 57 references were retrieved, comprising 28 pairs of duplicates plus a citation for the whole volume. Author, title, source and descriptors were requested off-line and the citations were paired with their duplicates. The 4 categories of descriptors-major descriptors, minor descriptors, subheadings and check-tags-were compared for depth and consistency of indexing and lessons that might be learnt from the study are discussed.
    Source
    Health libraries review. 7(1990) no.1, S.22-26
  5. Harter, S.P.; Cheng, Y.-R.: Colinked descriptors : improving vocabulary selection for end-user searching (1996) 0.01
    0.008308224 = product of:
      0.024924671 = sum of:
        0.010709076 = weight(_text_:in in 4216) [ClassicSimilarity], result of:
          0.010709076 = score(doc=4216,freq=8.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.18034597 = fieldWeight in 4216, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=4216)
        0.014215595 = weight(_text_:und in 4216) [ClassicSimilarity], result of:
          0.014215595 = score(doc=4216,freq=2.0), product of:
            0.09675359 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.043654136 = queryNorm
            0.14692576 = fieldWeight in 4216, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=4216)
      0.33333334 = coord(2/6)
    
    Abstract
    This article introduces a new concept and technique for information retrieval called 'colinked descriptors'. Borrowed from an analogous idea in bibliometrics - cocited references - colinked descriptors provide a theory and method for identifying search terms that, by hypothesis, will be superior to those entered initially by a searcher. The theory suggests a means of moving automatically from 2 or more initial search terms, to other terms that should be superior in retrieval performance to the 2 original terms. A research project designed to test this colinked descriptor hypothesis is reported. The results suggest that the approach is effective, although methodological problems in testing the idea are reported. Algorithms to generate colinked descriptors can be incorporated easily into system interfaces, front-end or pre-search systems, or help software, in any database that employs a thesaurus. The potential use of colinked descriptors is a strong argument for building richer and more complex thesauri that reflect as many legitimate links among descriptors as possible
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  6. Iivonen, M.; Kivimäki, K.: Common entities and missing properties : similarities and differences in the indexing of concepts (1998) 0.00
    0.0030914443 = product of:
      0.018548666 = sum of:
        0.018548666 = weight(_text_:in in 3074) [ClassicSimilarity], result of:
          0.018548666 = score(doc=3074,freq=24.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.3123684 = fieldWeight in 3074, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=3074)
      0.16666667 = coord(1/6)
    
    Abstract
    The selection and representation of concepts in indexing of the same documents in 2 databases of library and information studies are considered. the authors compare the indexing of 49 documents in KINF and LISA. They focus on the types of concepts presented in indexing, the degree of concept consistency in indexing, and similarities and differences in the indexing of concepts. The largest group of indexed concepts in both databases was the category of entities while concepts belonging to the category of properties were almost missing in both databases. The second largest group of indexed concepts in KINF was the category of activities and in LISA the category of dimensions. Although the concept consistency between KINF and LISA remained rather low and was only 34%, there were approximately 2,2 concepts per document which were indexed from the same documents in both databses. These common concepts belonged mostly to the category of entities
  7. Braam, R.R.; Bruil, J.: Quality of indexing information : authors' views on indexing of their articles in chemical abstracts online CA-file (1992) 0.00
    0.0029598325 = product of:
      0.017758995 = sum of:
        0.017758995 = weight(_text_:in in 2638) [ClassicSimilarity], result of:
          0.017758995 = score(doc=2638,freq=22.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.29906997 = fieldWeight in 2638, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=2638)
      0.16666667 = coord(1/6)
    
    Abstract
    Studies the quality of subject indexing by Chemical Abstracts Indexing Service by confronting authors with the particular indexing terms attributed to their computer, for 270 articles published in 54 journals, 5 articles out of each journal. Responses (80%) indicate the superior quality of keywords, both as content descriptors and as retrieval tools. Author judgements on these 2 different aspects do not always converge, however. CAS's indexing policy to cover only 'new' aspects is reflected in author's judgements that index lists are somewhat incomplete, in particular in the case of thesaurus terms (index headings). The large effort expanded by CAS in maintaining and using a subject thesuaurs, in order to select valid index headings, as compared to quick and cheap keyword postings, does not lead to clear superior quality of thesaurus terms for document description nor in retrieval. Some 20% of papers were not placed in 'proper' CA main section, according to authors. As concerns the use of indexing data by third parties, in bibliometrics, users should be aware of the indexing policies behind the data, in order to prevent invalid interpretations
  8. Morris, L.R.: ¬The frequency of use of Library of Congress Classification numbers and Dewey Decimal Classification numbers in the MARC file in the field of library science (1991) 0.00
    0.0025503114 = product of:
      0.015301868 = sum of:
        0.015301868 = weight(_text_:in in 2308) [ClassicSimilarity], result of:
          0.015301868 = score(doc=2308,freq=12.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.2576908 = fieldWeight in 2308, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2308)
      0.16666667 = coord(1/6)
    
    Abstract
    The LCC and DDC systems were devised and updated by librarians who had and have no access to the eventual frequency of use of each number in those classification systems. 80% of the monographs in a MARC file of over 1.000.000 records are classified into 20% of the classification numbers in the field of library science and only 20% of the mongraphs are classified into 80% of the classification numbers in the field of library science. Classification of monographs coulld be made easier and performed more accurately if many of the little used and unused numbers were eliminated and many of the most crowded numbers were expanded. A number of examples are included
  9. Rowley, J.: ¬The controlled versus natural indexing languages debate revisited : a perspective on information retrieval practice and research (1994) 0.00
    0.0023517415 = product of:
      0.014110449 = sum of:
        0.014110449 = weight(_text_:in in 7151) [ClassicSimilarity], result of:
          0.014110449 = score(doc=7151,freq=20.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.2376267 = fieldWeight in 7151, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0390625 = fieldNorm(doc=7151)
      0.16666667 = coord(1/6)
    
    Abstract
    This article revisits the debate concerning controlled and natural indexing languages, as used in searching the databases of the online hosts, in-house information retrieval systems, online public access catalogues and databases stored on CD-ROM. The debate was first formulated in the early days of information retrieval more than a century ago but, despite significant advance in technology, remains unresolved. The article divides the history of the debate into four eras. Era one was characterised by the introduction of controlled vocabulary. Era two focused on comparisons between different indexing languages in order to assess which was best. Era three saw a number of case studies of limited generalisability and a general recognition that the best search performance can be achieved by the parallel use of the two types of indexing languages. The emphasis in Era four has been on the development of end-user-based systems, including online public access catalogues and databases on CD-ROM. Recent developments in the use of expert systems techniques to support the representation of meaning may lead to systems which offer significant support to the user in end-user searching. In the meantime, however, information retrieval in practice involves a mixture of natural and controlled indexing languages used to search a wide variety of different kinds of databases
  10. Ellis, D.; Furner, J.; Willett, P.: On the creation of hypertext links in full-text documents : measurement of retrieval effectiveness (1996) 0.00
    0.0023517415 = product of:
      0.014110449 = sum of:
        0.014110449 = weight(_text_:in in 4214) [ClassicSimilarity], result of:
          0.014110449 = score(doc=4214,freq=20.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.2376267 = fieldWeight in 4214, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4214)
      0.16666667 = coord(1/6)
    
    Abstract
    An important stage in the process or retrieval of objects from a hypertext database is the creation of a set of internodal links that are intended to represent the relationships existing between objects; this operation is often undertaken manually, just as index terms are often manually assigned to documents in a conventional retrieval system. In an earlier article (1994), the results were published of a study in which several different sets of links were inserted, each by a different person, between the paragraphs of each of a number of full-text documents. These results showed little similarity between the link-sets, a finding that was comparable with those of studies of inter-indexer consistency, which suggest that there is generally only a low level of agreement between the sets of index terms assigned to a document by different indexers. In this article, a description is provided of an investigation into the nature of the relationship existing between (i) the levels of inter-linker consistency obtaining among the group of hypertext databases used in our earlier experiments, and (ii) the levels of effectiveness of a number of searches carried out in those databases. An account is given of the implementation of the searches and of the methods used in the calculation of numerical values expressing their effectiveness. Analysis of the results of a comparison between recorded levels of consistency and those of effectiveness does not allow us to draw conclusions about the consistency - effectiveness relationship that are equivalent to those drawn in comparable studies of inter-indexer consistency
  11. Krovetz, R.; Croft, W.B.: Lexical ambiguity and information retrieval (1992) 0.00
    0.0023281053 = product of:
      0.013968632 = sum of:
        0.013968632 = weight(_text_:in in 4028) [ClassicSimilarity], result of:
          0.013968632 = score(doc=4028,freq=10.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.23523843 = fieldWeight in 4028, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4028)
      0.16666667 = coord(1/6)
    
    Abstract
    Reports on an analysis of lexical ambiguity in information retrieval text collections and on experiments to determine the utility of word meanings for separating relevant from nonrelevant documents. Results show that there is considerable ambiguity even in a specialised database. Word senses provide a significant separation between relevant and nonrelevant documents, but several factors contribute to determining whether disambiguation will make an improvement in performance such as: resolving lexical ambiguity was found to have little impact on retrieval effectiveness for documents that have many words in common with the query. Discusses other uses of word sense disambiguation in an information retrieval context
  12. Tseng, Y.-H.: Keyword extraction techniques and relevance feedback (1997) 0.00
    0.0023281053 = product of:
      0.013968632 = sum of:
        0.013968632 = weight(_text_:in in 1830) [ClassicSimilarity], result of:
          0.013968632 = score(doc=1830,freq=10.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.23523843 = fieldWeight in 1830, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1830)
      0.16666667 = coord(1/6)
    
    Abstract
    Automatic keyword extraction is an important and fundamental technology in an advanced information retrieval systems. Briefly compares several major keyword extraction methods, lists their advantages and disadvantages, and reports recent research progress in Taiwan. Also describes the application of a keyword extraction algorithm in an information retrieval system for relevance feedback. Preliminary analysis shows that the error rate of extracting relevant keywords is 18%, and that the precision rate is over 50%. The main disadvantage of this approach is that the extraction results depend on the retrieval results, which in turn depend on the data held by the database. Apart from collecting more data, this problem can be alleviated by the application of a thesaurus constructed by the same keyword extraction algorithm
    Footnote
    [In Chinesisch]
  13. Mann, T.: 'Cataloging must change!' and indexer consistency studies : misreading the evidence at our peril (1997) 0.00
    0.0021859813 = product of:
      0.013115887 = sum of:
        0.013115887 = weight(_text_:in in 492) [ClassicSimilarity], result of:
          0.013115887 = score(doc=492,freq=12.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.22087781 = fieldWeight in 492, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=492)
      0.16666667 = coord(1/6)
    
    Abstract
    An earlier article ('Cataloging must change' by D. Gregor and C. Mandel in: Library journal 116(1991) no.6, S.42-47) has popularized the belief that there is low consistency (only 10-20% agreement) among subject cataloguers in assigning LCSH. Because of this alleged lack og consistency, the article suggests, cataloguers 'can be more accepting in variations in subject choices' in copy cataloguing. Argues that this inference is based on a serious misreading of previous studies of indexer consistency. The 10-20% figure actually derives from studies of people trying to guess the same natural language key words, precisely in the absence of vocabulary control mechanisms such as thesauri or LCSH. Concludes that sources cited fail support their conclusion and some directly contradict it. Raises the concern that a naive acceptance by the library profession of the 10-20% claim can only have negative consequences for the quality of subject cataloguing created, and accepted throughout the country
  14. Svenonius, E.; McGarry, D.: Objectivity in evaluating subject heading assignment (1993) 0.00
    0.0020823204 = product of:
      0.012493922 = sum of:
        0.012493922 = weight(_text_:in in 5612) [ClassicSimilarity], result of:
          0.012493922 = score(doc=5612,freq=8.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.21040362 = fieldWeight in 5612, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5612)
      0.16666667 = coord(1/6)
    
    Abstract
    Recent papers have called attention to discrepancies in the assignment of LCSH. While philosophical arguments can be made that subject analysis, if not a logical impossibility, at least is point-of-view dependent, subject headings continue to be assigned and continue to be useful. The hypothesis advanced in the present project is that to a considerable degree there is a clear-cut right and wrong to LCSH subject heading assignment. To test the hypothesis, it was postulated that the assignment of a subject heading is correct if it is supported by textual warrant (at least 20% of the book being cataloged is on the topic) and is constructed in accordance with the LoC Subject Cataloging Manual: Subject Headings. A sample of 100 books on scientific subjects was used to test the hypothesis
  15. Connell, T.H.: Use of the LCSH system : realities (1996) 0.00
    0.0020823204 = product of:
      0.012493922 = sum of:
        0.012493922 = weight(_text_:in in 6941) [ClassicSimilarity], result of:
          0.012493922 = score(doc=6941,freq=8.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.21040362 = fieldWeight in 6941, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=6941)
      0.16666667 = coord(1/6)
    
    Abstract
    Explores the question of whether academic libraries keep up with the changes in the LCSH system. Analysis of the handling of 15 subject headings in 50 academic library catalogues available via the Internet found that libraries are not consistently maintaining subject authority control, or making syndetic references and scope notes in their catalogues. Discusses the results from the perspective of the libraries' performance, performance on the headings overall, performance on references, performance on the type of change made to the headings,a nd performance within 3 widely used onlien catalogue systems (DRA, INNOPAC and NOTIS). Discusses the implications of the findings in relationship to expressions of dissatisfaction with the effectiveness of subject cataloguing expressed by discussion groups on the Internet
  16. Reich, P.; Biever, E.J.: Indexing consistency : The input/output function of thesauri (1991) 0.00
    0.0020609628 = product of:
      0.012365777 = sum of:
        0.012365777 = weight(_text_:in in 2258) [ClassicSimilarity], result of:
          0.012365777 = score(doc=2258,freq=6.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.2082456 = fieldWeight in 2258, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=2258)
      0.16666667 = coord(1/6)
    
    Abstract
    This study measures inter-indexer consistency as determined by the number of identical terms assigned to the same document by two different indexing organizations using the same thesaurus as a source for the entry vocabulary. The authors derive consistency figures of 24 percent and 45 percent for two samples. Factors in the consistency failures include variations in indexing depth, differences in choice of concepts for indexing, different indexing policies, and a highly specific indexing vocabulray. Results indicate that broad search strategies are often necessary for adequate search yields.
  17. Haanen, E.: Specificiteit en consistentie : een kwantitatief oderzoek naar trefwoordtoekenning door UBA en UBN (1991) 0.00
    0.0020609628 = product of:
      0.012365777 = sum of:
        0.012365777 = weight(_text_:in in 4778) [ClassicSimilarity], result of:
          0.012365777 = score(doc=4778,freq=6.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.2082456 = fieldWeight in 4778, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=4778)
      0.16666667 = coord(1/6)
    
    Abstract
    Online public access catalogues enable users to undertake subject searching by classification schedules, natural language, or controlled language terminology. In practice the 1st method is little used. Controlled language systems require indexers to index specifically and consistently. A comparative survey was made of indexing practices at Amsterdam and Mijmegen university libraries. On average Amsterdam assigned each document 3.5 index terms against 1.8 at Nijmegen. This discrepancy in indexing policy is the result of long-standing practices in each institution. Nijmegen has failed to utilise the advantages offered by online cataloges
  18. Kautto, V.: Classing and indexing : a comparative time study (1992) 0.00
    0.0017848461 = product of:
      0.010709076 = sum of:
        0.010709076 = weight(_text_:in in 2670) [ClassicSimilarity], result of:
          0.010709076 = score(doc=2670,freq=8.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.18034597 = fieldWeight in 2670, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=2670)
      0.16666667 = coord(1/6)
    
    Abstract
    A total of 16 classifiers made a subject analysis of a set of books such that some of the books were first classified by the UDC anf then indexed with terms from the General Finnish Subject Headings while another set were processed in the opposite order. Finally books on the same subject were either classifies or indexed. The total number of books processed was 581. A comparison was made of the time required for processing in different situations and of the number of classes or subject headings used. The time figures were compared with corresponding data from the British Library (1972) and the Library of Congress (1990 and 1991). The author finds that the contents analysis requires one third, classification one third and indexing obe third of the time, if the document is both classified and indexed. There was a plausible correlation (o.51) between the length of experience in classification and the decrease in the time required for classing. The average number of UDC numbers was 4,3 and the average number of terms from the list of subject headings was 4,0
  19. Gregor, D.; Mandel, C.: Cataloging must change! (1991) 0.00
    0.0017848461 = product of:
      0.010709076 = sum of:
        0.010709076 = weight(_text_:in in 1999) [ClassicSimilarity], result of:
          0.010709076 = score(doc=1999,freq=2.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.18034597 = fieldWeight in 1999, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.09375 = fieldNorm(doc=1999)
      0.16666667 = coord(1/6)
    
    Footnote
    Vgl. auch die Erwiderung von T. Mann in: Cataloging and classification quarterly 23(1997) nos.3/4, S.3-45
  20. Tonta, Y.: ¬A study of indexing consistency between Library of Congress and British Library catalogers (1991) 0.00
    0.001682769 = product of:
      0.010096614 = sum of:
        0.010096614 = weight(_text_:in in 2277) [ClassicSimilarity], result of:
          0.010096614 = score(doc=2277,freq=4.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.17003182 = fieldWeight in 2277, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=2277)
      0.16666667 = coord(1/6)
    
    Abstract
    Indexing consistency between Library of Congress and British Library catalogers using the LCSH is compared.82 titles published in 1987 in the field of library and information science were identified for comparison, and for each title its LC subject headings, assigned by both LC and BL catalogers, were compared. By applying Hooper's 'consistency of a pair' equation, the average indexing consistency value was calculated for the 82 titles. The average indexing value between LC and BL catalogers is 16% for exact matches, and 36% for partial matches