Search (35 results, page 1 of 2)

  • × theme_ss:"Indexierungsstudien"
  1. Veenema, F.: To index or not to index (1996) 0.12
    0.11743011 = product of:
      0.17614517 = sum of:
        0.14859696 = weight(_text_:index in 7247) [ClassicSimilarity], result of:
          0.14859696 = score(doc=7247,freq=6.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.6689808 = fieldWeight in 7247, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0625 = fieldNorm(doc=7247)
        0.027548207 = product of:
          0.055096414 = sum of:
            0.055096414 = weight(_text_:22 in 7247) [ClassicSimilarity], result of:
              0.055096414 = score(doc=7247,freq=2.0), product of:
                0.17800546 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.05083213 = queryNorm
                0.30952093 = fieldWeight in 7247, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=7247)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Describes an experiment comparing the performance of automatic full-text indexing software for personal computers with the human intellectual assignment of indexing terms in each document in a collection. Considers the times required to index the document, to retrieve documents satisfying 5 typical foreseen information needs, and the recall and precision ratios of searching. The software used is QuickFinder facility in WordPerfect 6.1 for Windows
    Source
    Canadian journal of information and library science. 21(1996) no.2, S.1-22
  2. Haanen, E.: Specificiteit en consistentie : een kwantitatief oderzoek naar trefwoordtoekenning door UBA en UBN (1991) 0.10
    0.0960757 = product of:
      0.14411354 = sum of:
        0.1213289 = weight(_text_:index in 4778) [ClassicSimilarity], result of:
          0.1213289 = score(doc=4778,freq=4.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.5462205 = fieldWeight in 4778, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0625 = fieldNorm(doc=4778)
        0.022784641 = product of:
          0.045569282 = sum of:
            0.045569282 = weight(_text_:classification in 4778) [ClassicSimilarity], result of:
              0.045569282 = score(doc=4778,freq=2.0), product of:
                0.16188543 = queryWeight, product of:
                  3.1847067 = idf(docFreq=4974, maxDocs=44218)
                  0.05083213 = queryNorm
                0.28149095 = fieldWeight in 4778, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1847067 = idf(docFreq=4974, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4778)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Online public access catalogues enable users to undertake subject searching by classification schedules, natural language, or controlled language terminology. In practice the 1st method is little used. Controlled language systems require indexers to index specifically and consistently. A comparative survey was made of indexing practices at Amsterdam and Mijmegen university libraries. On average Amsterdam assigned each document 3.5 index terms against 1.8 at Nijmegen. This discrepancy in indexing policy is the result of long-standing practices in each institution. Nijmegen has failed to utilise the advantages offered by online cataloges
  3. Lancaster, F.W.; Mills, J.: Testing indexes and index language devices : the ASLIB Cranfield project (1964) 0.06
    0.05719499 = product of:
      0.17158496 = sum of:
        0.17158496 = weight(_text_:index in 2261) [ClassicSimilarity], result of:
          0.17158496 = score(doc=2261,freq=2.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.7724724 = fieldWeight in 2261, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.125 = fieldNorm(doc=2261)
      0.33333334 = coord(1/3)
    
  4. Ballard, R.M.: Indexing and its relevance to technical processing (1993) 0.05
    0.052190267 = product of:
      0.078285396 = sum of:
        0.0536203 = weight(_text_:index in 554) [ClassicSimilarity], result of:
          0.0536203 = score(doc=554,freq=2.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.24139762 = fieldWeight in 554, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0390625 = fieldNorm(doc=554)
        0.024665099 = product of:
          0.049330197 = sum of:
            0.049330197 = weight(_text_:classification in 554) [ClassicSimilarity], result of:
              0.049330197 = score(doc=554,freq=6.0), product of:
                0.16188543 = queryWeight, product of:
                  3.1847067 = idf(docFreq=4974, maxDocs=44218)
                  0.05083213 = queryNorm
                0.3047229 = fieldWeight in 554, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.1847067 = idf(docFreq=4974, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=554)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The development of regional on-line catalogs and in-house information systems for retrieval of references provide examples of the impact of indexing theory and applications on technical processing. More emphasis must be given to understanding the techniques for evaluating the effectiveness of a file, irrespective of whether that file was created as a library catalog or an index to information sources. The most significant advances in classification theory in recent decades has been as a result of efforts to improve effectiveness of indexing systems. Library classification systems are indexing languages or systems. Courses offered for the preparation of indexers in the United States and the United Kingdom are reviewed. A point of congruence for both the indexer and the library classifier would appear to be the need for a thorough preparation in the techniques of subject analysis. Any subject heading list will suffer from omissions as well as the inclusion of terms which the patron will never use. Indexing theory has provided the technical services department with methods for evaluation of effectiveness. The writer does not believe that these techniques are used, nor do current courses, workshops, and continuing education programs stress them. When theory is totally subjugated to practice, critical thinking and maximum effectiveness will suffer.
    Source
    Cataloging and classification quarterly. 16(1993) no.3, S.79-88
  5. Cleverdon, C.W.: ¬The Cranfield tests on index language devices (1967) 0.04
    0.04289624 = product of:
      0.12868872 = sum of:
        0.12868872 = weight(_text_:index in 1957) [ClassicSimilarity], result of:
          0.12868872 = score(doc=1957,freq=2.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.5793543 = fieldWeight in 1957, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.09375 = fieldNorm(doc=1957)
      0.33333334 = coord(1/3)
    
  6. Braam, R.R.; Bruil, J.: Quality of indexing information : authors' views on indexing of their articles in chemical abstracts online CA-file (1992) 0.04
    0.03714924 = product of:
      0.111447714 = sum of:
        0.111447714 = weight(_text_:index in 2638) [ClassicSimilarity], result of:
          0.111447714 = score(doc=2638,freq=6.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.50173557 = fieldWeight in 2638, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.046875 = fieldNorm(doc=2638)
      0.33333334 = coord(1/3)
    
    Abstract
    Studies the quality of subject indexing by Chemical Abstracts Indexing Service by confronting authors with the particular indexing terms attributed to their computer, for 270 articles published in 54 journals, 5 articles out of each journal. Responses (80%) indicate the superior quality of keywords, both as content descriptors and as retrieval tools. Author judgements on these 2 different aspects do not always converge, however. CAS's indexing policy to cover only 'new' aspects is reflected in author's judgements that index lists are somewhat incomplete, in particular in the case of thesaurus terms (index headings). The large effort expanded by CAS in maintaining and using a subject thesuaurs, in order to select valid index headings, as compared to quick and cheap keyword postings, does not lead to clear superior quality of thesaurus terms for document description nor in retrieval. Some 20% of papers were not placed in 'proper' CA main section, according to authors. As concerns the use of indexing data by third parties, in bibliometrics, users should be aware of the indexing policies behind the data, in order to prevent invalid interpretations
  7. Losee, R.: ¬A performance model of the length and number of subject headings and index phrases (2004) 0.04
    0.03714924 = product of:
      0.111447714 = sum of:
        0.111447714 = weight(_text_:index in 3725) [ClassicSimilarity], result of:
          0.111447714 = score(doc=3725,freq=6.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.50173557 = fieldWeight in 3725, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.046875 = fieldNorm(doc=3725)
      0.33333334 = coord(1/3)
    
    Abstract
    When assigning subject headings or index terms to a document, how many terms or phrases should be used to represent the document? The contribution of an indexing phrase to locating and ordering documents can be compared to the contribution of a full-text query to finding documents. The length and number of phrases needed to equal the contribution of a full-text query is the subject of this paper. The appropriate number of phrases is determined in part by the length of the phrases. We suggest several rules that may be used to determine how many subject headings should be assigned, given index phrase lengths, and provide a general model for this process. A difference between characteristics of indexing "hard" science and "social" science literature is suggested.
  8. Broxis, P.F.: ASSIA social science information service (1989) 0.04
    0.03574687 = product of:
      0.1072406 = sum of:
        0.1072406 = weight(_text_:index in 1511) [ClassicSimilarity], result of:
          0.1072406 = score(doc=1511,freq=2.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.48279524 = fieldWeight in 1511, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.078125 = fieldNorm(doc=1511)
      0.33333334 = coord(1/3)
    
    Abstract
    ASSIA (Applied Social Science Index and Abtracts) started in 1987 as a bimonthly indexing and abstracting service in the society field, aimed at practitioners as well as sociologists. Considers the following aspects of the service: arrangement of ASSIA; journal coverage; indexing approach; services for subscribers; and who are the users?
  9. Qin, J.: Semantic similarities between a keyword database and a controlled vocabulary database : an investigation in the antibiotic resistance literature (2000) 0.03
    0.030957699 = product of:
      0.0928731 = sum of:
        0.0928731 = weight(_text_:index in 4386) [ClassicSimilarity], result of:
          0.0928731 = score(doc=4386,freq=6.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.418113 = fieldWeight in 4386, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4386)
      0.33333334 = coord(1/3)
    
    Abstract
    The 'KeyWords Plus' in the Science Citation Index database represents an approach to combining citation and semantic indexing in describing the document content. This paper explores the similariites or dissimilarities between citation-semantic and analytic indexing. The dataset consisted of over 400 matching records in the SCI and MEDLINE databases on antibiotic resistance in pneumonia. The degree of similarity in indexing terms was found to vary on a scale from completely different to completely identical with various levels in between. The within-document similarity in the 2 databases was measured by a variation on the Jaccard coefficient - the Inclusion Index. The average inclusion coefficient was 0,4134 for SCI and 0,3371 for Medline. The 20 terms occuring most frequently in each database were identified. The 2 groups of terms shared the same terms that consist of the 'intellectual base' for the subject. conceptual similarity was analyzed through scatterplots of matching and nonmatching terms vs. partially identical and broader/narrower terms. The study also found that both databases differed in assigning terms in various semantic categories. Implications of this research and further studies are suggested
    Object
    Science Citation Index
  10. Neshat, N.; Horri, A.: ¬A study of subject indexing consistency between the National Library of Iran and Humanities Libraries in the area of Iranian studies (2006) 0.03
    0.029360829 = product of:
      0.088082485 = sum of:
        0.088082485 = sum of:
          0.039873123 = weight(_text_:classification in 230) [ClassicSimilarity], result of:
            0.039873123 = score(doc=230,freq=2.0), product of:
              0.16188543 = queryWeight, product of:
                3.1847067 = idf(docFreq=4974, maxDocs=44218)
                0.05083213 = queryNorm
              0.24630459 = fieldWeight in 230, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1847067 = idf(docFreq=4974, maxDocs=44218)
                0.0546875 = fieldNorm(doc=230)
          0.04820936 = weight(_text_:22 in 230) [ClassicSimilarity], result of:
            0.04820936 = score(doc=230,freq=2.0), product of:
              0.17800546 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05083213 = queryNorm
              0.2708308 = fieldWeight in 230, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=230)
      0.33333334 = coord(1/3)
    
    Date
    4. 1.2007 10:22:26
    Source
    Cataloging and classification quarterly. 43(2006) no.1, S.67-76
  11. David, C.; Giroux, L.; Bertrand-Gastaldy, S.; Lanteigne, D.: Indexing as problem solving : a cognitive approach to consistency (1995) 0.03
    0.028597495 = product of:
      0.08579248 = sum of:
        0.08579248 = weight(_text_:index in 3833) [ClassicSimilarity], result of:
          0.08579248 = score(doc=3833,freq=2.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.3862362 = fieldWeight in 3833, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0625 = fieldNorm(doc=3833)
      0.33333334 = coord(1/3)
    
    Abstract
    Presents results of an experiment in which 8 indexers (4 beginners and 4 experts) were asked to index the same 4 documents with 2 different thesauri. The 3 kind of verbal reports provide complementary data on strategic behaviour. it is of prime importance to consider the indexing task as an ill-defined problem, where the solutionm is partly defined by the indexer
  12. Prasher, R.G.: Evaluation of indexing system (1989) 0.03
    0.028597495 = product of:
      0.08579248 = sum of:
        0.08579248 = weight(_text_:index in 4998) [ClassicSimilarity], result of:
          0.08579248 = score(doc=4998,freq=2.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.3862362 = fieldWeight in 4998, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0625 = fieldNorm(doc=4998)
      0.33333334 = coord(1/3)
    
    Abstract
    Describes information system and its various components-index file construstion, query formulation and searching. Discusses an indexing system, and brings out the need for its evaluation. Explains the concept of the efficiency of indexing systems and discusses factors which control this efficiency. Gives criteria for evaluation. Discusses recall and precision ratios, as also noise ratio, novelty ratio, and exhaustivity and specificity and the impact of each on the efficiency of indexing system. Mention also various steps for evaluation.
  13. Ellis, D.; Furner, J.; Willett, P.: On the creation of hypertext links in full-text documents : measurement of retrieval effectiveness (1996) 0.03
    0.025276855 = product of:
      0.075830564 = sum of:
        0.075830564 = weight(_text_:index in 4214) [ClassicSimilarity], result of:
          0.075830564 = score(doc=4214,freq=4.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.3413878 = fieldWeight in 4214, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4214)
      0.33333334 = coord(1/3)
    
    Abstract
    An important stage in the process or retrieval of objects from a hypertext database is the creation of a set of internodal links that are intended to represent the relationships existing between objects; this operation is often undertaken manually, just as index terms are often manually assigned to documents in a conventional retrieval system. In an earlier article (1994), the results were published of a study in which several different sets of links were inserted, each by a different person, between the paragraphs of each of a number of full-text documents. These results showed little similarity between the link-sets, a finding that was comparable with those of studies of inter-indexer consistency, which suggest that there is generally only a low level of agreement between the sets of index terms assigned to a document by different indexers. In this article, a description is provided of an investigation into the nature of the relationship existing between (i) the levels of inter-linker consistency obtaining among the group of hypertext databases used in our earlier experiments, and (ii) the levels of effectiveness of a number of searches carried out in those databases. An account is given of the implementation of the searches and of the methods used in the calculation of numerical values expressing their effectiveness. Analysis of the results of a comparison between recorded levels of consistency and those of effectiveness does not allow us to draw conclusions about the consistency - effectiveness relationship that are equivalent to those drawn in comparable studies of inter-indexer consistency
  14. Hughes, A.V.; Rafferty, P.: Inter-indexer consistency in graphic materials indexing at the National Library of Wales (2011) 0.03
    0.025276855 = product of:
      0.075830564 = sum of:
        0.075830564 = weight(_text_:index in 4488) [ClassicSimilarity], result of:
          0.075830564 = score(doc=4488,freq=4.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.3413878 = fieldWeight in 4488, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4488)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - This paper seeks to report a project to investigate the degree of inter-indexer consistency in the assignment of controlled vocabulary topical subject index terms to identical graphical images by different indexers at the National Library of Wales (NLW). Design/methodology/approach - An experimental quantitative methodology was devised to investigate inter-indexer consistency. Additionally, the project investigated the relationship, if any, between indexing exhaustivity and consistency, and the relationship, if any, between indexing consistency/exhaustivity and broad category of graphic format. Findings - Inter-indexer consistency in the assignment of topical subject index terms to graphic materials at the NLW was found to be generally low and highly variable. Inter-indexer consistency fell within the range 10.8 per cent to 48.0 per cent. Indexing exhaustivity varied substantially from indexer to indexer, with a mean assignment of 3.8 terms by each indexer to each image, falling within the range 2.5 to 4.7 terms. The broad category of graphic format, whether photographic or non-photographic, was found to have little influence on either inter-indexer consistency or indexing exhaustivity. Indexing exhaustivity and inter-indexer consistency exhibited a tendency toward a direct, positive relationship. The findings are necessarily limited as this is a small-scale study within a single institution. Originality/value - Previous consistency studies have almost exclusively investigated the indexing of print materials, with very little research published for non-print media. With the literature also rich in discussion of the added complexities of subjectively representing the intellectual content of visual media, this study attempts to enrich existing knowledge on indexing consistency for graphic materials and to address a noticeable gap in information theory.
  15. Leininger, K.: Interindexer consistency in PsychINFO (2000) 0.03
    0.025166426 = product of:
      0.075499274 = sum of:
        0.075499274 = sum of:
          0.034176964 = weight(_text_:classification in 2552) [ClassicSimilarity], result of:
            0.034176964 = score(doc=2552,freq=2.0), product of:
              0.16188543 = queryWeight, product of:
                3.1847067 = idf(docFreq=4974, maxDocs=44218)
                0.05083213 = queryNorm
              0.21111822 = fieldWeight in 2552, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1847067 = idf(docFreq=4974, maxDocs=44218)
                0.046875 = fieldNorm(doc=2552)
          0.04132231 = weight(_text_:22 in 2552) [ClassicSimilarity], result of:
            0.04132231 = score(doc=2552,freq=2.0), product of:
              0.17800546 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05083213 = queryNorm
              0.23214069 = fieldWeight in 2552, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2552)
      0.33333334 = coord(1/3)
    
    Abstract
    Reports results of a study to examine interindexer consistency (the degree to which indexers, when assigning terms to a chosen record, will choose the same terms to reflect that record) in the PsycINFO database using 60 records that were inadvertently processed twice between 1996 and 1998. Five aspects of interindexer consistency were analysed. Two methods were used to calculate interindexer consistency: one posited by Hooper (1965) and the other by Rollin (1981). Aspects analysed were: checktag consistency (66.24% using Hooper's calculation and 77.17% using Rollin's); major-to-all term consistency (49.31% and 62.59% respectively); overall indexing consistency (49.02% and 63.32%); classification code consistency (44.17% and 45.00%); and major-to-major term consistency (43.24% and 56.09%). The average consistency across all categories was 50.4% using Hooper's method and 60.83% using Rollin's. Although comparison with previous studies is difficult due to methodological variations in the overall study of indexing consistency and the specific characteristics of the database, results generally support previous findings when trends and similar studies are analysed.
    Date
    9. 2.1997 18:44:22
  16. Boyce, B.R.; McLain, J.P.: Entry point depth and online search using a controlled vocabulary (1989) 0.03
    0.025022808 = product of:
      0.07506842 = sum of:
        0.07506842 = weight(_text_:index in 2287) [ClassicSimilarity], result of:
          0.07506842 = score(doc=2287,freq=2.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.33795667 = fieldWeight in 2287, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2287)
      0.33333334 = coord(1/3)
    
    Abstract
    The depth of indexing, the number of terms assigned on average to each document in a retrieval system as entry points, has a significantly effect on the standard retrieval performance measures in modern commercial retrieval systems, just as it did in previous experimental work. Tests on the effect of basic index search, as opposed to controlled vocabulary search, in these real systems are quite different than traditional comparisons of free text searching with controlled vocabulary searching. In modern commercial systems the controlled vocabulary serves as a precision device, since the strucure of the default for unqualified search terms in these systems requires that it do so.
  17. Subrahmanyam, B.: Library of Congress Classification numbers : issues of consistency and their implications for union catalogs (2006) 0.02
    0.0249044 = product of:
      0.0747132 = sum of:
        0.0747132 = sum of:
          0.040277936 = weight(_text_:classification in 5784) [ClassicSimilarity], result of:
            0.040277936 = score(doc=5784,freq=4.0), product of:
              0.16188543 = queryWeight, product of:
                3.1847067 = idf(docFreq=4974, maxDocs=44218)
                0.05083213 = queryNorm
              0.24880521 = fieldWeight in 5784, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.1847067 = idf(docFreq=4974, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5784)
          0.03443526 = weight(_text_:22 in 5784) [ClassicSimilarity], result of:
            0.03443526 = score(doc=5784,freq=2.0), product of:
              0.17800546 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05083213 = queryNorm
              0.19345059 = fieldWeight in 5784, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5784)
      0.33333334 = coord(1/3)
    
    Abstract
    This study examined Library of Congress Classification (LCC)-based class numbers assigned to a representative sample of 200 titles in 52 American library systems to determine the level of consistency within and across those systems. The results showed that under the condition that a library system has a title, the probability of that title having the same LCC-based class number across library systems is greater than 85 percent. An examination of 121 titles displaying variations in class numbers among library systems showed certain titles (for example, multi-foci titles, titles in series, bibliographies, and fiction) lend themselves to alternate class numbers. Others were assigned variant numbers either due to latitude in the schedules or for reasons that cannot be pinpointed. With increasing dependence on copy cataloging, the size of such variations may continue to decrease. As the preferred class number with its alternates represents a title more fully than just the preferred class number, this paper argues for continued use of alternates by library systems and for finding a method to link alternate class numbers to preferred class numbers for enriched subject access through local and union catalogs.
    Date
    10. 9.2000 17:38:22
  18. Hudon, M.: Conceptual compatibility in controlled language tools used to index and access the content of moving image collections (2004) 0.02
    0.02144812 = product of:
      0.06434436 = sum of:
        0.06434436 = weight(_text_:index in 2655) [ClassicSimilarity], result of:
          0.06434436 = score(doc=2655,freq=2.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.28967714 = fieldWeight in 2655, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.046875 = fieldNorm(doc=2655)
      0.33333334 = coord(1/3)
    
  19. Wolfram, D.; Zhang, J.: ¬An investigation of the influence of indexing exhaustivity and term distributions on a document space (2002) 0.02
    0.017873434 = product of:
      0.0536203 = sum of:
        0.0536203 = weight(_text_:index in 5238) [ClassicSimilarity], result of:
          0.0536203 = score(doc=5238,freq=2.0), product of:
            0.2221244 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.05083213 = queryNorm
            0.24139762 = fieldWeight in 5238, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5238)
      0.33333334 = coord(1/3)
    
    Abstract
    Wolfram and Zhang are interested in the effect of different indexing exhaustivity, by which they mean the number of terms chosen, and of different index term distributions and different term weighting methods on the resulting document cluster organization. The Distance Angle Retrieval Environment, DARE, which provides a two dimensional display of retrieved documents was used to represent the document clusters based upon a document's distance from the searcher's main interest, and on the angle formed by the document, a point representing a minor interest, and the point representing the main interest. If the centroid and the origin of the document space are assigned as major and minor points the average distance between documents and the centroid can be measured providing an indication of cluster organization. in the form of a size normalized similarity measure. Using 500 records from NTIS and nine models created by intersecting low, observed, and high exhaustivity levels (based upon a negative binomial distribution) with shallow, observed, and steep term distributions (based upon a Zipf distribution) simulation runs were preformed using inverse document frequency, inter-document term frequency, and inverse document frequency based upon both inter and intra-document frequencies. Low exhaustivity and shallow distributions result in a more dense document space and less effective retrieval. High exhaustivity and steeper distributions result in a more diffuse space.
  20. Morris, L.R.: ¬The frequency of use of Library of Congress Classification numbers and Dewey Decimal Classification numbers in the MARC file in the field of library science (1991) 0.02
    0.016278135 = product of:
      0.048834402 = sum of:
        0.048834402 = product of:
          0.097668804 = sum of:
            0.097668804 = weight(_text_:classification in 2308) [ClassicSimilarity], result of:
              0.097668804 = score(doc=2308,freq=12.0), product of:
                0.16188543 = queryWeight, product of:
                  3.1847067 = idf(docFreq=4974, maxDocs=44218)
                  0.05083213 = queryNorm
                0.60332054 = fieldWeight in 2308, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.1847067 = idf(docFreq=4974, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2308)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The LCC and DDC systems were devised and updated by librarians who had and have no access to the eventual frequency of use of each number in those classification systems. 80% of the monographs in a MARC file of over 1.000.000 records are classified into 20% of the classification numbers in the field of library science and only 20% of the mongraphs are classified into 80% of the classification numbers in the field of library science. Classification of monographs coulld be made easier and performed more accurately if many of the little used and unused numbers were eliminated and many of the most crowded numbers were expanded. A number of examples are included