Search (22 results, page 1 of 2)

  • × author_ss:"Bookstein, A."
  1. Bookstein, A.: Informetric distributions : I. Unified overview (1990) 0.05
    0.05037771 = product of:
      0.10075542 = sum of:
        0.10075542 = sum of:
          0.01339476 = weight(_text_:a in 6902) [ClassicSimilarity], result of:
            0.01339476 = score(doc=6902,freq=4.0), product of:
              0.053105544 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046056706 = queryNorm
              0.25222903 = fieldWeight in 6902, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.109375 = fieldNorm(doc=6902)
          0.087360665 = weight(_text_:22 in 6902) [ClassicSimilarity], result of:
            0.087360665 = score(doc=6902,freq=2.0), product of:
              0.16128273 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046056706 = queryNorm
              0.5416616 = fieldWeight in 6902, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.109375 = fieldNorm(doc=6902)
      0.5 = coord(1/2)
    
    Date
    22. 7.2006 18:55:29
    Type
    a
  2. Bookstein, A.: Informetric distributions : II. Resilience to ambiguity (1990) 0.05
    0.05037771 = product of:
      0.10075542 = sum of:
        0.10075542 = sum of:
          0.01339476 = weight(_text_:a in 4689) [ClassicSimilarity], result of:
            0.01339476 = score(doc=4689,freq=4.0), product of:
              0.053105544 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046056706 = queryNorm
              0.25222903 = fieldWeight in 4689, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.109375 = fieldNorm(doc=4689)
          0.087360665 = weight(_text_:22 in 4689) [ClassicSimilarity], result of:
            0.087360665 = score(doc=4689,freq=2.0), product of:
              0.16128273 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046056706 = queryNorm
              0.5416616 = fieldWeight in 4689, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.109375 = fieldNorm(doc=4689)
      0.5 = coord(1/2)
    
    Date
    22. 7.2006 18:55:55
    Type
    a
  3. Bookstein, A.; Klein, S.T.: Compression, information theory, and grammars : a unified approach (1990) 0.00
    0.0046871896 = product of:
      0.009374379 = sum of:
        0.009374379 = product of:
          0.018748758 = sum of:
            0.018748758 = weight(_text_:a in 2970) [ClassicSimilarity], result of:
              0.018748758 = score(doc=2970,freq=6.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.3530471 = fieldWeight in 2970, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=2970)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  4. Bookstein, A.: Probability and Fuzzy-set applications to information retrieval (1985) 0.00
    0.003827074 = product of:
      0.007654148 = sum of:
        0.007654148 = product of:
          0.015308296 = sum of:
            0.015308296 = weight(_text_:a in 781) [ClassicSimilarity], result of:
              0.015308296 = score(doc=781,freq=4.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.28826174 = fieldWeight in 781, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=781)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  5. Bookstein, A.: Relevance (1979) 0.00
    0.003827074 = product of:
      0.007654148 = sum of:
        0.007654148 = product of:
          0.015308296 = sum of:
            0.015308296 = weight(_text_:a in 839) [ClassicSimilarity], result of:
              0.015308296 = score(doc=839,freq=4.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.28826174 = fieldWeight in 839, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=839)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  6. Bookstein, A.; Swanson, D.R.: Probabilistic models for automatic indexing (1974) 0.00
    0.003827074 = product of:
      0.007654148 = sum of:
        0.007654148 = product of:
          0.015308296 = sum of:
            0.015308296 = weight(_text_:a in 5466) [ClassicSimilarity], result of:
              0.015308296 = score(doc=5466,freq=4.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.28826174 = fieldWeight in 5466, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=5466)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  7. Bookstein, A.: Scientometrics: new opportunities (1994) 0.00
    0.003827074 = product of:
      0.007654148 = sum of:
        0.007654148 = product of:
          0.015308296 = sum of:
            0.015308296 = weight(_text_:a in 5062) [ClassicSimilarity], result of:
              0.015308296 = score(doc=5062,freq=4.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.28826174 = fieldWeight in 5062, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=5062)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  8. Bookstein, A.: Set-oriented retrieval (1989) 0.00
    0.003827074 = product of:
      0.007654148 = sum of:
        0.007654148 = product of:
          0.015308296 = sum of:
            0.015308296 = weight(_text_:a in 203) [ClassicSimilarity], result of:
              0.015308296 = score(doc=203,freq=4.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.28826174 = fieldWeight in 203, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=203)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  9. Bookstein, A.: Fuzzy requests : an approach to weighted Boolean searches (1979) 0.00
    0.0035799001 = product of:
      0.0071598003 = sum of:
        0.0071598003 = product of:
          0.014319601 = sum of:
            0.014319601 = weight(_text_:a in 5504) [ClassicSimilarity], result of:
              0.014319601 = score(doc=5504,freq=14.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.26964417 = fieldWeight in 5504, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5504)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This article concerns the problem of how to permit a patron tp represent the relatice importance of various index terms in a Boolean request while retaining the desirable properties of a Boolean system. The character of classical Boolean systems is reviewed and related to the notion of fuzzy sets. The fuzzy set concept then forms the basis of the concept of a fuzzy request in which weights are assigned to index terms. The properties of such a system are discussed, and it is shown that such systems retain the manipulability of traditional Boolena requests
    Type
    a
  10. Bookstein, A.: ¬The bibliometric distributions (1976) 0.00
    0.0035799001 = product of:
      0.0071598003 = sum of:
        0.0071598003 = product of:
          0.014319601 = sum of:
            0.014319601 = weight(_text_:a in 5061) [ClassicSimilarity], result of:
              0.014319601 = score(doc=5061,freq=14.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.26964417 = fieldWeight in 5061, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5061)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    One of the most surprising findings in the information sciences is the recurrence of a small number of frequency distributions. In this paper, these distributions are described, and a point of view is adopted that allows us to understand them a being different versions of a single distribution. The empirical distributions are shown to be special cases of a single theoretic distribution. It is found that when random fluctuations are introduced, the distributions are not strongly influenced
    Type
    a
  11. Bookstein, A.; Swanson, D.R.: ¬A decision theoretic foundation for indexing (1975) 0.00
    0.0031324127 = product of:
      0.0062648254 = sum of:
        0.0062648254 = product of:
          0.012529651 = sum of:
            0.012529651 = weight(_text_:a in 76) [ClassicSimilarity], result of:
              0.012529651 = score(doc=76,freq=14.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.23593865 = fieldWeight in 76, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=76)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The indexing of a document is among the most crucial steps in preparing that document for retrieval. The adequacy of the indexing determines the ability of the system to respond to patron requests. This paper discusses this process, and document retrieval in general, on the basis of formal decision theory. The basic theoretical approach taken is illustrated by means of a model of word occurrences in documents in the context of a model information system; both models are fully defined in this paper. Through the main purpose of this papers is to provide insights into a very complex process, formulae are developed that may prove to be of value for an automated operating system. The paper concludes with an interpretation of recall and precision curves as seen from the point of view of decision theory
    Type
    a
  12. Bookstein, A.; Cooper, W.: ¬A general mathematical model for information retrieval systems (1976) 0.00
    0.0029000505 = product of:
      0.005800101 = sum of:
        0.005800101 = product of:
          0.011600202 = sum of:
            0.011600202 = weight(_text_:a in 5461) [ClassicSimilarity], result of:
              0.011600202 = score(doc=5461,freq=12.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.21843673 = fieldWeight in 5461, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5461)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper presents a mathematical model of an information retrieval system thought to be general enough to serve as an abstract representation of most document and reference retrieval systems. The model is made up of four components that, in one form or another, appear in every functioning system. It is proved that the basic organization of documents that the system provides for a user on receipt of a request follows from the properties and interrelations of the four components. Each component is then discussed in turn and it is seen that much of the existing theory regarding information systems can be viewed as an elaboration of this model
    Type
    a
  13. Bookstein, A.; Kulyukin, V.; Raita, T.; Nicholson, J.: Adapting measures of clumping strength to assess term-term similarity (2003) 0.00
    0.0028703054 = product of:
      0.005740611 = sum of:
        0.005740611 = product of:
          0.011481222 = sum of:
            0.011481222 = weight(_text_:a in 1609) [ClassicSimilarity], result of:
              0.011481222 = score(doc=1609,freq=16.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.2161963 = fieldWeight in 1609, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1609)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Automated information retrieval relies heavily an statistical regularities that emerge as terms are deposited to produce text. This paper examines statistical patterns expected of a pair of terms that are semantically related to each other. Guided by a conceptualization of the text generation process, we derive measures of how tightly two terms are semantically associated. Our main objective is to probe whether such measures yield reasonable results. Specifically, we examine how the tendency of a content bearing term to clump, as quantified by previously developed measures of term clumping, is influenced by the presence of other terms. This approach allows us to present a toolkit from which a range of measures can be constructed. As an illustration, one of several suggested measures is evaluated an a large text corpus built from an on-line encyclopedia.
    Type
    a
  14. Bookstein, A.: Informetric distributions : III. Ambiguity (1997) 0.00
    0.00270615 = product of:
      0.0054123 = sum of:
        0.0054123 = product of:
          0.0108246 = sum of:
            0.0108246 = weight(_text_:a in 6489) [ClassicSimilarity], result of:
              0.0108246 = score(doc=6489,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.20383182 = fieldWeight in 6489, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6489)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This article examines various kinds of uncertainty. The notion of ambiguity is defined and contratsed with the more familiar notions of randomness and fuzziness. Functional forms resistant to ambiguity are defined, and it is shown how to incorporate a random component, that is itself also rsistant to ambiguity, into a resilent, but deterministic model
    Type
    a
  15. Bookstein, A.; Moed, H.; Yitzahki, M.: Measures of international collaboration in scientific literature : part I (2006) 0.00
    0.00270615 = product of:
      0.0054123 = sum of:
        0.0054123 = product of:
          0.0108246 = sum of:
            0.0108246 = weight(_text_:a in 985) [ClassicSimilarity], result of:
              0.0108246 = score(doc=985,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.20383182 = fieldWeight in 985, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=985)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Research evaluating models of scientific productivity require coherent metrics that quantify various key relations among papers as revealed by patterns of citation. This paper focuses on the various conceptual problems inherent in measuring the degree to which papers tend to cite other papers written by authors of the same nationality. We suggest that measures can be given a degree of assurance of coherence by being based on mathematical models describing the citation process. A number of such models are developed.
    Type
    a
  16. Bookstein, A.: Bibliocryptography (1996) 0.00
    0.0026473717 = product of:
      0.0052947435 = sum of:
        0.0052947435 = product of:
          0.010589487 = sum of:
            0.010589487 = weight(_text_:a in 6502) [ClassicSimilarity], result of:
              0.010589487 = score(doc=6502,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.19940455 = fieldWeight in 6502, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=6502)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Because of concerns about the privacy of its patrons, it is common for libraries to systematically destroy historic information about book circulation. I argue that this information has great potential value for improving retrieval effectiveness, and give 2 examples of how this information can be used. Further, I show how use-data can be preserved and exploited while still giving a high degree of protection for patron privacy. The methods are analyzed and formulae are derived indicating the tradeoff between retrieval effectiveness and security. A second, contrasting application, indicating how to introduce 'fingerprints' into digitized audio-visual material in a tamper-resistant manner, is described
    Type
    a
  17. Bookstein, A.; Klein, S.T.; Raita, T.: Clumping properties of content-bearing words (1998) 0.00
    0.0023678814 = product of:
      0.0047357627 = sum of:
        0.0047357627 = product of:
          0.009471525 = sum of:
            0.009471525 = weight(_text_:a in 442) [ClassicSimilarity], result of:
              0.009471525 = score(doc=442,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.17835285 = fieldWeight in 442, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=442)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Information Retrieval Systems identify content bearing words, and possibly also assign weights, as part of the process of formulating requests. For optimal retrieval efficiency, it is desirable that this be done automatically. This article defines the notion of serial clustering of words in text, and explores the value of such clustering as an indicator of a word's bearing content. This approach is flexible in the sense that it is sensitive to context: a term may be assessed as content-bearing within one collection, but not another. Our approach, being numerical, may also be of value in assigning weights to terms in requests. Experimental support is obtained from natural text databases in three different languages
    Type
    a
  18. Bookstein, A.; Raita, T.: Discovering term occurence structure in text (2001) 0.00
    0.0023678814 = product of:
      0.0047357627 = sum of:
        0.0047357627 = product of:
          0.009471525 = sum of:
            0.009471525 = weight(_text_:a in 5751) [ClassicSimilarity], result of:
              0.009471525 = score(doc=5751,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.17835285 = fieldWeight in 5751, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5751)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This article examines some consequences for information control of the tendency of occurrences of contentbearing terms to appear together, or clump. Properties of previously defined clumping measures are reviewed and extended, and the significance of these measures for devising retrieval strategies discussed. A new type of clumping measure, which extends the earlier measures by permitting gaps within a clump, is defined, and several variants examined. Experiments are carried out that indicate the relation between the new measure and one of the earlier measures, as well as the ability of the two types of measure to predict compression efficiency
    Type
    a
  19. Bookstein, A.; Moed, H.; Yitzahki, M.: Measures of international collaboration in scientific literature : part II (2006) 0.00
    0.0023678814 = product of:
      0.0047357627 = sum of:
        0.0047357627 = product of:
          0.009471525 = sum of:
            0.009471525 = weight(_text_:a in 989) [ClassicSimilarity], result of:
              0.009471525 = score(doc=989,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.17835285 = fieldWeight in 989, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=989)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper continues the attempt of Part I to develop a coherent family of measures of influence between classes of documents, for example, language or nationality classes, as indicated by citation choice. In this paper we focus on situations in which there is some ambiguity as to how to assign items to a class. For simplicity, we change our focus from citations to co-authorship patterns, restricting most of our discussion to papers with two authors. Like the earlier paper, we propose very simple models of the citation decision, and base our measures on the parameters that appear in the model.
    Type
    a
  20. Bookstein, A.; Wright, B.: Ambiguity in measurement (1997) 0.00
    0.0023435948 = product of:
      0.0046871896 = sum of:
        0.0046871896 = product of:
          0.009374379 = sum of:
            0.009374379 = weight(_text_:a in 1036) [ClassicSimilarity], result of:
              0.009374379 = score(doc=1036,freq=6.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.17652355 = fieldWeight in 1036, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1036)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Gives an overview of the role of ambiguity in measurement and explores analytical methods for exploring its impact. Argues that certain functional forms are more resilient than others to problems of ambiguity, and that these should be preferred when ambiguity is a serious concern
    Type
    a