Search (20 results, page 1 of 1)

  • × author_ss:"Bookstein, A."
  1. Bookstein, A.: Probability and Fuzzy-set applications to information retrieval (1985) 0.01
    0.0067306077 = product of:
      0.02692243 = sum of:
        0.02692243 = weight(_text_:information in 781) [ClassicSimilarity], result of:
          0.02692243 = score(doc=781,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.43886948 = fieldWeight in 781, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=781)
      0.25 = coord(1/4)
    
    Source
    Annual review of information science and technology. 20(1985), S.117-151
  2. Bookstein, A.; Klein, S.T.: Compression, information theory, and grammars : a unified approach (1990) 0.01
    0.0067306077 = product of:
      0.02692243 = sum of:
        0.02692243 = weight(_text_:information in 2970) [ClassicSimilarity], result of:
          0.02692243 = score(doc=2970,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.43886948 = fieldWeight in 2970, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=2970)
      0.25 = coord(1/4)
    
    Source
    ACM transaction on information systems. 8(1990), S.27-49
  3. Bookstein, A.: Relevance (1979) 0.00
    0.004759258 = product of:
      0.019037032 = sum of:
        0.019037032 = weight(_text_:information in 839) [ClassicSimilarity], result of:
          0.019037032 = score(doc=839,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.3103276 = fieldWeight in 839, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=839)
      0.25 = coord(1/4)
    
    Source
    Journal of the American Society for Information Science. 30(1979), S.269-273
  4. Bookstein, A.; Swanson, D.R.: Probabilistic models for automatic indexing (1974) 0.00
    0.004759258 = product of:
      0.019037032 = sum of:
        0.019037032 = weight(_text_:information in 5466) [ClassicSimilarity], result of:
          0.019037032 = score(doc=5466,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.3103276 = fieldWeight in 5466, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=5466)
      0.25 = coord(1/4)
    
    Source
    Journal of the American Society for Information Science. 25(1974), S.312-318
  5. Bookstein, A.: Set-oriented retrieval (1989) 0.00
    0.004759258 = product of:
      0.019037032 = sum of:
        0.019037032 = weight(_text_:information in 203) [ClassicSimilarity], result of:
          0.019037032 = score(doc=203,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.3103276 = fieldWeight in 203, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=203)
      0.25 = coord(1/4)
    
    Source
    Information processing and management. 25(1989) no.5, S.465-475
  6. Bookstein, A.: Informetric distributions : I. Unified overview (1990) 0.00
    0.004164351 = product of:
      0.016657405 = sum of:
        0.016657405 = weight(_text_:information in 6902) [ClassicSimilarity], result of:
          0.016657405 = score(doc=6902,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.27153665 = fieldWeight in 6902, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=6902)
      0.25 = coord(1/4)
    
    Source
    Journal of the American Society for Information Science. 41(1990) no.5, S.368-375
  7. Bookstein, A.: Bibliocryptography (1996) 0.00
    0.004164351 = product of:
      0.016657405 = sum of:
        0.016657405 = weight(_text_:information in 6502) [ClassicSimilarity], result of:
          0.016657405 = score(doc=6502,freq=8.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.27153665 = fieldWeight in 6502, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=6502)
      0.25 = coord(1/4)
    
    Abstract
    Because of concerns about the privacy of its patrons, it is common for libraries to systematically destroy historic information about book circulation. I argue that this information has great potential value for improving retrieval effectiveness, and give 2 examples of how this information can be used. Further, I show how use-data can be preserved and exploited while still giving a high degree of protection for patron privacy. The methods are analyzed and formulae are derived indicating the tradeoff between retrieval effectiveness and security. A second, contrasting application, indicating how to introduce 'fingerprints' into digitized audio-visual material in a tamper-resistant manner, is described
    Source
    Journal of the American Society for Information Science. 47(1996) no.12, S.886-895
  8. Bookstein, A.: Informetric distributions : II. Resilience to ambiguity (1990) 0.00
    0.004164351 = product of:
      0.016657405 = sum of:
        0.016657405 = weight(_text_:information in 4689) [ClassicSimilarity], result of:
          0.016657405 = score(doc=4689,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.27153665 = fieldWeight in 4689, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=4689)
      0.25 = coord(1/4)
    
    Source
    Journal of the American Society for Information Science. 41(1990) no.5, S.376-386
  9. Bookstein, A.: Implications of ambiguity for scientometric measurement (2001) 0.00
    0.004121639 = product of:
      0.016486555 = sum of:
        0.016486555 = weight(_text_:information in 5199) [ClassicSimilarity], result of:
          0.016486555 = score(doc=5199,freq=6.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.2687516 = fieldWeight in 5199, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=5199)
      0.25 = coord(1/4)
    
    Abstract
    Finally, Bookstein points to the ambiguity of our measurements that seems to present a structural impediment to the development of social science theory. Our theory always seems to be at an early stage, information science still at the frontier.
    Footnote
    Beitrag eines Themenheftes: Still the Frontier: Information Science at the Millenium
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.1, S.74-79
  10. Bookstein, A.; Cooper, W.: ¬A general mathematical model for information retrieval systems (1976) 0.00
    0.003606434 = product of:
      0.014425736 = sum of:
        0.014425736 = weight(_text_:information in 5461) [ClassicSimilarity], result of:
          0.014425736 = score(doc=5461,freq=6.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.23515764 = fieldWeight in 5461, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5461)
      0.25 = coord(1/4)
    
    Abstract
    This paper presents a mathematical model of an information retrieval system thought to be general enough to serve as an abstract representation of most document and reference retrieval systems. The model is made up of four components that, in one form or another, appear in every functioning system. It is proved that the basic organization of documents that the system provides for a user on receipt of a request follows from the properties and interrelations of the four components. Each component is then discussed in turn and it is seen that much of the existing theory regarding information systems can be viewed as an elaboration of this model
  11. Bookstein, A.; Swanson, D.R.: ¬A decision theoretic foundation for indexing (1975) 0.00
    0.0029446408 = product of:
      0.011778563 = sum of:
        0.011778563 = weight(_text_:information in 76) [ClassicSimilarity], result of:
          0.011778563 = score(doc=76,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.1920054 = fieldWeight in 76, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=76)
      0.25 = coord(1/4)
    
    Abstract
    The indexing of a document is among the most crucial steps in preparing that document for retrieval. The adequacy of the indexing determines the ability of the system to respond to patron requests. This paper discusses this process, and document retrieval in general, on the basis of formal decision theory. The basic theoretical approach taken is illustrated by means of a model of word occurrences in documents in the context of a model information system; both models are fully defined in this paper. Through the main purpose of this papers is to provide insights into a very complex process, formulae are developed that may prove to be of value for an automated operating system. The paper concludes with an interpretation of recall and precision curves as seen from the point of view of decision theory
    Source
    Journal of the American Society for Information Science. 26(1975), S.45-50
  12. Bookstein, A.; Klein, S.T.; Raita, T.: Clumping properties of content-bearing words (1998) 0.00
    0.0029446408 = product of:
      0.011778563 = sum of:
        0.011778563 = weight(_text_:information in 442) [ClassicSimilarity], result of:
          0.011778563 = score(doc=442,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.1920054 = fieldWeight in 442, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=442)
      0.25 = coord(1/4)
    
    Abstract
    Information Retrieval Systems identify content bearing words, and possibly also assign weights, as part of the process of formulating requests. For optimal retrieval efficiency, it is desirable that this be done automatically. This article defines the notion of serial clustering of words in text, and explores the value of such clustering as an indicator of a word's bearing content. This approach is flexible in the sense that it is sensitive to context: a term may be assessed as content-bearing within one collection, but not another. Our approach, being numerical, may also be of value in assigning weights to terms in requests. Experimental support is obtained from natural text databases in three different languages
    Source
    Journal of the American Society for Information Science. 49(1998) no.2, S.102-114
  13. Bookstein, A.; Raita, T.: Discovering term occurence structure in text (2001) 0.00
    0.0029446408 = product of:
      0.011778563 = sum of:
        0.011778563 = weight(_text_:information in 5751) [ClassicSimilarity], result of:
          0.011778563 = score(doc=5751,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.1920054 = fieldWeight in 5751, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5751)
      0.25 = coord(1/4)
    
    Abstract
    This article examines some consequences for information control of the tendency of occurrences of contentbearing terms to appear together, or clump. Properties of previously defined clumping measures are reviewed and extended, and the significance of these measures for devising retrieval strategies discussed. A new type of clumping measure, which extends the earlier measures by permitting gaps within a clump, is defined, and several variants examined. Experiments are carried out that indicate the relation between the new measure and one of the earlier measures, as well as the ability of the two types of measure to predict compression efficiency
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.6, S.476-486
  14. Swanson, D.R.; Smalheiser, N.R.; Bookstein, A.: Information discovery from complementary literatures : categorizing viruses as potential weapons (2001) 0.00
    0.0025760243 = product of:
      0.010304097 = sum of:
        0.010304097 = weight(_text_:information in 6513) [ClassicSimilarity], result of:
          0.010304097 = score(doc=6513,freq=6.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.16796975 = fieldWeight in 6513, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6513)
      0.25 = coord(1/4)
    
    Abstract
    Using novel informatics techniques to process the Output of Medline searches, we have generated a list of viruses that may have the potential for development as weapons. Our findings are intended as a guide to the virus literature to support further studies that might then lead to appropriate defense and public health measures. This article stresses methods that are more generally relevant to information science. Initial Medline searches identified two kinds of virus literaturesthe first concerning the genetic aspects of virulence, and the second concerning the transmission of viral diseases. Both literatures taken together are of central importance in identifying research relevant to the development of biological weapons. Yet, the two literatures had very few articles in common. We downloaded the Medline records for each of the two literatures and used a computer to extract all virus terms common to both. The fact that the resulting virus list includes most of an earlier independently published list of viruses considered by military experts to have the highest threat as potential biological weapons served as a test of the method; the test outcome showed a high degree of statistical significance, thus supporting an inference that the new viruses an the list share certain important characteristics with viruses of known biological
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.10, S.797-812
  15. Bookstein, A.; Kulyukin, V.; Raita, T.; Nicholson, J.: Adapting measures of clumping strength to assess term-term similarity (2003) 0.00
    0.0025239778 = product of:
      0.010095911 = sum of:
        0.010095911 = weight(_text_:information in 1609) [ClassicSimilarity], result of:
          0.010095911 = score(doc=1609,freq=4.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.16457605 = fieldWeight in 1609, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1609)
      0.25 = coord(1/4)
    
    Abstract
    Automated information retrieval relies heavily an statistical regularities that emerge as terms are deposited to produce text. This paper examines statistical patterns expected of a pair of terms that are semantically related to each other. Guided by a conceptualization of the text generation process, we derive measures of how tightly two terms are semantically associated. Our main objective is to probe whether such measures yield reasonable results. Specifically, we examine how the tendency of a content bearing term to clump, as quantified by previously developed measures of term clumping, is influenced by the presence of other terms. This approach allows us to present a toolkit from which a range of measures can be constructed. As an illustration, one of several suggested measures is evaluated an a large text corpus built from an on-line encyclopedia.
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.7, S.611-620
  16. Bookstein, A.: Fuzzy requests : an approach to weighted Boolean searches (1979) 0.00
    0.002379629 = product of:
      0.009518516 = sum of:
        0.009518516 = weight(_text_:information in 5504) [ClassicSimilarity], result of:
          0.009518516 = score(doc=5504,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.1551638 = fieldWeight in 5504, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=5504)
      0.25 = coord(1/4)
    
    Source
    Journal of the American Society for Information Science. 30(1979), S.240-247
  17. Bookstein, A.: ¬The bibliometric distributions (1976) 0.00
    0.002379629 = product of:
      0.009518516 = sum of:
        0.009518516 = weight(_text_:information in 5061) [ClassicSimilarity], result of:
          0.009518516 = score(doc=5061,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.1551638 = fieldWeight in 5061, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=5061)
      0.25 = coord(1/4)
    
    Abstract
    One of the most surprising findings in the information sciences is the recurrence of a small number of frequency distributions. In this paper, these distributions are described, and a point of view is adopted that allows us to understand them a being different versions of a single distribution. The empirical distributions are shown to be special cases of a single theoretic distribution. It is found that when random fluctuations are introduced, the distributions are not strongly influenced
  18. Bookstein, A.: Informetric distributions : III. Ambiguity (1997) 0.00
    0.002379629 = product of:
      0.009518516 = sum of:
        0.009518516 = weight(_text_:information in 6489) [ClassicSimilarity], result of:
          0.009518516 = score(doc=6489,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.1551638 = fieldWeight in 6489, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=6489)
      0.25 = coord(1/4)
    
    Source
    Journal of the American Society for Information Science. 48(1997) no.1, S.2-10
  19. Bookstein, A.; Moed, H.; Yitzahki, M.: Measures of international collaboration in scientific literature : part I (2006) 0.00
    0.002379629 = product of:
      0.009518516 = sum of:
        0.009518516 = weight(_text_:information in 985) [ClassicSimilarity], result of:
          0.009518516 = score(doc=985,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.1551638 = fieldWeight in 985, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=985)
      0.25 = coord(1/4)
    
    Source
    Information processing and management. 42(2006) no.6, S.1408-1421
  20. Bookstein, A.; Moed, H.; Yitzahki, M.: Measures of international collaboration in scientific literature : part II (2006) 0.00
    0.0020821756 = product of:
      0.008328702 = sum of:
        0.008328702 = weight(_text_:information in 989) [ClassicSimilarity], result of:
          0.008328702 = score(doc=989,freq=2.0), product of:
            0.06134496 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.034944877 = queryNorm
            0.13576832 = fieldWeight in 989, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=989)
      0.25 = coord(1/4)
    
    Source
    Information processing and management. 42(2006) no.6, S.1422-1427