Search (10 results, page 1 of 1)

  • × author_ss:"Bookstein, A."
  1. Bookstein, A.: Informetric distributions : I. Unified overview (1990) 0.02
    0.02475569 = product of:
      0.04951138 = sum of:
        0.04951138 = product of:
          0.09902276 = sum of:
            0.09902276 = weight(_text_:22 in 6902) [ClassicSimilarity], result of:
              0.09902276 = score(doc=6902,freq=2.0), product of:
                0.18281296 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052204985 = queryNorm
                0.5416616 = fieldWeight in 6902, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6902)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 7.2006 18:55:29
  2. Bookstein, A.: Informetric distributions : II. Resilience to ambiguity (1990) 0.02
    0.02475569 = product of:
      0.04951138 = sum of:
        0.04951138 = product of:
          0.09902276 = sum of:
            0.09902276 = weight(_text_:22 in 4689) [ClassicSimilarity], result of:
              0.09902276 = score(doc=4689,freq=2.0), product of:
                0.18281296 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052204985 = queryNorm
                0.5416616 = fieldWeight in 4689, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4689)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 7.2006 18:55:55
  3. Bookstein, A.: Probability and Fuzzy-set applications to information retrieval (1985) 0.02
    0.021110734 = product of:
      0.042221468 = sum of:
        0.042221468 = product of:
          0.084442936 = sum of:
            0.084442936 = weight(_text_:retrieval in 781) [ClassicSimilarity], result of:
              0.084442936 = score(doc=781,freq=2.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.5347345 = fieldWeight in 781, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.125 = fieldNorm(doc=781)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  4. Bookstein, A.: Set-oriented retrieval (1989) 0.02
    0.021110734 = product of:
      0.042221468 = sum of:
        0.042221468 = product of:
          0.084442936 = sum of:
            0.084442936 = weight(_text_:retrieval in 203) [ClassicSimilarity], result of:
              0.084442936 = score(doc=203,freq=2.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.5347345 = fieldWeight in 203, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.125 = fieldNorm(doc=203)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  5. Bookstein, A.; Cooper, W.: ¬A general mathematical model for information retrieval systems (1976) 0.02
    0.015997129 = product of:
      0.031994257 = sum of:
        0.031994257 = product of:
          0.063988514 = sum of:
            0.063988514 = weight(_text_:retrieval in 5461) [ClassicSimilarity], result of:
              0.063988514 = score(doc=5461,freq=6.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.40520695 = fieldWeight in 5461, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5461)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper presents a mathematical model of an information retrieval system thought to be general enough to serve as an abstract representation of most document and reference retrieval systems. The model is made up of four components that, in one form or another, appear in every functioning system. It is proved that the basic organization of documents that the system provides for a user on receipt of a request follows from the properties and interrelations of the four components. Each component is then discussed in turn and it is seen that much of the existing theory regarding information systems can be viewed as an elaboration of this model
  6. Bookstein, A.; Swanson, D.R.: ¬A decision theoretic foundation for indexing (1975) 0.01
    0.0130616 = product of:
      0.0261232 = sum of:
        0.0261232 = product of:
          0.0522464 = sum of:
            0.0522464 = weight(_text_:retrieval in 76) [ClassicSimilarity], result of:
              0.0522464 = score(doc=76,freq=4.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.33085006 = fieldWeight in 76, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=76)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The indexing of a document is among the most crucial steps in preparing that document for retrieval. The adequacy of the indexing determines the ability of the system to respond to patron requests. This paper discusses this process, and document retrieval in general, on the basis of formal decision theory. The basic theoretical approach taken is illustrated by means of a model of word occurrences in documents in the context of a model information system; both models are fully defined in this paper. Through the main purpose of this papers is to provide insights into a very complex process, formulae are developed that may prove to be of value for an automated operating system. The paper concludes with an interpretation of recall and precision curves as seen from the point of view of decision theory
  7. Bookstein, A.: Bibliocryptography (1996) 0.01
    0.0130616 = product of:
      0.0261232 = sum of:
        0.0261232 = product of:
          0.0522464 = sum of:
            0.0522464 = weight(_text_:retrieval in 6502) [ClassicSimilarity], result of:
              0.0522464 = score(doc=6502,freq=4.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.33085006 = fieldWeight in 6502, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=6502)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Because of concerns about the privacy of its patrons, it is common for libraries to systematically destroy historic information about book circulation. I argue that this information has great potential value for improving retrieval effectiveness, and give 2 examples of how this information can be used. Further, I show how use-data can be preserved and exploited while still giving a high degree of protection for patron privacy. The methods are analyzed and formulae are derived indicating the tradeoff between retrieval effectiveness and security. A second, contrasting application, indicating how to introduce 'fingerprints' into digitized audio-visual material in a tamper-resistant manner, is described
  8. Bookstein, A.; Klein, S.T.; Raita, T.: Clumping properties of content-bearing words (1998) 0.01
    0.0130616 = product of:
      0.0261232 = sum of:
        0.0261232 = product of:
          0.0522464 = sum of:
            0.0522464 = weight(_text_:retrieval in 442) [ClassicSimilarity], result of:
              0.0522464 = score(doc=442,freq=4.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.33085006 = fieldWeight in 442, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=442)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Information Retrieval Systems identify content bearing words, and possibly also assign weights, as part of the process of formulating requests. For optimal retrieval efficiency, it is desirable that this be done automatically. This article defines the notion of serial clustering of words in text, and explores the value of such clustering as an indicator of a word's bearing content. This approach is flexible in the sense that it is sensitive to context: a term may be assessed as content-bearing within one collection, but not another. Our approach, being numerical, may also be of value in assigning weights to terms in requests. Experimental support is obtained from natural text databases in three different languages
  9. Bookstein, A.; Raita, T.: Discovering term occurence structure in text (2001) 0.01
    0.009235946 = product of:
      0.018471893 = sum of:
        0.018471893 = product of:
          0.036943786 = sum of:
            0.036943786 = weight(_text_:retrieval in 5751) [ClassicSimilarity], result of:
              0.036943786 = score(doc=5751,freq=2.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.23394634 = fieldWeight in 5751, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5751)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This article examines some consequences for information control of the tendency of occurrences of contentbearing terms to appear together, or clump. Properties of previously defined clumping measures are reviewed and extended, and the significance of these measures for devising retrieval strategies discussed. A new type of clumping measure, which extends the earlier measures by permitting gaps within a clump, is defined, and several variants examined. Experiments are carried out that indicate the relation between the new measure and one of the earlier measures, as well as the ability of the two types of measure to predict compression efficiency
  10. Bookstein, A.; Kulyukin, V.; Raita, T.; Nicholson, J.: Adapting measures of clumping strength to assess term-term similarity (2003) 0.01
    0.007916525 = product of:
      0.01583305 = sum of:
        0.01583305 = product of:
          0.0316661 = sum of:
            0.0316661 = weight(_text_:retrieval in 1609) [ClassicSimilarity], result of:
              0.0316661 = score(doc=1609,freq=2.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.20052543 = fieldWeight in 1609, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1609)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Automated information retrieval relies heavily an statistical regularities that emerge as terms are deposited to produce text. This paper examines statistical patterns expected of a pair of terms that are semantically related to each other. Guided by a conceptualization of the text generation process, we derive measures of how tightly two terms are semantically associated. Our main objective is to probe whether such measures yield reasonable results. Specifically, we examine how the tendency of a content bearing term to clump, as quantified by previously developed measures of term clumping, is influenced by the presence of other terms. This approach allows us to present a toolkit from which a range of measures can be constructed. As an illustration, one of several suggested measures is evaluated an a large text corpus built from an on-line encyclopedia.