Search (7 results, page 1 of 1)

  • × author_ss:"Losee, R.M."
  1. Losee, R.M.: Determining information retrieval and filtering performance without experimentation (1995) 0.03
    0.03485952 = product of:
      0.13943808 = sum of:
        0.13943808 = sum of:
          0.09428916 = weight(_text_:model in 3368) [ClassicSimilarity], result of:
            0.09428916 = score(doc=3368,freq=6.0), product of:
              0.1830527 = queryWeight, product of:
                3.845226 = idf(docFreq=2569, maxDocs=44218)
                0.047605187 = queryNorm
              0.51509297 = fieldWeight in 3368, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                3.845226 = idf(docFreq=2569, maxDocs=44218)
                0.0546875 = fieldNorm(doc=3368)
          0.045148917 = weight(_text_:22 in 3368) [ClassicSimilarity], result of:
            0.045148917 = score(doc=3368,freq=2.0), product of:
              0.16670525 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.047605187 = queryNorm
              0.2708308 = fieldWeight in 3368, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=3368)
      0.25 = coord(1/4)
    
    Abstract
    The performance of an information retrieval or text and media filtering system may be determined through analytic methods as well as by traditional simulation or experimental methods. These analytic methods can provide precise statements about expected performance. They can thus determine which of 2 similarly performing systems is superior. For both a single query terms and for a multiple query term retrieval model, a model for comparing the performance of different probabilistic retrieval methods is developed. This method may be used in computing the average search length for a query, given only knowledge of database parameter values. Describes predictive models for inverse document frequency, binary independence, and relevance feedback based retrieval and filtering. Simulation illustrate how the single term model performs and sample performance predictions are given for single term and multiple term problems
    Date
    22. 2.1996 13:14:10
  2. Losee, R.M.: Term dependence : truncating the Bahadur Lazarsfeld expansion (1994) 0.02
    0.016497165 = product of:
      0.06598866 = sum of:
        0.06598866 = product of:
          0.13197732 = sum of:
            0.13197732 = weight(_text_:model in 7390) [ClassicSimilarity], result of:
              0.13197732 = score(doc=7390,freq=4.0), product of:
                0.1830527 = queryWeight, product of:
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.047605187 = queryNorm
                0.72097987 = fieldWeight in 7390, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.09375 = fieldNorm(doc=7390)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Studies the performance of probabilistic information retrieval systems where differing statistical dependence assumptions are used when estimating the probabilities inherent in the retrieval model. Uses the Bahadur Lazarsfeld expansion model
  3. Losee, R.M.: ¬The effect of assigning a metadata or indexing term on document ordering (2013) 0.01
    0.01010241 = product of:
      0.04040964 = sum of:
        0.04040964 = product of:
          0.08081928 = sum of:
            0.08081928 = weight(_text_:model in 1100) [ClassicSimilarity], result of:
              0.08081928 = score(doc=1100,freq=6.0), product of:
                0.1830527 = queryWeight, product of:
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.047605187 = queryNorm
                0.44150823 = fieldWeight in 1100, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1100)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    The assignment of indexing terms and metadata to documents, data, and other information representations is considered useful, but the utility of including a single term is seldom discussed. The author discusses a simple model of document ordering and then shows how assigning index and metadata labels improves or decreases retrieval performance. The Indexing and Metadata Advantage (IMA) factor measures how indexing or assigning a metadata term helps (or hurts) ordering performance. Performance values and the associated IMA expressions are computed, consistent with several different assumptions. The economic value associated with various term assignment decisions is developed. The IMA term advantage model itself is empirically validated with computer software that shows that the analytic results obtained agree completely with the actual performance gains and losses found when ordering all sets of 14 or fewer documents. When the formulas in the software are changed to differ from this model, the predictions of the actual performance are erroneous.
  4. Losee, R.M.: Evaluating retrieval performance given database and query characteristics : analytic determination of performance surfaces (1996) 0.01
    0.0082485825 = product of:
      0.03299433 = sum of:
        0.03299433 = product of:
          0.06598866 = sum of:
            0.06598866 = weight(_text_:model in 4162) [ClassicSimilarity], result of:
              0.06598866 = score(doc=4162,freq=4.0), product of:
                0.1830527 = queryWeight, product of:
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.047605187 = queryNorm
                0.36048993 = fieldWeight in 4162, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4162)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    An analytic method of information retrieval and filtering evaluation can quantitatively predict the expected number of documents examined in retrieving a relevant document. It also allows researchers and practioners to qualitatively understand how varying different estimates of query parameter values affects retrieval performance. The incoorporation of relevance feedback to increase our knowledge about the parameters of relevant documents and the robustness of parameter estimates is modeled. Single term and two term independence models, as well as a complete term dependence model, are developed. An economic model of retrieval performance may be used to study the effects of database size and to provide analytic answers to questions comparing retrieval from small and large databases, as well as questions about the number of terms in a query. Results are presented as a performance surface, a three dimensional graph showing the effects of two independent variables on performance.
  5. Losee, R.M.: Browsing document collections : automatically organizing digital libraries and hypermedia using the Gray code (1997) 0.01
    0.0058326283 = product of:
      0.023330513 = sum of:
        0.023330513 = product of:
          0.046661027 = sum of:
            0.046661027 = weight(_text_:model in 146) [ClassicSimilarity], result of:
              0.046661027 = score(doc=146,freq=2.0), product of:
                0.1830527 = queryWeight, product of:
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.047605187 = queryNorm
                0.25490487 = fieldWeight in 146, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.046875 = fieldNorm(doc=146)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Relevance and economic feedback may be used to produce an ordering of documents that supports browsing in hypermedia and digital libraries. Document classification based on the Gray code provides paths through the entire collection, each path traversing each node in the set of documents exactly once. Examines systems organizing document based on weighted and unweighted Gray codes. Relevance feedback is used to conceptually organize the collection for an individual to browse, based on that individual's interests and information needs, as reflected by their relevance judgements and user supplied economic preferences. Applies Bayesian learning theory to estimating the characteristics of documents of interest to the user and supplying an analytic model of browsing performance, based on minimising the Expected Browsing Distance. Economic feedback may be used to change the ordering of documents to benefit the user. Using these techniques, a hypermedia or digital library may order any and all available documents, not just those examined, based on the information provided by the searcher or people with similar interests
  6. Losee, R.M.: Term dependence : a basis for Luhn and Zipf models (2001) 0.01
    0.0058326283 = product of:
      0.023330513 = sum of:
        0.023330513 = product of:
          0.046661027 = sum of:
            0.046661027 = weight(_text_:model in 6976) [ClassicSimilarity], result of:
              0.046661027 = score(doc=6976,freq=2.0), product of:
                0.1830527 = queryWeight, product of:
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.047605187 = queryNorm
                0.25490487 = fieldWeight in 6976, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.046875 = fieldNorm(doc=6976)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    There are regularities in the statistical information provided by natural language terms about neighboring terms. We find that when phrase rank increases, moving from common to less common phrases, the value of the expected mutual information measure (EMIM) between the terms regularly decreases. Luhn's model suggests that midrange terms are the best index terms and relevance discriminators. We suggest reasons for this principle based on the empirical relationships shown here between the rank of terms within phrases and the average mutual information between terms, which we refer to as the Inverse Representation- EMIM principle. We also suggest an Inverse EMIM term weight for indexing or retrieval applications that is consistent with Luhn's distribution. An information theoretic interpretation of Zipf's Law is provided. Using the regularity noted here, we suggest that Zipf's Law is a consequence of the statistical dependencies that exist between terms, described here using information theoretic concepts.
  7. Losee, R.M.: Improving collection browsing : small world networking and Gray code ordering (2017) 0.00
    0.0048605236 = product of:
      0.019442094 = sum of:
        0.019442094 = product of:
          0.03888419 = sum of:
            0.03888419 = weight(_text_:model in 5148) [ClassicSimilarity], result of:
              0.03888419 = score(doc=5148,freq=2.0), product of:
                0.1830527 = queryWeight, product of:
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.047605187 = queryNorm
                0.21242073 = fieldWeight in 5148, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5148)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Documents in digital and paper libraries may be arranged, based on their topics, in order to facilitate browsing. It may seem intuitively obvious that ordering documents by their subject should improve browsing performance; the results presented in this article suggest that ordering library materials by their Gray code values and through using links consistent with the small world model of document relationships is consistent with improving browsing performance. Below, library circulation data, including ordering with Library of Congress Classification numbers and Library of Congress Subject Headings, are used to provide information useful in generating user-centered document arrangements, as well as user-independent arrangements. Documents may be linearly arranged so they can be placed in a line by topic, such as on a library shelf, or in a list on a computer display. Crossover links, jumps between a document and another document to which it is not adjacent, can be used in library databases to allow additional paths that one might take when browsing. The improvement that is obtained with different combinations of document orderings and different crossovers is examined and applications suggested.