Search (10 results, page 1 of 1)

  • × author_ss:"Kantor, P.B."
  1. Elovici, Y.; Shapira, Y.B.; Kantor, P.B.: ¬A decision theoretic approach to combining information filters : an analytical and empirical evaluation. (2006) 0.02
    0.015250247 = product of:
      0.030500494 = sum of:
        0.00823978 = product of:
          0.03295912 = sum of:
            0.03295912 = weight(_text_:based in 5267) [ClassicSimilarity], result of:
              0.03295912 = score(doc=5267,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.23302436 = fieldWeight in 5267, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5267)
          0.25 = coord(1/4)
        0.022260714 = product of:
          0.04452143 = sum of:
            0.04452143 = weight(_text_:22 in 5267) [ClassicSimilarity], result of:
              0.04452143 = score(doc=5267,freq=2.0), product of:
                0.16438834 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04694356 = queryNorm
                0.2708308 = fieldWeight in 5267, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5267)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    The outputs of several information filtering (IF) systems can be combined to improve filtering performance. In this article the authors propose and explore a framework based on the so-called information structure (IS) model, which is frequently used in Information Economics, for combining the output of multiple IF systems according to each user's preferences (profile). The combination seeks to maximize the expected payoff to that user. The authors show analytically that the proposed framework increases users expected payoff from the combined filtering output for any user preferences. An experiment using the TREC-6 test collection confirms the theoretical findings.
    Date
    22. 7.2006 15:05:39
  2. Ng, K.B.; Kantor, P.B.; Strzalkowski, T.; Wacholder, N.; Tang, R.; Bai, B.; Rittman,; Song, P.; Sun, Y.: Automated judgment of document qualities (2006) 0.01
    0.011857167 = product of:
      0.047428668 = sum of:
        0.047428668 = product of:
          0.094857335 = sum of:
            0.094857335 = weight(_text_:assessment in 182) [ClassicSimilarity], result of:
              0.094857335 = score(doc=182,freq=2.0), product of:
                0.25917634 = queryWeight, product of:
                  5.52102 = idf(docFreq=480, maxDocs=44218)
                  0.04694356 = queryNorm
                0.36599535 = fieldWeight in 182, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.52102 = idf(docFreq=480, maxDocs=44218)
                  0.046875 = fieldNorm(doc=182)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    The authors report on a series of experiments to automate the assessment of document qualities such as depth and objectivity. The primary purpose is to develop a quality-sensitive functionality, orthogonal to relevance, to select documents for an interactive question-answering system. The study consisted of two stages. In the classifier construction stage, nine document qualities deemed important by information professionals were identified and classifiers were developed to predict their values. In the confirmative evaluation stage, the performance of the developed methods was checked using a different document collection. The quality prediction methods worked well in the second stage. The results strongly suggest that the best way to predict document qualities automatically is to construct classifiers on a person-by-person basis.
  3. Kantor, P.B.: Mathematical models in information science (2002) 0.01
    0.011130357 = product of:
      0.04452143 = sum of:
        0.04452143 = product of:
          0.08904286 = sum of:
            0.08904286 = weight(_text_:22 in 4112) [ClassicSimilarity], result of:
              0.08904286 = score(doc=4112,freq=2.0), product of:
                0.16438834 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04694356 = queryNorm
                0.5416616 = fieldWeight in 4112, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4112)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Source
    Bulletin of the American Society for Information Science. 28(2002) no.6, S.22-24
  4. Ng, K.B.; Loewenstern, D.; Basu, C.; Hirsh, H.; Kantor, P.B.: Data fusion of machine-learning methods for the TREC5 routing tak (and other work) (1997) 0.01
    0.007950256 = product of:
      0.031801023 = sum of:
        0.031801023 = product of:
          0.063602045 = sum of:
            0.063602045 = weight(_text_:22 in 3107) [ClassicSimilarity], result of:
              0.063602045 = score(doc=3107,freq=2.0), product of:
                0.16438834 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04694356 = queryNorm
                0.38690117 = fieldWeight in 3107, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3107)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    27. 2.1999 20:59:22
  5. Kantor, P.B.: Information retrieval techniques (1994) 0.00
    0.0024970302 = product of:
      0.009988121 = sum of:
        0.009988121 = product of:
          0.039952483 = sum of:
            0.039952483 = weight(_text_:based in 1056) [ClassicSimilarity], result of:
              0.039952483 = score(doc=1056,freq=4.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.28246817 = fieldWeight in 1056, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1056)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    State of the art review of information retrieval techniques viewed in terms of the growing effort to implement concept based retrieval in content based algorithms. Identifies trends in the automation of indexing, retrieval, and the interaction between systems and users. Identifies 3 central issues: ways in which systems describe documents for purposes of information retrieval; ways in which systems compute the degree of match between a given document and the current state of the query; amd what the systems do with the information that they obtain from the users. Looks at information retrieval techniques in terms of: location, navigation; indexing; documents; queries; structures; concepts; matching documents to queries; restoring query structure; algorithms and content versus concepts; formulation of concepts in terms of contents; formulation of concepts with the assistance of the users; complex system codes versus underlying principles; and system evaluation
  6. Menkov, V.; Ginsparg, P.; Kantor, P.B.: Recommendations and privacy in the arXiv system : a simulation experiment using historical data (2020) 0.00
    0.0024970302 = product of:
      0.009988121 = sum of:
        0.009988121 = product of:
          0.039952483 = sum of:
            0.039952483 = weight(_text_:based in 5671) [ClassicSimilarity], result of:
              0.039952483 = score(doc=5671,freq=4.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.28246817 = fieldWeight in 5671, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5671)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    Recommender systems may accelerate knowledge discovery in many fields. However, their users may be competitors guarding their ideas before publication or for other reasons. We describe a simulation experiment to assess user privacy against targeted attacks, modeling recommendations based on co-access data. The analysis uses an unusually long (14?years) set of anonymized historical data on user-item accesses. We introduce the notions of "visibility" and "discoverability." We find, based on historical data, that the majority of the actions of arXiv users would be potentially "visible" under targeted attack. However, "discoverability," which incorporates the difficulty of actually seeing a "visible" effect, is very much lower for nearly all users. We consider the effect of changes to the settings of the recommender algorithm on the visibility and discoverability of user actions and propose mitigation strategies that reduce both measures of risk.
  7. Saracevic, T.; Kantor, P.B.: Studying the value of library and information services : Part I: Establishing a theoretical framework (1997) 0.00
    0.0017656671 = product of:
      0.0070626684 = sum of:
        0.0070626684 = product of:
          0.028250674 = sum of:
            0.028250674 = weight(_text_:based in 352) [ClassicSimilarity], result of:
              0.028250674 = score(doc=352,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.19973516 = fieldWeight in 352, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.046875 = fieldNorm(doc=352)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Footnote
    1st part of a study to develop a taxonomy of value-in-use of library and information services based on users assessments and to propose methods and instruments for similar studies of library and information services in general
  8. Saracevic, T.; Kantor, P.B.: Studying the value of library and information services : Part II: Methodology and taxonomy (1997) 0.00
    0.0017656671 = product of:
      0.0070626684 = sum of:
        0.0070626684 = product of:
          0.028250674 = sum of:
            0.028250674 = weight(_text_:based in 353) [ClassicSimilarity], result of:
              0.028250674 = score(doc=353,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.19973516 = fieldWeight in 353, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.046875 = fieldNorm(doc=353)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Footnote
    2nd part of a study to develop a taxonomy of value-in-use of library and information services based on users assessments and to propose methods and instruments for similar studies of library and information services in general
  9. Shim, W.; Kantor, P.B.: Evaluation of digital libraries : a DEA approach (1999) 0.00
    0.0014713892 = product of:
      0.005885557 = sum of:
        0.005885557 = product of:
          0.023542227 = sum of:
            0.023542227 = weight(_text_:based in 6701) [ClassicSimilarity], result of:
              0.023542227 = score(doc=6701,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.16644597 = fieldWeight in 6701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=6701)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    As libraries evolve from paper based to digitized collection, traditional measurement activities must change. To demonstrate the growth in library value during this transition period, libraries must be able to describe how library inputs are transformed into the services libraries render. We apply a complex tool, data envelopment analysis (DEA), to evaluate the relative efficiency of major academic research libraries that are members of the Association of Research Libraries (ARL). An efficient library is defined as the one which produces same output with less input or, for a given input, produces more output. We report the results of a two-year base line study using traditional measures taken from 1995-1996 ARL statistics. We observe the patterns of efficiency scores of both individual libraries and libraries in peer groups (private vs. public). In particular we study the consistency over the years of specific DEA measures. This consistency provides justification for extending DEA as libraries undergo revolutionary digital transformation. The results are also corroborated using standard statistical measures. DEA application in the new digital library environment is discussed
  10. Sun, Y.; Kantor, P.B.: Cross-evaluation : a new model for information system evaluation (2006) 0.00
    0.0014713892 = product of:
      0.005885557 = sum of:
        0.005885557 = product of:
          0.023542227 = sum of:
            0.023542227 = weight(_text_:based in 5048) [ClassicSimilarity], result of:
              0.023542227 = score(doc=5048,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.16644597 = fieldWeight in 5048, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5048)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    In this article, we introduce a new information system evaluation method and report on its application to a collaborative information seeking system, AntWorld. The key innovation of the new method is to use precisely the same group of users who work with the system as judges, a system we call Cross-Evaluation. In the new method, we also propose to assess the system at the level of task completion. The obvious potential limitation of this method is that individuals may be inclined to think more highly of the materials that they themselves have found and are almost certain to think more highly of their own work product than they do of the products built by others. The keys to neutralizing this problem are careful design and a corresponding analytical model based on analysis of variance. We model the several measures of task completion with a linear model of five effects, describing the users who interact with the system, the system used to finish the task, the task itself, the behavior of individuals as judges, and the selfjudgment bias. Our analytical method successfully isolates the effect of each variable. This approach provides a successful model to make concrete the "threerealities" paradigm, which calls for "real tasks," "real users," and "real systems."