Search (6 results, page 1 of 1)

  • × author_ss:"Kantor, P.B."
  • × language_ss:"e"
  • × type_ss:"a"
  1. Sun, Y.; Kantor, P.B.; Morse, E.L.: Using cross-evaluation to evaluate interactive QA systems (2011) 0.04
    0.036013078 = product of:
      0.14405231 = sum of:
        0.14405231 = weight(_text_:assess in 4744) [ClassicSimilarity], result of:
          0.14405231 = score(doc=4744,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.39077166 = fieldWeight in 4744, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.046875 = fieldNorm(doc=4744)
      0.25 = coord(1/4)
    
    Abstract
    In this article, we report on an experiment to assess the possibility of rigorous evaluation of interactive question-answering (QA) systems using the cross-evaluation method. This method takes into account the effects of tasks and context, and of the users of the systems. Statistical techniques are used to remove these effects, isolating the effect of the system itself. The results show that this approach yields meaningful measurements of the impact of systems on user task performance, using a surprisingly small number of subjects and without relying on predetermined judgments of the quality, or of the relevance of materials. We conclude that the method is indeed effective for comparing end-to-end QA systems, and for comparing interactive systems with high efficiency.
  2. Menkov, V.; Ginsparg, P.; Kantor, P.B.: Recommendations and privacy in the arXiv system : a simulation experiment using historical data (2020) 0.04
    0.036013078 = product of:
      0.14405231 = sum of:
        0.14405231 = weight(_text_:assess in 5671) [ClassicSimilarity], result of:
          0.14405231 = score(doc=5671,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.39077166 = fieldWeight in 5671, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.046875 = fieldNorm(doc=5671)
      0.25 = coord(1/4)
    
    Abstract
    Recommender systems may accelerate knowledge discovery in many fields. However, their users may be competitors guarding their ideas before publication or for other reasons. We describe a simulation experiment to assess user privacy against targeted attacks, modeling recommendations based on co-access data. The analysis uses an unusually long (14?years) set of anonymized historical data on user-item accesses. We introduce the notions of "visibility" and "discoverability." We find, based on historical data, that the majority of the actions of arXiv users would be potentially "visible" under targeted attack. However, "discoverability," which incorporates the difficulty of actually seeing a "visible" effect, is very much lower for nearly all users. We consider the effect of changes to the settings of the recommender algorithm on the visibility and discoverability of user actions and propose mitigation strategies that reduce both measures of risk.
  3. Sun, Y.; Kantor, P.B.: Cross-evaluation : a new model for information system evaluation (2006) 0.03
    0.030010901 = product of:
      0.120043606 = sum of:
        0.120043606 = weight(_text_:assess in 5048) [ClassicSimilarity], result of:
          0.120043606 = score(doc=5048,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.32564306 = fieldWeight in 5048, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5048)
      0.25 = coord(1/4)
    
    Abstract
    In this article, we introduce a new information system evaluation method and report on its application to a collaborative information seeking system, AntWorld. The key innovation of the new method is to use precisely the same group of users who work with the system as judges, a system we call Cross-Evaluation. In the new method, we also propose to assess the system at the level of task completion. The obvious potential limitation of this method is that individuals may be inclined to think more highly of the materials that they themselves have found and are almost certain to think more highly of their own work product than they do of the products built by others. The keys to neutralizing this problem are careful design and a corresponding analytical model based on analysis of variance. We model the several measures of task completion with a linear model of five effects, describing the users who interact with the system, the system used to finish the task, the task itself, the behavior of individuals as judges, and the selfjudgment bias. Our analytical method successfully isolates the effect of each variable. This approach provides a successful model to make concrete the "threerealities" paradigm, which calls for "real tasks," "real users," and "real systems."
  4. Kantor, P.B.: Mathematical models in information science (2002) 0.03
    0.029654698 = product of:
      0.118618794 = sum of:
        0.118618794 = weight(_text_:22 in 4112) [ClassicSimilarity], result of:
          0.118618794 = score(doc=4112,freq=2.0), product of:
            0.21899058 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.062536046 = queryNorm
            0.5416616 = fieldWeight in 4112, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.109375 = fieldNorm(doc=4112)
      0.25 = coord(1/4)
    
    Source
    Bulletin of the American Society for Information Science. 28(2002) no.6, S.22-24
  5. Ng, K.B.; Loewenstern, D.; Basu, C.; Hirsh, H.; Kantor, P.B.: Data fusion of machine-learning methods for the TREC5 routing tak (and other work) (1997) 0.02
    0.021181928 = product of:
      0.08472771 = sum of:
        0.08472771 = weight(_text_:22 in 3107) [ClassicSimilarity], result of:
          0.08472771 = score(doc=3107,freq=2.0), product of:
            0.21899058 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.062536046 = queryNorm
            0.38690117 = fieldWeight in 3107, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.078125 = fieldNorm(doc=3107)
      0.25 = coord(1/4)
    
    Date
    27. 2.1999 20:59:22
  6. Elovici, Y.; Shapira, Y.B.; Kantor, P.B.: ¬A decision theoretic approach to combining information filters : an analytical and empirical evaluation. (2006) 0.01
    0.014827349 = product of:
      0.059309397 = sum of:
        0.059309397 = weight(_text_:22 in 5267) [ClassicSimilarity], result of:
          0.059309397 = score(doc=5267,freq=2.0), product of:
            0.21899058 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.062536046 = queryNorm
            0.2708308 = fieldWeight in 5267, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5267)
      0.25 = coord(1/4)
    
    Date
    22. 7.2006 15:05:39