Search (2 results, page 1 of 1)

  • × author_ss:"Bai, B."
  1. Ng, K.B.; Kantor, P.B.; Strzalkowski, T.; Wacholder, N.; Tang, R.; Bai, B.; Rittman,; Song, P.; Sun, Y.: Automated judgment of document qualities (2006) 0.00
    0.0037164674 = product of:
      0.01486587 = sum of:
        0.01486587 = weight(_text_:information in 182) [ClassicSimilarity], result of:
          0.01486587 = score(doc=182,freq=4.0), product of:
            0.09032827 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051455088 = queryNorm
            0.16457605 = fieldWeight in 182, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=182)
      0.25 = coord(1/4)
    
    Abstract
    The authors report on a series of experiments to automate the assessment of document qualities such as depth and objectivity. The primary purpose is to develop a quality-sensitive functionality, orthogonal to relevance, to select documents for an interactive question-answering system. The study consisted of two stages. In the classifier construction stage, nine document qualities deemed important by information professionals were identified and classifiers were developed to predict their values. In the confirmative evaluation stage, the performance of the developed methods was checked using a different document collection. The quality prediction methods worked well in the second stage. The results strongly suggest that the best way to predict document qualities automatically is to construct classifiers on a person-by-person basis.
    Source
    Journal of the American Society for Information Science and Technology. 57(2006) no.9, S.1155-1164
  2. Wacholder, N.; Kelly, D.; Kantor, P.; Rittman, R.; Sun, Y.; Bai, B.; Small, S.; Yamrom, B.; Strzalkowski, T.: ¬A model for quantitative evaluation of an end-to-end question-answering system (2007) 0.00
    0.0037164674 = product of:
      0.01486587 = sum of:
        0.01486587 = weight(_text_:information in 435) [ClassicSimilarity], result of:
          0.01486587 = score(doc=435,freq=4.0), product of:
            0.09032827 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051455088 = queryNorm
            0.16457605 = fieldWeight in 435, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=435)
      0.25 = coord(1/4)
    
    Abstract
    We describe a procedure for quantitative evaluation of interactive question-answering systems and illustrate it with application to the High-Quality Interactive QuestionAnswering (HITIQA) system. Our objectives were (a) to design a method to realistically and reliably assess interactive question-answering systems by comparing the quality of reports produced using different systems, (b) to conduct a pilot test of this method, and (c) to perform a formative evaluation of the HITIQA system. Far more important than the specific information gathered from this pilot evaluation is the development of (a) a protocol for evaluating an emerging technology, (b) reusable assessment instruments, and (c) the knowledge gained in conducting the evaluation. We conclude that this method, which uses a surprisingly small number of subjects and does not rely on predetermined relevance judgments, measures the impact of system change on work produced by users. Therefore this method can be used to compare the product of interactive systems that use different underlying technologies.
    Source
    Journal of the American Society for Information Science and Technology. 58(2007) no.8, S.1082-1099