Zajic, D.; Dorr, B.J.; Lin, J.; Schwartz, R.: Multi-candidate reduction : sentence compression as a tool for document summarization tasks (2007)
0.01
0.008779775 = product of:
0.03219251 = sum of:
0.0072322977 = weight(_text_:a in 944) [ClassicSimilarity], result of:
0.0072322977 = score(doc=944,freq=14.0), product of:
0.030653298 = queryWeight, product of:
1.153047 = idf(docFreq=37942, maxDocs=44218)
0.026584605 = queryNorm
0.23593865 = fieldWeight in 944, product of:
3.7416575 = tf(freq=14.0), with freq of:
14.0 = termFreq=14.0
1.153047 = idf(docFreq=37942, maxDocs=44218)
0.0546875 = fieldNorm(doc=944)
0.022529786 = weight(_text_:r in 944) [ClassicSimilarity], result of:
0.022529786 = score(doc=944,freq=2.0), product of:
0.088001914 = queryWeight, product of:
3.3102584 = idf(docFreq=4387, maxDocs=44218)
0.026584605 = queryNorm
0.25601473 = fieldWeight in 944, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.3102584 = idf(docFreq=4387, maxDocs=44218)
0.0546875 = fieldNorm(doc=944)
0.0024304248 = weight(_text_:s in 944) [ClassicSimilarity], result of:
0.0024304248 = score(doc=944,freq=2.0), product of:
0.028903782 = queryWeight, product of:
1.0872376 = idf(docFreq=40523, maxDocs=44218)
0.026584605 = queryNorm
0.08408674 = fieldWeight in 944, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
1.0872376 = idf(docFreq=40523, maxDocs=44218)
0.0546875 = fieldNorm(doc=944)
0.27272728 = coord(3/11)
- Abstract
- This article examines the application of two single-document sentence compression techniques to the problem of multi-document summarization-a "parse-and-trim" approach and a statistical noisy-channel approach. We introduce the multi-candidate reduction (MCR) framework for multi-document summarization, in which many compressed candidates are generated for each source sentence. These candidates are then selected for inclusion in the final summary based on a combination of static and dynamic features. Evaluations demonstrate that sentence compression is a valuable component of a larger multi-document summarization framework.
- Source
- Information processing and management. 43(2007) no.6, S.1549-1570
- Type
- a
Hobson, S.P.; Dorr, B.J.; Monz, C.; Schwartz, R.: Task-based evaluation of text summarization using Relevance Prediction (2007)
0.01
0.008138846 = product of:
0.029842433 = sum of:
0.008447966 = weight(_text_:a in 938) [ClassicSimilarity], result of:
0.008447966 = score(doc=938,freq=26.0), product of:
0.030653298 = queryWeight, product of:
1.153047 = idf(docFreq=37942, maxDocs=44218)
0.026584605 = queryNorm
0.27559727 = fieldWeight in 938, product of:
5.0990195 = tf(freq=26.0), with freq of:
26.0 = termFreq=26.0
1.153047 = idf(docFreq=37942, maxDocs=44218)
0.046875 = fieldNorm(doc=938)
0.019311246 = weight(_text_:r in 938) [ClassicSimilarity], result of:
0.019311246 = score(doc=938,freq=2.0), product of:
0.088001914 = queryWeight, product of:
3.3102584 = idf(docFreq=4387, maxDocs=44218)
0.026584605 = queryNorm
0.2194412 = fieldWeight in 938, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.3102584 = idf(docFreq=4387, maxDocs=44218)
0.046875 = fieldNorm(doc=938)
0.0020832212 = weight(_text_:s in 938) [ClassicSimilarity], result of:
0.0020832212 = score(doc=938,freq=2.0), product of:
0.028903782 = queryWeight, product of:
1.0872376 = idf(docFreq=40523, maxDocs=44218)
0.026584605 = queryNorm
0.072074346 = fieldWeight in 938, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
1.0872376 = idf(docFreq=40523, maxDocs=44218)
0.046875 = fieldNorm(doc=938)
0.27272728 = coord(3/11)
- Abstract
- This article introduces a new task-based evaluation measure called Relevance Prediction that is a more intuitive measure of an individual's performance on a real-world task than interannotator agreement. Relevance Prediction parallels what a user does in the real world task of browsing a set of documents using standard search tools, i.e., the user judges relevance based on a short summary and then that same user - not an independent user - decides whether to open (and judge) the corresponding document. This measure is shown to be a more reliable measure of task performance than LDC Agreement, a current gold-standard based measure used in the summarization evaluation community. Our goal is to provide a stable framework within which developers of new automatic measures may make stronger statistical statements about the effectiveness of their measures in predicting summary usefulness. We demonstrate - as a proof-of-concept methodology for automatic metric developers - that a current automatic evaluation measure has a better correlation with Relevance Prediction than with LDC Agreement and that the significance level for detected differences is higher for the former than for the latter.
- Source
- Information processing and management. 43(2007) no.6, S.1482-1499
- Type
- a