Search (2 results, page 1 of 1)
- Did you mean:
- bliss%3a%2281.68 computereinsatz in unterricht und ausbildung%22 2
- bliss%3a%2181.68 computereinsatz in unterricht und ausbildung%22 2
- bliss%3a%2281.68 computereinsatz in unterrichts und ausbildung%22 2
- bliss%3a%2281.68 computereinsatzes in unterricht und ausbildung%22 2
- bross%3a%2281.68 computereinsatz in unterricht und ausbildung%22 2
-
Toutanova, K.; Manning, C.D.: Enriching the knowledge sources used in a maximum entropy Part-of-Speech Tagger (2000)
0.00
0.0018033426 = product of: 0.010820055 = sum of: 0.010820055 = weight(_text_:in in 1060) [ClassicSimilarity], result of: 0.010820055 = score(doc=1060,freq=6.0), product of: 0.059380736 = queryWeight, product of: 1.3602545 = idf(docFreq=30841, maxDocs=44218) 0.043654136 = queryNorm 0.1822149 = fieldWeight in 1060, product of: 2.4494898 = tf(freq=6.0), with freq of: 6.0 = termFreq=6.0 1.3602545 = idf(docFreq=30841, maxDocs=44218) 0.0546875 = fieldNorm(doc=1060) 0.16666667 = coord(1/6)
- Abstract
- This paper presents results for a maximumentropy-based part of speech tagger, which achieves superior performance principally by enriching the information sources used for tagging. In particular, we get improved results by incorporating these features: (i) more extensive treatment of capitalization for unknown words; (ii) features for the disambiguation of the tense forms of verbs; (iii) features for disambiguating particles from prepositions and adverbs. The best resulting accuracy for the tagger on the Penn Treebank is 96.86% overall, and 86.91% on previously unseen words.
- Source
- Proceedings of the Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora (EMNLP/VLC-2000)
-
Toutanova, K.; Klein, D.; Manning, C.D.; Singer, Y.: Feature-rich Part-of-Speech Tagging with a cyclic dependency network (2003)
0.00
0.0010411602 = product of: 0.006246961 = sum of: 0.006246961 = weight(_text_:in in 1059) [ClassicSimilarity], result of: 0.006246961 = score(doc=1059,freq=2.0), product of: 0.059380736 = queryWeight, product of: 1.3602545 = idf(docFreq=30841, maxDocs=44218) 0.043654136 = queryNorm 0.10520181 = fieldWeight in 1059, product of: 1.4142135 = tf(freq=2.0), with freq of: 2.0 = termFreq=2.0 1.3602545 = idf(docFreq=30841, maxDocs=44218) 0.0546875 = fieldNorm(doc=1059) 0.16666667 = coord(1/6)
- Abstract
- We present a new part-of-speech tagger that demonstrates the following ideas: (i) explicit use of both preceding and following tag contexts via a dependency network representation, (ii) broad use of lexical features, including jointly conditioning on multiple consecutive words, (iii) effective use of priors in conditional loglinear models, and (iv) fine-grained modeling of unknown word features. Using these ideas together, the resulting tagger gives a 97.24%accuracy on the Penn TreebankWSJ, an error reduction of 4.4% on the best previous single automatically learned tagging result.
Authors
- Manning, C.D. 2
- Klein, D. 1
- Singer, Y. 1