Abstract
This paper investigates several augmented mixture models that are competitive alternatives to standard Bayesian models and prove to be very suitable to word sense disambiguation and related classification tasks. We present a new classification correction technique that successfully addresses the problem of under-estimation of infrequent classes in the training data. We show that the mixture models are boosting-friendly and that both Adaboost and our original correction technique can improve the results of the raw model significantly, achieving state-of-the-art performance on several standard test sets in four languages. With substantially different output to Naïve Bayes and other statistical methods, the investigated models are also shown to be effective participants in classifier combination.
Cite
CITATION STYLE
Cucerzan, S., & Yarowsky, D. (2002). Augmented Mixture Models for Lexical Disambiguation. In Proceedings of the 2002 Conference on Empirical Methods in Natural Language Processing, EMNLP 2002 (pp. 33–40). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1118693.1118698
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.