Abstract
We propose a new approach to language modeling which utilizes discriminative learning methods. Our approach is an iterative one: starting with an initial language model, in each iteration we generate 'false' sentences from the current model, and then train a classifier to discriminate between them and sentences from the training corpus. To the extent that this succeeds, the classifier is incorporated into the model by lowering the probability of sentences classified as false, and the process is repeated. We demonstrate the effectiveness of this approach on a natural language corpus and show it provides an 11.4% improvement in perplexity over a modified kneser-ney smoothed trigram. © 2008 Association for Computational Linguistics.
Cite
CITATION STYLE
Sandbank, B. (2008). Refining generative language models using discriminative learning. In EMNLP 2008 - 2008 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference: A Meeting of SIGDAT, a Special Interest Group of the ACL (pp. 51–58). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1613715.1613723
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.