Decoder integration and expected BLEU training for recurrent neural network language models

30Citations
Citations of this article
130Readers
Mendeley users who have this article in their library.

Abstract

Neural network language models are often trained by optimizing likelihood, but we would prefer to optimize for a task specific metric, such as BLEU in machine translation. We show how a recurrent neural network language model can be optimized towards an expected BLEU loss instead of the usual cross-entropy criterion. Furthermore, we tackle the issue of directly integrating a recurrent network into firstpass decoding under an efficient approximation. Our best results improve a phrasebased statistical machine translation system trained onWMT2012 French-English data by up to 2.0 BLEU, and the expected BLEU objective improves over a crossentropy trained model by up to 0.6 BLEU in a single reference setup. © 2014 Association for Computational Linguistics.

Cite

CITATION STYLE

APA

Auli, M., & Gao, J. (2014). Decoder integration and expected BLEU training for recurrent neural network language models. In 52nd Annual Meeting of the Association for Computational Linguistics, ACL 2014 - Proceedings of the Conference (Vol. 2, pp. 136–142). Association for Computational Linguistics (ACL). https://doi.org/10.3115/v1/p14-2023

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free