Neural network language models are often trained by optimizing likelihood, but we would prefer to optimize for a task specific metric, such as BLEU in machine translation. We show how a recurrent neural network language model can be optimized towards an expected BLEU loss instead of the usual cross-entropy criterion. Furthermore, we tackle the issue of directly integrating a recurrent network into firstpass decoding under an efficient approximation. Our best results improve a phrasebased statistical machine translation system trained onWMT2012 French-English data by up to 2.0 BLEU, and the expected BLEU objective improves over a crossentropy trained model by up to 0.6 BLEU in a single reference setup. © 2014 Association for Computational Linguistics.
CITATION STYLE
Auli, M., & Gao, J. (2014). Decoder integration and expected BLEU training for recurrent neural network language models. In 52nd Annual Meeting of the Association for Computational Linguistics, ACL 2014 - Proceedings of the Conference (Vol. 2, pp. 136–142). Association for Computational Linguistics (ACL). https://doi.org/10.3115/v1/p14-2023
Mendeley helps you to discover research relevant for your work.