A simple and effective approach to coverage-aware neural machine translation

15Citations
Citations of this article
111Readers
Mendeley users who have this article in their library.

Abstract

We offer a simple and effective method to seek a better balance between model confidence and length preference for Neural Machine Translation (NMT). Unlike the popular length normalization and coverage models, our model does not require training nor reranking the limited n-best outputs. Moreover, it is robust to large beam sizes, which is not well studied in previous work. On the Chinese-English and English-German translation tasks, our approach yields +0.4 ∼ 1.5 BLEU improvements over the state-of-the-art baselines.

Cite

CITATION STYLE

APA

Li, Y., Xiao, T., Li, Y., Wang, Q., Xu, C., & Lu, X. (2018). A simple and effective approach to coverage-aware neural machine translation. In ACL 2018 - 56th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 2, pp. 292–297). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p18-2047

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free