Abstract
While back-translation is simple and effective in exploiting abundant monolingual corpora to improve low-resource neural machine translation (NMT), the synthetic bilingual corpora generated by NMT models trained on limited authentic bilingual data are inevitably noisy. In this work, we propose to quantify the confidence of NMT model predictions based on model uncertainty. With word- and sentence-level confidence measures based on uncertainty, it is possible for back-translation to better cope with noise in synthetic bilingual corpora. Experiments on Chinese-English and English-German translation tasks show that uncertainty-based confidence estimation significantly improves the performance of backtranslation. 1.
Cite
CITATION STYLE
Wang, S., Liu, Y., Wang, C., Luan, H., & Sun, M. (2019). Improving back-translation with uncertainty-based confidence estimation. In EMNLP-IJCNLP 2019 - 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint Conference on Natural Language Processing, Proceedings of the Conference (pp. 791–802). Association for Computational Linguistics. https://doi.org/10.18653/v1/D19-1073
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.