Improving back-translation with uncertainty-based confidence estimation

60Citations
Citations of this article
165Readers
Mendeley users who have this article in their library.

Abstract

While back-translation is simple and effective in exploiting abundant monolingual corpora to improve low-resource neural machine translation (NMT), the synthetic bilingual corpora generated by NMT models trained on limited authentic bilingual data are inevitably noisy. In this work, we propose to quantify the confidence of NMT model predictions based on model uncertainty. With word- and sentence-level confidence measures based on uncertainty, it is possible for back-translation to better cope with noise in synthetic bilingual corpora. Experiments on Chinese-English and English-German translation tasks show that uncertainty-based confidence estimation significantly improves the performance of backtranslation. 1.

Cite

CITATION STYLE

APA

Wang, S., Liu, Y., Wang, C., Luan, H., & Sun, M. (2019). Improving back-translation with uncertainty-based confidence estimation. In EMNLP-IJCNLP 2019 - 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint Conference on Natural Language Processing, Proceedings of the Conference (pp. 791–802). Association for Computational Linguistics. https://doi.org/10.18653/v1/D19-1073

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free