On the Complementarity between Pre-Training and Back-Translation for Neural Machine Translation

22Citations
Citations of this article
63Readers
Mendeley users who have this article in their library.

Abstract

Pre-training (PT) and back-translation (BT) are two simple and powerful methods to utilize monolingual data for improving the model performance of neural machine translation (NMT). This paper takes the first step to investigate the complementarity between PT and BT. We introduce two probing tasks for PT and BT respectively and find that PT mainly contributes to the encoder module while BT brings more benefits to the decoder. Experimental results show that PT and BT are nicely complementary to each other, establishing state-ofthe-art performances on the WMT16 EnglishRomanian and English-Russian benchmarks. Through extensive analyses on sentence originality and word frequency, we also demonstrate that combining Tagged BT with PT is more helpful to their complementarity, leading to better translation quality. Source code is freely available at https://github.com/ SunbowLiu/PTvsBT.

Cite

CITATION STYLE

APA

Liu, X., Wang, L., Wong, D. F., Ding, L., Chao, L. S., Shi, S., & Tu, Z. (2021). On the Complementarity between Pre-Training and Back-Translation for Neural Machine Translation. In Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 (pp. 2900–2907). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-emnlp.247

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free