Bridging the gap between training and inference for neural machine translation

6Citations
Citations of this article
532Readers
Mendeley users who have this article in their library.

Abstract

Neural Machine Translation (NMT) generates target words sequentially in the way of predicting the next word conditioned on the context words. At training time, it predicts with the ground truth words as context while at inference it has to generate the entire sequence from scratch. This discrepancy of the fed context leads to error accumulation among the translation. Furthermore, word-level training requires strict matching between the generated sequence and the ground truth sequence which leads to overcorrection over different but reasonable translations. In this paper, we address these issues by sampling context words not only from the ground truth sequence but also from the predicted sequence during training1. Experimental results on NIST Chinese?English and WMT2014 English?German translation tasks demonstrate that our method can achieve significant improvements on multiple data sets compared to strong baselines.

Cite

CITATION STYLE

APA

Zhang, W., Feng, Y., & Liu, Q. (2020). Bridging the gap between training and inference for neural machine translation. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2021-January, pp. 4790–4794). International Joint Conferences on Artificial Intelligence. https://doi.org/10.18653/v1/p19-1426

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free