Regularized Training Objective for Continued Training for Domain Adaptation in Neural Machine Translation

48Citations
Citations of this article
87Readers
Mendeley users who have this article in their library.

Abstract

Supervised domain adaptation-where a large generic corpus and a smaller in-domain corpus are both available for training-is a challenge for neural machine translation (NMT). Standard practice is to train a generic model and use it to initialize a second model, then continue training the second model on in-domain data to produce an in-domain model. We add an auxiliary term to the training objective during continued training that minimizes the cross entropy between the in-domain model's output word distribution and that of the out-of-domain model to prevent the model's output from differing too much from the original out-of-domain model. We perform experiments on EMEA (descriptions of medicines) and TED (rehearsed presentations), initialized from a general domain (WMT) model. Our method shows improvements over standard continued training by up to 1.5 BLEU.

Cite

CITATION STYLE

APA

Khayrallah, H., Thompson, B., Duh, K., & Koehn, P. (2018). Regularized Training Objective for Continued Training for Domain Adaptation in Neural Machine Translation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 36–41). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w18-2705

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free