Freezing Subnetworks to Analyze Domain Adaptation in Neural Machine Translation

33Citations
Citations of this article
127Readers
Mendeley users who have this article in their library.

Abstract

To better understand the effectiveness of continued training, we analyze the major components of a neural machine translation system (the encoder, decoder, and each embedding space) and consider each component’s contribution to, and capacity for, domain adaptation. We find that freezing any single component during continued training has minimal impact on performance, and that performance is surprisingly good when a single component is adapted while holding the rest of the model fixed. We also find that continued training does not move the model very far from the out-of-domain model, compared to a sensitivity analysis metric, suggesting that the out-of-domain model can provide a good generic initialization for the new domain.

Cite

CITATION STYLE

APA

Thompson, B., Khayrallah, H., Anastasopoulos, A., McCarthy, A. D., Duh, K., Marvin, R., … Koehn, P. (2018). Freezing Subnetworks to Analyze Domain Adaptation in Neural Machine Translation. In WMT 2018 - 3rd Conference on Machine Translation, Proceedings of the Conference (Vol. 1, pp. 124–132). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w18-6313

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free