Specializing Multi-domain NMT via Penalizing Low Mutual Information

1Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.

Abstract

Multi-domain Neural Machine Translation (NMT) trains a single model with multiple domains. It is appealing because of its efficacy in handling multiple domains within one model. An ideal multi-domain NMT should learn distinctive domain characteristics simultaneously, however, grasping the domain peculiarity is a non-trivial task. In this paper, we investigate domain-specific information through the lens of mutual information (MI) and propose a new objective that penalizes low MI to become higher. Our method achieved the state-of-the-art performance among the current competitive multi-domain NMT models. Also, we empirically show our objective promotes low MI to be higher resulting in domain-specialized multi-domain NMT.

Cite

CITATION STYLE

APA

Lee, J., Kim, H., Cho, H., Choi, E., & Park, C. (2022). Specializing Multi-domain NMT via Penalizing Low Mutual Information. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 10015–10026). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.680

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free