Multi-domain Neural Machine Translation (NMT) trains a single model with multiple domains. It is appealing because of its efficacy in handling multiple domains within one model. An ideal multi-domain NMT should learn distinctive domain characteristics simultaneously, however, grasping the domain peculiarity is a non-trivial task. In this paper, we investigate domain-specific information through the lens of mutual information (MI) and propose a new objective that penalizes low MI to become higher. Our method achieved the state-of-the-art performance among the current competitive multi-domain NMT models. Also, we empirically show our objective promotes low MI to be higher resulting in domain-specialized multi-domain NMT.
CITATION STYLE
Lee, J., Kim, H., Cho, H., Choi, E., & Park, C. (2022). Specializing Multi-domain NMT via Penalizing Low Mutual Information. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 10015–10026). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.680
Mendeley helps you to discover research relevant for your work.