Learn and Consolidate: Continual Adaptation for Zero-Shot and Multilingual Neural Machine Translation

0Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.

Abstract

Although existing multilingual neural machine translation (MNMT) models have demonstrated remarkable performance to handle multiple translation directions in a single model and achieved zero-shot translation between language pairs unseen in training, they still suffer from relatively poor translation qualities for some language pairs. A practical scenario is that how to continually update MNMT models for both supervised and zero-shot translations when limited new data arrives. To this end, we propose a two-stage approach that encourages original models to acquire language-agnostic multilingual representations from new data, and preserves the model architecture without introducing parameters. Experimental results and further analysis demonstrate that our method can efficiently improve performance of existing MNMT models in translation directions where they are initially weak, and mitigates the degeneration in the original well-performing translation directions, offering flexibility in the real-world scenario.

Cite

CITATION STYLE

APA

Huang, K., Li, P., Liu, J., Sun, M., & Liu, Y. (2023). Learn and Consolidate: Continual Adaptation for Zero-Shot and Multilingual Neural Machine Translation. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 13938–13951). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-main.860

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free