CLLE: A Benchmark for Continual Language Learning Evaluation in Multilingual Machine Translation

5Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Continual Language Learning (CLL) in multilingual translation is inevitable when new languages are required to be translated. Due to the lack of unified and generalized benchmarks, the evaluation of existing methods is greatly influenced by experimental design which usually has a big gap from the industrial demands. In this work, we propose the first Continual Language Learning Evaluation benchmark CLLE in multilingual translation. CLLE consists of a Chinese-centric corpus - CN-25 and two CLL tasks - the close-distance language continual learning task and the language family continual learning task designed for real and disparate demands. Different from existing translation benchmarks, CLLE considers several restrictions for CLL, including domain distribution alignment, content overlap, language diversity, and the balance of corpus. Furthermore, we propose a novel framework COMETA based on Constrained Optimization and META-learning to alleviate catastrophic forgetting and dependency on historical training data by using a meta-model to retain the important parameters for old languages. Our experiments prove that CLLE is a challenging CLL benchmark and that our proposed method is effective when compared with other strong baselines. Due to the construction of corpus, the task designing and the evaluation method are independent of the central language, we also construct and release the English-centric corpus EN-25 to facilitate academic research.

Cite

CITATION STYLE

APA

Zhang, H., Zhang, S., Xiang, Y., Liang, B., Su, J., Miao, Z., … Xu, R. (2022). CLLE: A Benchmark for Continual Language Learning Evaluation in Multilingual Machine Translation. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 428–443). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.368

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free