Adapting Multilingual Models for Code-Mixed Translation

2Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.

Abstract

The scarcity of gold standard code-mixed to pure language parallel data makes it difficult to train translation models reliably. Prior work has addressed the paucity of parallel data with data augmentation techniques. Such methods rely heavily on external resources making systems difficult to train and scale effectively for multiple languages. We present a simple yet highly effective two-stage back-translation based training scheme for adapting multilingual models to the task of code-mixed translation which eliminates dependence on external resources. We show a substantial improvement in translation quality (measured through BLEU), beating existing prior work by up to +3.8 BLEU on code-mixed Hi→En, Mr→En, and Bn→En tasks. On the LinCE Machine Translation leader board, we achieve the highest score for code-mixed Es→En, beating existing best baseline by +6.5 BLEU, and our own stronger baseline by +1.1 BLEU.

Cite

CITATION STYLE

APA

Vavre, A., Gupta, A., & Sarawagi, S. (2022). Adapting Multilingual Models for Code-Mixed Translation. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 7162–7170). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.528

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free