Evaluating the cross-lingual effectiveness of massively multilingual neural machine translation

38Citations
Citations of this article
82Readers
Mendeley users who have this article in their library.

Abstract

The recently proposed massively multilingual neural machine translation (NMT) system has been shown to be capable of translating over 100 languages to and from English within a single model (Aharoni, Johnson, and Firat 2019). Its improved translation performance on low resource languages hints at potential cross-lingual transfer capability for downstream tasks. In this paper, we evaluate the cross-lingual effectiveness of representations from the encoder of a massively multilingual NMT model on 5 downstream classification and sequence labeling tasks covering a diverse set of over 50 languages. We compare against a strong baseline, multilingual BERT (mBERT) (Devlin et al. 2018), in different cross-lingual transfer learning scenarios and show gains in zero-shot transfer in 4 out of these 5 tasks.

Cite

CITATION STYLE

APA

Siddhant, A., Johnson, M., Tsai, H., Ari, N., Riesa, J., Bapna, A., … Raman, K. (2020). Evaluating the cross-lingual effectiveness of massively multilingual neural machine translation. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 8854–8861). AAAI press. https://doi.org/10.1609/aaai.v34i05.6414

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free