XLM-E: Cross-lingual Language Model Pre-training via ELECTRA

50Citations
Citations of this article
124Readers
Mendeley users who have this article in their library.

Abstract

In this paper, we introduce ELECTRA-style tasks (Clark et al., 2020b) to cross-lingual language model pre-training. Specifically, we present two pre-training tasks, namely multilingual replaced token detection, and translation replaced token detection. Besides, we pretrain the model, named as XLM-E, on both multilingual and parallel corpora. Our model outperforms the baseline models on various cross-lingual understanding tasks with much less computation cost. Moreover, analysis shows that XLM-E tends to obtain better cross-lingual transferability.

Cite

CITATION STYLE

APA

Chi, Z., Huang, S., Dong, L., Ma, S., Zheng, B., Singhal, S., … Wei, F. (2022). XLM-E: Cross-lingual Language Model Pre-training via ELECTRA. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 6170–6182). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.427

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free