mT6: Multilingual Pretrained Text-to-Text Transformer with Translation Pairs

36Citations
Citations of this article
113Readers
Mendeley users who have this article in their library.

Abstract

Multilingual T5 (MT5; Xue et al. 2020) pretrains a sequence-to-sequence model on massive monolingual texts, which has shown promising results on many cross-lingual tasks. In this paper, we improve multilingual text-to-text transfer Transformer with translation pairs (MT6). Specifically, we explore three cross-lingual text-to-text pre-training tasks, namely, machine translation, translation pair span corruption, and translation span corruption. In addition, we propose a partially non-autoregressive objective for text-to-text pretraining. We evaluate the methods on eight multilingual benchmark datasets, including sentence classification, named entity recognition, question answering, and abstractive summarization. Experimental results show that the proposed MT6 improves cross-lingual transferability over MT5.

Cite

CITATION STYLE

APA

Chi, Z., Dong, L., Ma, S., Huang, S., Mao, X. L., Huang, H., & Wei, F. (2021). mT6: Multilingual Pretrained Text-to-Text Transformer with Translation Pairs. In EMNLP 2021 - 2021 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 1671–1683). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.emnlp-main.125

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free