Multilingual T5 (MT5; Xue et al. 2020) pretrains a sequence-to-sequence model on massive monolingual texts, which has shown promising results on many cross-lingual tasks. In this paper, we improve multilingual text-to-text transfer Transformer with translation pairs (MT6). Specifically, we explore three cross-lingual text-to-text pre-training tasks, namely, machine translation, translation pair span corruption, and translation span corruption. In addition, we propose a partially non-autoregressive objective for text-to-text pretraining. We evaluate the methods on eight multilingual benchmark datasets, including sentence classification, named entity recognition, question answering, and abstractive summarization. Experimental results show that the proposed MT6 improves cross-lingual transferability over MT5.
CITATION STYLE
Chi, Z., Dong, L., Ma, S., Huang, S., Mao, X. L., Huang, H., & Wei, F. (2021). mT6: Multilingual Pretrained Text-to-Text Transformer with Translation Pairs. In EMNLP 2021 - 2021 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 1671–1683). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.emnlp-main.125
Mendeley helps you to discover research relevant for your work.