MVP: Multi-task Supervised Pre-training for Natural Language Generation

14Citations
Citations of this article
51Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Pre-trained language models (PLMs) have achieved remarkable success in natural language generation (NLG) tasks. Up to now, most NLG-oriented PLMs are pre-trained in an unsupervised manner using the large-scale general corpus. In the meanwhile, an increasing number of models pre-trained with labeled data (i.e., “supervised pre-training”) showcase superior performance compared to unsupervised pre-trained models. Motivated by the success of supervised pre-training, we propose Multitask superVised Pre-training (MVP) for natural language generation. We collect a large-scale natural language generation corpus, MVPCorpus, from 77 datasets over 11 diverse NLG tasks. Then we unify these examples into a general text-to-text format to pre-train the text generation model MVP in a supervised manner. For each task, we further pre-train specific soft prompts to stimulate the model's capacity to perform a specific task. Our MVP model can be seen as a practice that utilizes recent instruction tuning on relatively small PLMs. Extensive experiments have demonstrated the effectiveness and generality of our MVP model in a number of NLG tasks, which achieves state-of-the-art performance on 13 out of 17 datasets, outperforming BART by 9.3% and Flan-T5 by 5.8%.

Cite

CITATION STYLE

APA

Tang, T., Li, J., Zhao, W. X., & Wen, J. R. (2023). MVP: Multi-task Supervised Pre-training for Natural Language Generation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 8758–8794). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.558

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free