Exploring versatile generative language model via parameter-efficient transfer learning

85Citations
Citations of this article
108Readers
Mendeley users who have this article in their library.

Abstract

Fine-tuning pre-trained generative language models to down-stream language generation tasks has shown promising results. However, this comes with the cost of having a single, large model for each task, which is not ideal in low-memory/power scenarios (e.g., mobile). In this paper, we propose an effective way to fine-tune multiple down-stream generation tasks simultaneously using a single, large pre-trained model. The experiments on five diverse language generation tasks show that by just using an additional 2-3% parameters for each task, our model can maintain or even improve the performance of fine-tuning the whole model1

Cite

CITATION STYLE

APA

Lin, Z., Madotto, A., & Fung, P. (2020). Exploring versatile generative language model via parameter-efficient transfer learning. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 441–459). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.41

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free