On Transferability of Prompt Tuning for Natural Language Processing

74Citations
Citations of this article
104Readers
Mendeley users who have this article in their library.

Abstract

Prompt tuning (PT) is a promising parameter-efficient method to utilize extremely large pre-trained language models (PLMs), which can achieve comparable performance to full-parameter fine-tuning by only tuning a few soft prompts. However, PT requires much more training time than fine-tuning. Intuitively, knowledge transfer can help to improve the efficiency. To explore whether we can improve PT via prompt transfer, we empirically investigate the transferability of soft prompts across different downstream tasks and PLMs in this work. We find that (1) in zero-shot setting, trained soft prompts can effectively transfer to similar tasks on the same PLM and also to other PLMs with a cross-model projector trained on similar tasks; (2) when used as initialization, trained soft prompts of similar tasks and projected prompts of other PLMs can significantly accelerate training and also improve the performance of PT. Moreover, to explore what decides prompt transferability, we investigate various transferability indicators and find that the overlapping rate of activated neurons strongly reflects the transferability, which suggests how the prompts stimulate PLMs is essential. Our findings show that prompt transfer is promising for improving PT, and further research shall focus more on prompts' stimulation to PLMs. The source code can be obtained from https://github.com/thunlp/Prompt-Transferability.

References Powered by Scopus

SQuad: 100,000+ questions for machine comprehension of text

3979Citations
N/AReaders
Get full text

A large annotated corpus for learning natural language inference

2534Citations
N/AReaders
Get full text

Prefix-tuning: Optimizing continuous prompts for generation

1712Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Parameter-efficient fine-tuning of large-scale pre-trained language models

338Citations
N/AReaders
Get full text

Dual-Path Adaptation from Image to Video Transformers

24Citations
N/AReaders
Get full text

Towards a comprehensive understanding of the impact of large language models on natural language processing: challenges, opportunities and future directions

22Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Su, Y., Wang, X., Qin, Y., Chan, C. M., Lin, Y., Wang, H., … Zhou, J. (2022). On Transferability of Prompt Tuning for Natural Language Processing. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 3949–3969). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.290

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 27

69%

Researcher 8

21%

Professor / Associate Prof. 2

5%

Lecturer / Post doc 2

5%

Readers' Discipline

Tooltip

Computer Science 39

85%

Engineering 3

7%

Agricultural and Biological Sciences 2

4%

Linguistics 2

4%

Save time finding and organizing research with Mendeley

Sign up for free