Clip-Tuning: Towards Derivative-free Prompt Learning with a Mixture of Rewards

5Citations
Citations of this article
25Readers
Mendeley users who have this article in their library.

Abstract

Derivative-free prompt learning has emerged as a lightweight alternative to prompt tuning, which only requires model inference to optimize the prompts. However, existing work did not take full advantage of the over-parameterized characteristics of large pretrained language models (PLMs). In this paper, we propose Clip-Tuning, a simple yet effective method that adopts diverse frozen “thinned” networks of PLMs to obtain a mixture of rewards and thus advance the derivative-free prompt learning. The thinned networks consist of all the hidden units that survive a stationary dropout strategy, whose inference predictions reflect an ensemble of partial views over prompted training samples. Our method outperforms previous gradient-free prompt learning methods and achieves parity with gradient-based counterparts on seven language understanding benchmarks under few-shot settings.

Cite

CITATION STYLE

APA

Chai, Y., Wang, S., Sun, Y., Tian, H., Wu, H., & Wang, H. (2022). Clip-Tuning: Towards Derivative-free Prompt Learning with a Mixture of Rewards. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 108–117). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.8

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free