Derivative-free prompt learning has emerged as a lightweight alternative to prompt tuning, which only requires model inference to optimize the prompts. However, existing work did not take full advantage of the over-parameterized characteristics of large pretrained language models (PLMs). In this paper, we propose Clip-Tuning, a simple yet effective method that adopts diverse frozen “thinned” networks of PLMs to obtain a mixture of rewards and thus advance the derivative-free prompt learning. The thinned networks consist of all the hidden units that survive a stationary dropout strategy, whose inference predictions reflect an ensemble of partial views over prompted training samples. Our method outperforms previous gradient-free prompt learning methods and achieves parity with gradient-based counterparts on seven language understanding benchmarks under few-shot settings.
CITATION STYLE
Chai, Y., Wang, S., Sun, Y., Tian, H., Wu, H., & Wang, H. (2022). Clip-Tuning: Towards Derivative-free Prompt Learning with a Mixture of Rewards. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 108–117). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.8
Mendeley helps you to discover research relevant for your work.