Towards Unified Prompt Tuning for Few-shot Text Classification

20Citations
Citations of this article
57Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-shot text classification by employing task-specific prompts. Yet, PLMs are unfamiliar with prompt-style expressions during pre-training, which limits the few-shot learning performance on downstream tasks. It would be desirable if the models can acquire some prompting knowledge before adapting to specific NLP tasks. We present the Unified Prompt Tuning (UPT) framework, leading to better few-shot text classification for BERT-style models by explicitly capturing prompting semantics from non-target NLP datasets. In UPT, a novel paradigm Prompt-Options-Verbalizer is proposed for joint prompt learning across different NLP tasks, forcing PLMs to capture task-invariant prompting knowledge. We further design a self-supervised task named Knowledge-enhanced Selective Masked Language Modeling to improve the PLM's generalization abilities for accurate adaptation to previously unseen tasks. After multi-task learning across multiple tasks, the PLM can be better prompt-tuned towards any dissimilar target tasks in low-resourced settings. Experiments over a variety of NLP tasks show that UPT consistently outperforms state-of-the-arts for prompt-based fine-tuning.

Cite

CITATION STYLE

APA

Wang, J., Wang, C., Luo, F., Tan, C., Qiu, M., Yang, F., … Gao, M. (2022). Towards Unified Prompt Tuning for Few-shot Text Classification. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 524–536). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.76

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free