CLICK: Controllable Text Generation with Sequence Likelihood Contrastive Learning

16Citations
Citations of this article
29Readers
Mendeley users who have this article in their library.
Get full text

Abstract

It has always been an important yet challenging problem to control language models to avoid generating texts with undesirable attributes, such as toxic language and unnatural repetition. We introduce CLICK for controllable text generation, which needs no modification to the model architecture and facilitates out-of-the-box use of trained models. It employs a contrastive loss on sequence likelihood, which fundamentally decreases the generation probability of negative samples (i.e., generations with undesirable attributes). It also adopts a novel likelihood ranking-based strategy to construct contrastive samples from model generations. On the tasks of language detoxification, sentiment steering, and repetition reduction, we show that CLICK outperforms strong baselines of controllable text generation and demonstrate the superiority of CLICK's sample construction strategy.

Cite

CITATION STYLE

APA

Zheng, C., Ke, P., Zhang, Z., & Huang, M. (2023). CLICK: Controllable Text Generation with Sequence Likelihood Contrastive Learning. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 1022–1040). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.65

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free