GPTs Don't Keep Secrets: Searching for Backdoor Watermark Triggers in Autoregressive Language Models

1Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

This work analyzes backdoor watermarks in an autoregressive transformer fine-tuned to perform a generative sequence-to-sequence task, specifically summarization. We propose and demonstrate an attack to identify trigger words or phrases by analyzing open ended generations from autoregressive models that have backdoor watermarks inserted. It is shown in our work that triggers based on random common words are easier to identify than those based on single, rare tokens. The attack proposed is easy to implement and only requires access to the model weights.

Cite

CITATION STYLE

APA

Lucas, E., & Havens, T. C. (2023). GPTs Don’t Keep Secrets: Searching for Backdoor Watermark Triggers in Autoregressive Language Models. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 242–248). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.trustnlp-1.21

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free