Watermarking LLMs with Weight Quantization

5Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Abuse of large language models reveals high risks as large language models are being deployed at an astonishing speed. It is important to protect the model weights to avoid malicious usage that violates licenses of open-source large language models. This paper proposes a novel watermarking strategy that plants watermarks in the quantization process of large language models without pre-defined triggers during inference. The watermark works when the model is used in the fp32 mode and remains hidden when the model is quantized to int8, in this way, the users can only inference the model without further supervised fine-tuning of the model. We successfully plant the watermark into open-source large language model weights including GPT-Neo and LLaMA. We hope our proposed method can provide a potential direction for protecting model weights in the era of large language model applications.

Cite

CITATION STYLE

APA

Li, L., Jiang, B., Wang, P., Ren, K., Yan, H., & Qiu, X. (2023). Watermarking LLMs with Weight Quantization. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 3368–3378). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.220

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free