Publicly Detectable Watermarking for Language Models

Published: 01 Jan 2023, Last Modified: 29 Sept 2024IACR Cryptol. ePrint Arch. 2023EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: We present a highly detectable, trustless watermarking scheme for LLMs: the detection algorithm contains no secret information, and it is executable by anyone. We embed a publicly-verifiable cryptographic signature into LLM output using rejection sampling. We prove that our scheme is cryptographically correct, sound, and distortion-free. We make novel uses of error-correction techniques to overcome periods of low entropy, a barrier for all prior watermarking schemes. We implement our scheme and make empirical measurements over open models in the 2.7B to 70B parameter range. Our experiments suggest that our formal claims are met in practice.
Loading