A Resilient and Accessible Distribution-Preserving Watermark for Large Language Models

Abstract

Watermarking techniques offer a promising way to identify machine-generated content via embedding covert information into the contents generated from language models. A challenge in the domain lies in preserving the distribution of original generated content after watermarking. Our research extends and improves upon existing watermarking framework, placing emphasis on the importance of a Distribution-Preserving (DiP) watermark. Contrary to the current strategies, our proposed DiPmark simultaneously preserves the original token distribution during watermarking (distribution-preserving), is detectable without access to the language model API and prompts (accessible), and is provably robust to moderate changes of tokens (resilient). DiPmark operates by selecting a random set of tokens prior to the generation of a word, then modifying the token distribution through a distribution-preserving reweight function to enhance the probability of these selected tokens during the sampling process. Extensive empirical evaluation on various language models and tasks demonstrates our approach’s distribution-preserving property, accessibility, and resilience, making it a effective solution for watermarking tasks that demand impeccable quality preservation.

Cite

Text

Wu et al. "A Resilient and Accessible Distribution-Preserving Watermark for Large Language Models." International Conference on Machine Learning, 2024.

Markdown

[Wu et al. "A Resilient and Accessible Distribution-Preserving Watermark for Large Language Models." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/wu2024icml-resilient/)

BibTeX

@inproceedings{wu2024icml-resilient,
  title     = {{A Resilient and Accessible Distribution-Preserving Watermark for Large Language Models}},
  author    = {Wu, Yihan and Hu, Zhengmian and Guo, Junfeng and Zhang, Hongyang and Huang, Heng},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {53443-53470},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/wu2024icml-resilient/}
}