Differentially Private Latent Diffusion Models

Abstract

Diffusion models (DMs) are one of the most widely used generative models for producing high quality images. However, a flurry of recent papers points out that DMs are least private forms of image generators, by extracting a significant number of near-identical replicas of training images from DMs. Existing privacy-enhancing techniques for DMs, unfortunately, do not provide a good privacy-utility tradeoff. To address this challenge, a recent paper suggest pre-training DMs with public data, then fine-tuning them with private data using DP-SGD for a relatively short period. In this paper, we further improve the current state of DMs with differential privacy (DP) by adopting the $\textit{Latent}$ Diffusion Models (LDMs). LDMs are equipped with powerful pre-trained autoencoders that map the high-dimensional pixels into lower-dimensional latent representations, in which DMs are trained, yielding a more efficient and fast training of DMs. Rather than fine-tuning the entire LDMs, we fine-tune only the $\textit{attention}$ modules of LDMs with DP-SGD, reducing the number of trainable parameters by roughly 90% and achieving a better privacy-accuracy trade-off. Our approach allows us to generate realistic, high-dimensional images (256x256) conditioned on text prompts with DP guarantees, which, to the best of our knowledge, has not been attempted before. Our approach provides a promising direction for training more powerful, yet training-efficient differentially private DMs, producing high-quality DP images.

Cite

Text

Lyu et al. "Differentially Private Latent Diffusion Models." ICLR 2024 Workshops: PML, 2024.

Markdown

[Lyu et al. "Differentially Private Latent Diffusion Models." ICLR 2024 Workshops: PML, 2024.](https://mlanthology.org/iclrw/2024/lyu2024iclrw-differentially/)

BibTeX

@inproceedings{lyu2024iclrw-differentially,
  title     = {{Differentially Private Latent Diffusion Models}},
  author    = {Lyu, Saiyue and Liu, Michael F and Vinaroz, Margarita and Park, Mijung},
  booktitle = {ICLR 2024 Workshops: PML},
  year      = {2024},
  url       = {https://mlanthology.org/iclrw/2024/lyu2024iclrw-differentially/}
}