Scalable DP-SGD: Shuffling vs. Poisson Subsampling

Abstract

We provide new lower bounds on the privacy guarantee of multi-epoch Adaptive Batch Linear Queries (ABLQ) mechanism with shuffled batch sampling, demonstrating substantial gaps when compared to Poisson subsampling; prior analysis was limited to a single epoch.Since the privacy analysis of Differentially Private Stochastic Gradient Descent (DP-SGD) is obtained by analyzing the ABLQ mechanism, this brings into serious question the common practice of implementing Shuffling based DP-SGD, but reporting privacy parameters as if Poisson subsampling was used.To understand the impact of this gap on the utility of trained machine learning models, we introduce a novel practical approach to implement Poisson subsampling at scale using massively parallel computation, and efficiently train models with the same.We provide a comparison between the utility of models trained with Poisson subsampling based DP-SGD, and the optimistic estimates of utility when using shuffling, via our new lower bounds on the privacy guarantee of ABLQ with shuffling.

Cite

Text

Chua et al. "Scalable DP-SGD: Shuffling vs. Poisson Subsampling." Neural Information Processing Systems, 2024. doi:10.52202/079017-2238

Markdown

[Chua et al. "Scalable DP-SGD: Shuffling vs. Poisson Subsampling." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/chua2024neurips-scalable/) doi:10.52202/079017-2238

BibTeX

@inproceedings{chua2024neurips-scalable,
  title     = {{Scalable DP-SGD: Shuffling vs. Poisson Subsampling}},
  author    = {Chua, Lynn and Ghazi, Badih and Kamath, Pritish and Kumar, Ravi and Manurangsi, Pasin and Sinha, Amer and Zhang, Chiyuan},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-2238},
  url       = {https://mlanthology.org/neurips/2024/chua2024neurips-scalable/}
}