Shuffle Private Stochastic Convex Optimization

Abstract

In shuffle privacy, each user sends a collection of randomized messages to a trusted shuffler, the shuffler randomly permutes these messages, and the resulting shuffled collection of messages must satisfy differential privacy. Prior work in this model has largely focused on protocols that use a single round of communication to compute algorithmic primitives like means, histograms, and counts. In this work, we present interactive shuffle protocols for stochastic convex optimization. Our optimization protocols rely on a new noninteractive protocol for summing vectors of bounded $\ell_2$ norm. By combining this sum subroutine with techniques including mini-batch stochastic gradient descent, accelerated gradient descent, and Nesterov's smoothing method, we obtain loss guarantees for a variety of convex loss functions that significantly improve on those of the local model and sometimes match those of the central model.

Cite

Text

Cheu et al. "Shuffle Private Stochastic Convex Optimization." International Conference on Learning Representations, 2022.

Markdown

[Cheu et al. "Shuffle Private Stochastic Convex Optimization." International Conference on Learning Representations, 2022.](https://mlanthology.org/iclr/2022/cheu2022iclr-shuffle/)

BibTeX

@inproceedings{cheu2022iclr-shuffle,
  title     = {{Shuffle Private Stochastic Convex Optimization}},
  author    = {Cheu, Albert and Joseph, Matthew and Mao, Jieming and Peng, Binghui},
  booktitle = {International Conference on Learning Representations},
  year      = {2022},
  url       = {https://mlanthology.org/iclr/2022/cheu2022iclr-shuffle/}
}