Provable Membership Inference Privacy

Abstract

In applications involving sensitive data, such as finance and healthcare, the necessity for preserving data privacy can be a significant barrier to machine learning model development. Differential privacy (DP) has emerged as one canonical standard for provable privacy. However, DP's strong theoretical guarantees often come at the cost of a large drop in its utility for machine learning; and DP guarantees themselves can be difficult to interpret. In this work, we propose a novel privacy notion, membership inference privacy (MIP), to address these challenges. We give a precise characterization of the relationship between MIP and DP, and show that MIP can be achieved using less randomness compared to the amount required for guaranteeing DP, leading to smaller drop in utility. MIP also guarantees are easily interpretable in terms of the success rate of membership inference attacks. Our theoretical results also give rise to a simple algorithm for guaranteeing MIP which can be used as a wrapper around any algorithm with a continuous output, including parametric model training.

Cite

Text

Izzo et al. "Provable Membership Inference Privacy." NeurIPS 2022 Workshops: TSRML, 2022.

Markdown

[Izzo et al. "Provable Membership Inference Privacy." NeurIPS 2022 Workshops: TSRML, 2022.](https://mlanthology.org/neuripsw/2022/izzo2022neuripsw-provable/)

BibTeX

@inproceedings{izzo2022neuripsw-provable,
  title     = {{Provable Membership Inference Privacy}},
  author    = {Izzo, Zachary and Yoon, Jinsung and Arik, Sercan O and Zou, James},
  booktitle = {NeurIPS 2022 Workshops: TSRML},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/izzo2022neuripsw-provable/}
}