Analyzing Privacy Leakage in Machine Learning via Multiple Hypothesis Testing: A Lesson from Fano

Abstract

Differential privacy (DP) is by far the most widely accepted framework for mitigating privacy risks in machine learning. However, exactly how small the privacy parameter $\epsilon$ needs to be to protect against certain privacy risks in practice is still not well-understood. In this work, we study data reconstruction attacks for discrete data and analyze it under the framework of multiple hypothesis testing. For a learning algorithm satisfying $(\alpha, \epsilon)$-Renyi DP, we utilize different variants of the celebrated Fano’s inequality to upper bound the attack advantage of a data reconstruction adversary. Our bound can be numerically computed to relate the parameter $\epsilon$ to the desired level of privacy protection in practice, and complements the empirical evidence for the effectiveness of DP against data reconstruction attacks even at relatively large values of $\epsilon$.

Cite

Text

Guo et al. "Analyzing Privacy Leakage in Machine Learning via Multiple Hypothesis Testing: A Lesson from Fano." International Conference on Machine Learning, 2023.

Markdown

[Guo et al. "Analyzing Privacy Leakage in Machine Learning via Multiple Hypothesis Testing: A Lesson from Fano." International Conference on Machine Learning, 2023.](https://mlanthology.org/icml/2023/guo2023icml-analyzing/)

BibTeX

@inproceedings{guo2023icml-analyzing,
  title     = {{Analyzing Privacy Leakage in Machine Learning via Multiple Hypothesis Testing: A Lesson from Fano}},
  author    = {Guo, Chuan and Sablayrolles, Alexandre and Sanjabi, Maziar},
  booktitle = {International Conference on Machine Learning},
  year      = {2023},
  pages     = {11998-12011},
  volume    = {202},
  url       = {https://mlanthology.org/icml/2023/guo2023icml-analyzing/}
}