Instance-Optimality in Differential Privacy via Approximate Inverse Sensitivity Mechanisms

Abstract

We study and provide instance-optimal algorithms in differential privacy by extending and approximating the inverse sensitivity mechanism. We provide two approximation frameworks, one which only requires knowledge of local sensitivities, and a gradient-based approximation for optimization problems, which are efficiently computable for a broad class of functions. We complement our analysis with instance-specific lower bounds for vector-valued functions, which demonstrate that our mechanisms are (nearly) instance-optimal under certain assumptions and that minimax lower bounds may not provide an accurate estimate of the hardness of a problem in general: our algorithms can significantly outperform minimax bounds for well-behaved instances. Finally, we use our approximation framework to develop private mechanisms for unbounded-range mean estimation, principal component analysis, and linear regression. For PCA, our mechanisms give an efficient (pure) differentially private algorithm with near-optimal rates.

Cite

Text

Asi and Duchi. "Instance-Optimality in Differential Privacy via Approximate Inverse Sensitivity Mechanisms." Neural Information Processing Systems, 2020.

Markdown

[Asi and Duchi. "Instance-Optimality in Differential Privacy via Approximate Inverse Sensitivity Mechanisms." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/asi2020neurips-instanceoptimality/)

BibTeX

@inproceedings{asi2020neurips-instanceoptimality,
  title     = {{Instance-Optimality in Differential Privacy via Approximate Inverse Sensitivity Mechanisms}},
  author    = {Asi, Hilal and Duchi, John C.},
  booktitle = {Neural Information Processing Systems},
  year      = {2020},
  url       = {https://mlanthology.org/neurips/2020/asi2020neurips-instanceoptimality/}
}