Privately Publishable Per-Instance Privacy
Abstract
We consider how to privately share the personalized privacy losses incurred by objective perturbation, using per-instance differential privacy (pDP). Standard differential privacy (DP) gives us a worst-case bound that might be orders of magnitude larger than the privacy loss to a particular individual relative to a fixed dataset. The pDP framework provides a more fine-grained analysis of the privacy guarantee to a target individual, but the per-instance privacy loss itself might be a function of sensitive data. In this paper, we analyze the per-instance privacy loss of releasing a private empirical risk minimizer learned via objective perturbation, and propose a group of methods to privately and accurately publish the pDP losses at little to no additional privacy cost.
Cite
Text
Redberg and Wang. "Privately Publishable Per-Instance Privacy." Neural Information Processing Systems, 2021.Markdown
[Redberg and Wang. "Privately Publishable Per-Instance Privacy." Neural Information Processing Systems, 2021.](https://mlanthology.org/neurips/2021/redberg2021neurips-privately/)BibTeX
@inproceedings{redberg2021neurips-privately,
title = {{Privately Publishable Per-Instance Privacy}},
author = {Redberg, Rachel and Wang, Yu-Xiang},
booktitle = {Neural Information Processing Systems},
year = {2021},
url = {https://mlanthology.org/neurips/2021/redberg2021neurips-privately/}
}