Bias and Variance of Post-Processing in Differential Privacy

Abstract

Post-processing immunity is a fundamental property of differential privacy: it enables the application of arbitrary data-independent transformations to the results of differentially private outputs without affecting their privacy guarantees. When query outputs must satisfy domain constraints, post-processing can be used to project them back onto the feasibility region. Moreover, when the feasible region is convex, a widely adopted class of post-processing steps is also guaranteed to improve accuracy. Post-processing has been applied successfully in many applications including census data, energy systems, and mobility. However, its effects on the noise distribution is poorly understood: It is often argued that post-processing may introduce bias and increase variance. This paper takes a first step towards understanding the properties of post-processing. It considers the release of census data and examines, both empirically and theoretically, the behavior of a widely adopted class of post-processing functions.

Cite

Text

Zhu et al. "Bias and Variance of Post-Processing in Differential Privacy." AAAI Conference on Artificial Intelligence, 2021. doi:10.1609/AAAI.V35I12.17333

Markdown

[Zhu et al. "Bias and Variance of Post-Processing in Differential Privacy." AAAI Conference on Artificial Intelligence, 2021.](https://mlanthology.org/aaai/2021/zhu2021aaai-bias/) doi:10.1609/AAAI.V35I12.17333

BibTeX

@inproceedings{zhu2021aaai-bias,
  title     = {{Bias and Variance of Post-Processing in Differential Privacy}},
  author    = {Zhu, Keyu and Van Hentenryck, Pascal and Fioretto, Ferdinando},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2021},
  pages     = {11177-11184},
  doi       = {10.1609/AAAI.V35I12.17333},
  url       = {https://mlanthology.org/aaai/2021/zhu2021aaai-bias/}
}