Bias and Variance of Post-Processing in Differential Privacy
Abstract
Post-processing immunity is a fundamental property of differential privacy: it enables the application of arbitrary data-independent transformations to the results of differentially private outputs without affecting their privacy guarantees. When query outputs must satisfy domain constraints, post-processing can be used to project them back onto the feasibility region. Moreover, when the feasible region is convex, a widely adopted class of post-processing steps is also guaranteed to improve accuracy. Post-processing has been applied successfully in many applications including census data, energy systems, and mobility. However, its effects on the noise distribution is poorly understood: It is often argued that post-processing may introduce bias and increase variance. This paper takes a first step towards understanding the properties of post-processing. It considers the release of census data and examines, both empirically and theoretically, the behavior of a widely adopted class of post-processing functions.
Cite
Text
Zhu et al. "Bias and Variance of Post-Processing in Differential Privacy." AAAI Conference on Artificial Intelligence, 2021. doi:10.1609/AAAI.V35I12.17333Markdown
[Zhu et al. "Bias and Variance of Post-Processing in Differential Privacy." AAAI Conference on Artificial Intelligence, 2021.](https://mlanthology.org/aaai/2021/zhu2021aaai-bias/) doi:10.1609/AAAI.V35I12.17333BibTeX
@inproceedings{zhu2021aaai-bias,
title = {{Bias and Variance of Post-Processing in Differential Privacy}},
author = {Zhu, Keyu and Van Hentenryck, Pascal and Fioretto, Ferdinando},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2021},
pages = {11177-11184},
doi = {10.1609/AAAI.V35I12.17333},
url = {https://mlanthology.org/aaai/2021/zhu2021aaai-bias/}
}