Subset-Based Instance Optimality in Private Estimation

Abstract

We propose a new definition of instance optimality for differentially private estimation algorithms. Our definition requires an optimal algorithm to compete, simultaneously for every dataset $D$, with the best private benchmark algorithm that (a) knows $D$ in advance and (b) is evaluated by its worst-case performance on large subsets of $D$. That is, the benchmark algorithm need not perform well when potentially extreme points are added to $D$; it only has to handle the removal of a small number of real data points that already exist. This makes our benchmark significantly stronger than those proposed in prior work. We nevertheless show, for real-valued datasets, how to construct private algorithms that achieve our notion of instance optimality when estimating a broad class of dataset properties, including means, quantiles, and $\ell_p$-norm minimizers. For means in particular, we provide a detailed analysis and show that our algorithm simultaneously matches or exceeds the asymptotic performance of existing algorithms under a range of distributional assumptions.

Cite

Text

Dick et al. "Subset-Based Instance Optimality in Private Estimation." International Conference on Machine Learning, 2023.

Markdown

[Dick et al. "Subset-Based Instance Optimality in Private Estimation." International Conference on Machine Learning, 2023.](https://mlanthology.org/icml/2023/dick2023icml-subsetbased/)

BibTeX

@inproceedings{dick2023icml-subsetbased,
  title     = {{Subset-Based Instance Optimality in Private Estimation}},
  author    = {Dick, Travis and Kulesza, Alex and Sun, Ziteng and Suresh, Ananda Theertha},
  booktitle = {International Conference on Machine Learning},
  year      = {2023},
  pages     = {7992-8014},
  volume    = {202},
  url       = {https://mlanthology.org/icml/2023/dick2023icml-subsetbased/}
}