A Computational Separation Between Private Learning and Online Learning

Abstract

A recent line of work has shown a qualitative equivalence between differentially private PAC learning and online learning: A concept class is privately learnable if and only if it is online learnable with a finite mistake bound. However, both directions of this equivalence incur significant losses in both sample and computational efficiency. Studying a special case of this connection, Gonen, Hazan, and Moran (NeurIPS 2019) showed that uniform or highly sample-efficient pure-private learners can be time-efficiently compiled into online learners. We show that, assuming the existence of one-way functions, such an efficient conversion is impossible even for general pure-private learners with polynomial sample complexity. This resolves a question of Neel, Roth, and Wu (FOCS 2019).

Cite

Text

Bun. "A Computational Separation Between Private Learning and Online Learning." Neural Information Processing Systems, 2020.

Markdown

[Bun. "A Computational Separation Between Private Learning and Online Learning." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/bun2020neurips-computational/)

BibTeX

@inproceedings{bun2020neurips-computational,
  title     = {{A Computational Separation Between Private Learning and Online Learning}},
  author    = {Bun, Mark},
  booktitle = {Neural Information Processing Systems},
  year      = {2020},
  url       = {https://mlanthology.org/neurips/2020/bun2020neurips-computational/}
}