Interpretable and Differentially Private Predictions

Abstract

Interpretable predictions, which clarify why a machine learning model makes a particular decision, can compromise privacy by revealing the characteristics of individual data points. This raises the central question addressed in this paper: Can models be interpretable without compromising privacy? For complex “big” data fit by correspondingly rich models, balancing privacy and explainability is particularly challenging, such that this question has remained largely unexplored. In this paper, we propose a family of simple models with the aim of approximating complex models using several locally linear maps per class to provide high classification accuracy, as well as differentially private explanations on the classification. We illustrate the usefulness of our approach on several image benchmark datasets as well as a medical dataset.

Cite

Text

Harder et al. "Interpretable and Differentially Private Predictions." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I04.5827

Markdown

[Harder et al. "Interpretable and Differentially Private Predictions." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/harder2020aaai-interpretable/) doi:10.1609/AAAI.V34I04.5827

BibTeX

@inproceedings{harder2020aaai-interpretable,
  title     = {{Interpretable and Differentially Private Predictions}},
  author    = {Harder, Frederik and Bauer, Matthias and Park, Mijung},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2020},
  pages     = {4083-4090},
  doi       = {10.1609/AAAI.V34I04.5827},
  url       = {https://mlanthology.org/aaai/2020/harder2020aaai-interpretable/}
}