Membership Inference Attacks via Adversarial Examples

Abstract

The raise of machine learning and deep learning led to significant improvement in several domains. This change is supported by both the dramatic rise in computation power and the collection of large datasets. Such massive datasets often include personal data which can represent a threat to privacy. Membership inference attacks are a novel direction of research which aims at recovering training data used by a learning algorithm. In this paper, we develop a mean to measure the leakage of training data leveraging a quantity appearing as a proxy of the total variation of a trained model near its training samples. We extend our work by providing a novel defense mechanism. Our contributions are supported by empirical evidence through convincing numerical experiments.

Cite

Text

Jalalzai et al. "Membership Inference Attacks via Adversarial Examples." NeurIPS 2022 Workshops: TSRML, 2022.

Markdown

[Jalalzai et al. "Membership Inference Attacks via Adversarial Examples." NeurIPS 2022 Workshops: TSRML, 2022.](https://mlanthology.org/neuripsw/2022/jalalzai2022neuripsw-membership/)

BibTeX

@inproceedings{jalalzai2022neuripsw-membership,
  title     = {{Membership Inference Attacks via Adversarial Examples}},
  author    = {Jalalzai, Hamid and Kadoche, Elie and Leluc, Rémi and Plassier, Vincent},
  booktitle = {NeurIPS 2022 Workshops: TSRML},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/jalalzai2022neuripsw-membership/}
}