Hard to Forget: Poisoning Attacks on Certified Machine Unlearning

Abstract

The right to erasure requires removal of a user's information from data held by organizations, with rigorous interpretations extending to downstream products such as learned models. Retraining from scratch with the particular user's data omitted fully removes its influence on the resulting model, but comes with a high computational cost. Machine "unlearning" mitigates the cost incurred by full retraining: instead, models are updated incrementally, possibly only requiring retraining when approximation errors accumulate. Rapid progress has been made towards privacy guarantees on the indistinguishability of unlearned and retrained models, but current formalisms do not place practical bounds on computation. In this paper we demonstrate how an attacker can exploit this oversight, highlighting a novel attack surface introduced by machine unlearning. We consider an attacker aiming to increase the computational cost of data removal. We derive and empirically investigate a poisoning attack on certified machine unlearning where strategically designed training data triggers complete retraining when removed.

Cite

Text

Marchant et al. "Hard to Forget: Poisoning Attacks on Certified Machine Unlearning." AAAI Conference on Artificial Intelligence, 2022. doi:10.1609/AAAI.V36I7.20736

Markdown

[Marchant et al. "Hard to Forget: Poisoning Attacks on Certified Machine Unlearning." AAAI Conference on Artificial Intelligence, 2022.](https://mlanthology.org/aaai/2022/marchant2022aaai-hard/) doi:10.1609/AAAI.V36I7.20736

BibTeX

@inproceedings{marchant2022aaai-hard,
  title     = {{Hard to Forget: Poisoning Attacks on Certified Machine Unlearning}},
  author    = {Marchant, Neil G. and Rubinstein, Benjamin I. P. and Alfeld, Scott},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2022},
  pages     = {7691-7700},
  doi       = {10.1609/AAAI.V36I7.20736},
  url       = {https://mlanthology.org/aaai/2022/marchant2022aaai-hard/}
}