Are We Really Unlearning? the Presence of Residual Knowledge in Machine Unlearning
Abstract
Machine unlearning seeks to remove a set of forget samples from a pre-trained model to comply with emerging privacy regulations. While existing machine unlearning algorithms focus on effectiveness by either achieving indistinguishability from a re-trained model or closely matching its accuracy, they often overlook the vulnerability of unlearned models to slight perturbations of forget samples. In this paper, we identify a novel privacy vulnerability in unlearning, which we term residual knowledge. We find that even when an unlearned model no longer recognizes a forget sample---effectively removing direct knowledge of the sample---residual knowledge often persists in its vicinity, which a re-trained model does not recognize at all. Addressing residual knowledge should become a key consideration in the design of future unlearning algorithms.
Cite
Text
Hsu et al. "Are We Really Unlearning? the Presence of Residual Knowledge in Machine Unlearning." ICLR 2025 Workshops: ICBINB, 2025.Markdown
[Hsu et al. "Are We Really Unlearning? the Presence of Residual Knowledge in Machine Unlearning." ICLR 2025 Workshops: ICBINB, 2025.](https://mlanthology.org/iclrw/2025/hsu2025iclrw-we/)BibTeX
@inproceedings{hsu2025iclrw-we,
title = {{Are We Really Unlearning? the Presence of Residual Knowledge in Machine Unlearning}},
author = {Hsu, Hsiang and Niroula, Pradeep and He, Zichang and Chen, Chun-Fu},
booktitle = {ICLR 2025 Workshops: ICBINB},
year = {2025},
url = {https://mlanthology.org/iclrw/2025/hsu2025iclrw-we/}
}