Learning to Forget Using Hypernetworks

Abstract

Machine unlearning is gaining increasing attention as a way to remove adversarial data poisoning attacks from already trained models and to comply with privacy and AI regulations. The objective is to unlearn the effect of undesired data from a trained model while maintaining performance on the remaining data. This paper introduces HyperForget, a novel machine unlearning framework that leverages hypernetworks– neural networks that generate parameters for other networks– to dynamically sample models that lack knowledge of targeted data while preserving essential capabilities. Leveraging diffusion models, we implement two Diffusion HyperForget Networks and used them to sample unlearned models in Proof-of-Concept experiments. The unlearned models obtained zero accuracy on the forget set, while preserving good accuracy on the retain sets, highlighting the potential of HyperForget for dynamic targeted data removal and a promising direction for developing adaptive machine unlearning algorithms.

Cite

Text

Rangel et al. "Learning to Forget Using Hypernetworks." NeurIPS 2024 Workshops: AdvML-Frontiers, 2024.

Markdown

[Rangel et al. "Learning to Forget Using Hypernetworks." NeurIPS 2024 Workshops: AdvML-Frontiers, 2024.](https://mlanthology.org/neuripsw/2024/rangel2024neuripsw-learning/)

BibTeX

@inproceedings{rangel2024neuripsw-learning,
  title     = {{Learning to Forget Using Hypernetworks}},
  author    = {Rangel, Jose Miguel Lara and Anwar, Usman and Schoepf, Stefan and Foster, Jack and Krueger, David},
  booktitle = {NeurIPS 2024 Workshops: AdvML-Frontiers},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/rangel2024neuripsw-learning/}
}