Low Compute Unlearning via Sparse Representations
Abstract
Machine unlearning, which involves erasing knowledge about a \emph{forget set} from a trained model, can prove to be costly and infeasible using existing techniques. We propose a low-compute unlearning technique based on a discrete representational bottleneck. We show that the proposed technique efficiently unlearns the forget set and incurs negligible damage to the model's performance on the rest of the dataset. We evaluate the proposed technique on the problem of class unlearning using four datasets: CIFAR-10, CIFAR-100, LACUNA-100 and ImageNet-1k. We compare the proposed technique to SCRUB, a state-of-the-art approach which uses knowledge distillation for unlearning. Across all four datasets, the proposed technique performs as well as, if not better than SCRUB while incurring almost no computational cost.
Cite
Text
Shah et al. "Low Compute Unlearning via Sparse Representations." Transactions on Machine Learning Research, 2025.Markdown
[Shah et al. "Low Compute Unlearning via Sparse Representations." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/shah2025tmlr-low/)BibTeX
@article{shah2025tmlr-low,
title = {{Low Compute Unlearning via Sparse Representations}},
author = {Shah, Vedant and Träuble, Frederik and Malik, Ashish and Larochelle, Hugo and Mozer, Michael Curtis and Arora, Sanjeev and Bengio, Yoshua and Goyal, Anirudh},
journal = {Transactions on Machine Learning Research},
year = {2025},
url = {https://mlanthology.org/tmlr/2025/shah2025tmlr-low/}
}