Rethinking LLM Unlearning Objectives: A Gradient Perspective and Go Beyond

Abstract

Large language models (LLMs) should undergo rigorous audits to identify potential risks, such as copyright and privacy infringements. Once these risks emerge, timely updates are crucial to remove undesirable responses, ensuring legal and safe model usage. It has spurred recent research into LLM unlearning, focusing on erasing targeted undesirable knowledge without compromising the integrity of other, non-targeted responses. Existing studies have introduced various unlearning objectives to pursue LLM unlearning without necessitating complete retraining. However, each of these objectives has unique properties, and no unified framework is currently available to comprehend them thoroughly. To fill the gap, we propose the metric of the G-effect, quantifying the impacts of unlearning objectives on model performance from a gradient lens. A significant advantage of our metric is its broad ability to detail the unlearning impacts from various aspects across instances, updating steps, and LLM layers. Accordingly, the G-effect offers new insights into identifying drawbacks of existing unlearning objectives, further motivating us to explore a series of candidate solutions for their mitigation and improvements. Finally, we outline promising directions that merit further studies, aiming at contributing to the community to advance this critical field.

Cite

Text

Wang et al. "Rethinking LLM Unlearning Objectives: A Gradient Perspective and Go Beyond." International Conference on Learning Representations, 2025.

Markdown

[Wang et al. "Rethinking LLM Unlearning Objectives: A Gradient Perspective and Go Beyond." International Conference on Learning Representations, 2025.](https://mlanthology.org/iclr/2025/wang2025iclr-rethinking/)

BibTeX

@inproceedings{wang2025iclr-rethinking,
  title     = {{Rethinking LLM Unlearning Objectives: A Gradient Perspective and Go Beyond}},
  author    = {Wang, Qizhou and Zhou, Jin Peng and Zhou, Zhanke and Shin, Saebyeol and Han, Bo and Weinberger, Kilian Q},
  booktitle = {International Conference on Learning Representations},
  year      = {2025},
  url       = {https://mlanthology.org/iclr/2025/wang2025iclr-rethinking/}
}