Controlling Forgetting with Test-Time Data in Continual Learning

Abstract

Foundational vision-language models excel in various tasks but require updates as new tasks or domains emerge. Current Continual Learning (CL) methods, which focus on supervised training, often suffer from significant forgetting, performing worse than the original models in zero-shot scenarios. This work proposes leveraging test-time, unsupervised data in a self-supervised manner to refresh the model’s memory of previously learned tasks, minimizing forgetting without additional labeling. By introducing a student-teacher framework with gradient-based sparse parameter updates, the approach enhances performance on prior tasks and reduces reliance on offline memory buffers, effectively improving continual learning outcomes.

Cite

Text

Singh et al. "Controlling Forgetting with Test-Time Data in Continual Learning." NeurIPS 2024 Workshops: AFM, 2024.

Markdown

[Singh et al. "Controlling Forgetting with Test-Time Data in Continual Learning." NeurIPS 2024 Workshops: AFM, 2024.](https://mlanthology.org/neuripsw/2024/singh2024neuripsw-controlling/)

BibTeX

@inproceedings{singh2024neuripsw-controlling,
  title     = {{Controlling Forgetting with Test-Time Data in Continual Learning}},
  author    = {Singh, Vaibhav and Aljundi, Rahaf and Belilovsky, Eugene},
  booktitle = {NeurIPS 2024 Workshops: AFM},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/singh2024neuripsw-controlling/}
}