MagMax: Leveraging Model Merging for Seamless Continual Learning
Abstract
This paper introduces a continual learning approach named , which utilizes model merging to enable large pre-trained models to continuously learn from new data without forgetting previously acquired knowledge. Distinct from traditional continual learning methods that aim to reduce forgetting during task training, combines sequential fine-tuning with a maximum magnitude weight selection for effective knowledge integration across tasks. Our initial contribution is an extensive examination of model merging techniques, revealing that simple approaches like weight averaging and random weight selection surprisingly hold up well in various continual learning contexts. More importantly, we present , a novel model-merging strategy that enables continual learning of large pre-trained models for successive tasks. Our thorough evaluation demonstrates the superiority of in various scenarios, including class- and domain-incremental learning settings. The code is available on github.
Cite
Text
Marczak et al. "MagMax: Leveraging Model Merging for Seamless Continual Learning." Proceedings of the European Conference on Computer Vision (ECCV), 2024. doi:10.1007/978-3-031-73013-9_22Markdown
[Marczak et al. "MagMax: Leveraging Model Merging for Seamless Continual Learning." Proceedings of the European Conference on Computer Vision (ECCV), 2024.](https://mlanthology.org/eccv/2024/marczak2024eccv-magmax/) doi:10.1007/978-3-031-73013-9_22BibTeX
@inproceedings{marczak2024eccv-magmax,
title = {{MagMax: Leveraging Model Merging for Seamless Continual Learning}},
author = {Marczak, Daniel and Twardowski, Bartlomiej and Trzcinski, Tomasz and Cygert, Sebastian},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2024},
doi = {10.1007/978-3-031-73013-9_22},
url = {https://mlanthology.org/eccv/2024/marczak2024eccv-magmax/}
}