Entropy-Based Stability-Plasticity for Lifelong Learning

Abstract

The ability to continuously learn remains elusive for deep learning models. Unlike humans, models cannot accumulate knowledge in their weights when learning new tasks, mainly due to an excess of plasticity and the low incentive to reuse weights when training a new task. To address the stability-plasticity dilemma in neural networks, we propose a novel method called Entropy-based Stability-Plasticity (ESP). Our approach can decide dynamically how much each model layer should be modified via a plasticity factor. We incorporate branch layers and an entropy-based criterion into the model to find such factor. Our experiments in the domains of natural language and vision show the effectiveness of our approach in leveraging prior knowledge by reducing interference. Also, in some cases, it is possible to freeze layers during training leading to speed up in training.

Cite

Text

Araujo et al. "Entropy-Based Stability-Plasticity for Lifelong Learning." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2022. doi:10.1109/CVPRW56347.2022.00416

Markdown

[Araujo et al. "Entropy-Based Stability-Plasticity for Lifelong Learning." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2022.](https://mlanthology.org/cvprw/2022/araujo2022cvprw-entropybased/) doi:10.1109/CVPRW56347.2022.00416

BibTeX

@inproceedings{araujo2022cvprw-entropybased,
  title     = {{Entropy-Based Stability-Plasticity for Lifelong Learning}},
  author    = {Araujo, Vladimir and Hurtado, Julio and Soto, Alvaro and Moens, Marie-Francine},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2022},
  pages     = {3720-3727},
  doi       = {10.1109/CVPRW56347.2022.00416},
  url       = {https://mlanthology.org/cvprw/2022/araujo2022cvprw-entropybased/}
}