RepAn: Enhanced Annealing Through Re-Parameterization

Abstract

The simulated annealing algorithm aims to improve model convergence through multiple restarts of training. However existing annealing algorithms overlook the correlation between different cycles neglecting the potential for incremental learning. We contend that a fixed network structure prevents the model from recognizing distinct features at different training stages. To this end we propose RepAn redesigning the irreversible re-parameterization (Rep) method and integrating it with annealing to enhance training. Specifically the network goes through Rep expansion restoration and backpropagation operations during training and iterating through these processes in each annealing round. Such a method exhibits good generalization and is easy to apply and we provide theoretical explanations for its effectiveness. Experiments demonstrate that our method improves baseline performance by 6.38% on the CIFAR-100 dataset and 2.80% on ImageNet achieving state-of-the-art performance in the Rep field. The code is available at https://github.com/xfey/RepAn.

Cite

Text

Fei et al. "RepAn: Enhanced Annealing Through Re-Parameterization." Conference on Computer Vision and Pattern Recognition, 2024. doi:10.1109/CVPR52733.2024.00554

Markdown

[Fei et al. "RepAn: Enhanced Annealing Through Re-Parameterization." Conference on Computer Vision and Pattern Recognition, 2024.](https://mlanthology.org/cvpr/2024/fei2024cvpr-repan/) doi:10.1109/CVPR52733.2024.00554

BibTeX

@inproceedings{fei2024cvpr-repan,
  title     = {{RepAn: Enhanced Annealing Through Re-Parameterization}},
  author    = {Fei, Xiang and Zheng, Xiawu and Wang, Yan and Chao, Fei and Wu, Chenglin and Cao, Liujuan},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2024},
  pages     = {5798-5808},
  doi       = {10.1109/CVPR52733.2024.00554},
  url       = {https://mlanthology.org/cvpr/2024/fei2024cvpr-repan/}
}