Growing a Brain: Fine-Tuning by Increasing Model Capacity
Abstract
CNNs have made an undeniable impact on computer vision through the ability to learn high-capacity models with large annotated training sets. One of their remarkable properties is the ability to transfer knowledge from a large source dataset to a (typically smaller) target dataset. This is usually accomplished through fine-tuning a fixed-size network on new target data. Indeed, virtually every contemporary visual recognition system makes use of fine-tuning to transfer knowledge from ImageNet. In this work, we analyze what components and parameters change during fine-tuning, and discover that increasing model capacity allows for more natural model adaptation through fine-tuning. By making an analogy to developmental learning, we demonstrate that growing a CNN with additional units, either by widening existing layers or deepening the overall network, significantly outperforms classic fine-tuning approaches. But in order to properly grow a network, we show that newly-added units must be appropriately normalized to allow for a pace of learning that is consistent with existing units. We empirically validate our approach on several benchmark datasets, producing state-of-the-art results.
Cite
Text
Wang et al. "Growing a Brain: Fine-Tuning by Increasing Model Capacity." Conference on Computer Vision and Pattern Recognition, 2017. doi:10.1109/CVPR.2017.323Markdown
[Wang et al. "Growing a Brain: Fine-Tuning by Increasing Model Capacity." Conference on Computer Vision and Pattern Recognition, 2017.](https://mlanthology.org/cvpr/2017/wang2017cvpr-growing/) doi:10.1109/CVPR.2017.323BibTeX
@inproceedings{wang2017cvpr-growing,
title = {{Growing a Brain: Fine-Tuning by Increasing Model Capacity}},
author = {Wang, Yu-Xiong and Ramanan, Deva and Hebert, Martial},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2017},
doi = {10.1109/CVPR.2017.323},
url = {https://mlanthology.org/cvpr/2017/wang2017cvpr-growing/}
}