AdaFilter: Adaptive Filter Fine-Tuning for Deep Transfer Learning

Abstract

There is an increasing number of pre-trained deep neural network models. However, it is still unclear how to effectively use these models for a new task. Transfer learning, which aims to transfer knowledge from source tasks to a target task, is an effective solution to this problem. Fine-tuning is a popular transfer learning technique for deep neural networks where a few rounds of training are applied to the parameters of a pre-trained model to adapt them to a new task. Despite its popularity, in this paper we show that fine-tuning suffers from several drawbacks. We propose an adaptive fine-tuning approach, called AdaFilter, which selects only a part of the convolutional filters in the pre-trained model to optimize on a per-example basis. We use a recurrent gated network to selectively fine-tune convolutional filters based on the activations of the previous layer. We experiment with 7 public image classification datasets and the results show that AdaFilter can reduce the average classification error of the standard fine-tuning by 2.54%.

Cite

Text

Guo et al. "AdaFilter: Adaptive Filter Fine-Tuning for Deep Transfer Learning." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I04.5824

Markdown

[Guo et al. "AdaFilter: Adaptive Filter Fine-Tuning for Deep Transfer Learning." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/guo2020aaai-adafilter/) doi:10.1609/AAAI.V34I04.5824

BibTeX

@inproceedings{guo2020aaai-adafilter,
  title     = {{AdaFilter: Adaptive Filter Fine-Tuning for Deep Transfer Learning}},
  author    = {Guo, Yunhui and Li, Yandong and Wang, Liqiang and Rosing, Tajana},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2020},
  pages     = {4060-4066},
  doi       = {10.1609/AAAI.V34I04.5824},
  url       = {https://mlanthology.org/aaai/2020/guo2020aaai-adafilter/}
}