RL-Tune: A Deep Reinforcement Learning Assisted Layer-Wise Fine-Tuning Approach for Transfer Learning
Abstract
Data scarcity is one of the major challenges in many real-world applications. To handle low-data regimes, practitioners often take an existing pre-trained network and fine-tune it on a data-deficient target task. In this setup, a network is pre-trained on a source dataset and fine-tuned on a different, potentially smaller, target dataset. We address two critical challenges with transfer learning via fine-tuning: (1) The required amount of fine-tuning greatly depends on the distribution shift from source to target dataset. (2) This distribution shift greatly varies by layer, thereby requiring layer-wise adjustments in fine-tuning to adapt to this distribution shift while preserving the pre-trained network's feature representation. To overcome these challenges, we propose RL-Tune, a layer-wise fine-tuning framework for transfer learning which leverages reinforcement learning to adjust learning rates as a function of the target data shift. In our RL framework, the state is a collection of the intermediate feature activations generated from training samples. The agent generates layer-wise learning rates as actions for fine-tuning based on the current state and obtains sample accuracy as the reward. RL-Tune outperforms other state-of-the-art approaches on standard transfer learning benchmarks by a large margin, e.g., 6% mean accuracy improvement on CUB-200-2011 with 15% data.
Cite
Text
Mahmud et al. "RL-Tune: A Deep Reinforcement Learning Assisted Layer-Wise Fine-Tuning Approach for Transfer Learning." ICML 2022 Workshops: Pre-Training, 2022.Markdown
[Mahmud et al. "RL-Tune: A Deep Reinforcement Learning Assisted Layer-Wise Fine-Tuning Approach for Transfer Learning." ICML 2022 Workshops: Pre-Training, 2022.](https://mlanthology.org/icmlw/2022/mahmud2022icmlw-rltune/)BibTeX
@inproceedings{mahmud2022icmlw-rltune,
title = {{RL-Tune: A Deep Reinforcement Learning Assisted Layer-Wise Fine-Tuning Approach for Transfer Learning}},
author = {Mahmud, Tanvir and Frumkin, Natalia and Marculescu, Diana},
booktitle = {ICML 2022 Workshops: Pre-Training},
year = {2022},
url = {https://mlanthology.org/icmlw/2022/mahmud2022icmlw-rltune/}
}