K for the Price of 1: Parameter-Efficient Multi-Task and Transfer Learning
Abstract
We introduce a novel method that enables parameter-efficient transfer and multi-task learning with deep neural networks. The basic approach is to learn a model patch - a small set of parameters - that will specialize to each task, instead of fine-tuning the last layer or the entire network. For instance, we show that learning a set of scales and biases is sufficient to convert a pretrained network to perform well on qualitatively different problems (e.g. converting a Single Shot MultiBox Detection (SSD) model into a 1000-class image classification model while reusing 98% of parameters of the SSD feature extractor). Similarly, we show that re-learning existing low-parameter layers (such as depth-wise convolutions) while keeping the rest of the network frozen also improves transfer-learning accuracy significantly. Our approach allows both simultaneous (multi-task) as well as sequential transfer learning. In several multi-task learning problems, despite using much fewer parameters than traditional logits-only fine-tuning, we match single-task performance.
Cite
Text
Mudrakarta et al. "K for the Price of 1: Parameter-Efficient Multi-Task and Transfer Learning." International Conference on Learning Representations, 2019.Markdown
[Mudrakarta et al. "K for the Price of 1: Parameter-Efficient Multi-Task and Transfer Learning." International Conference on Learning Representations, 2019.](https://mlanthology.org/iclr/2019/mudrakarta2019iclr-price/)BibTeX
@inproceedings{mudrakarta2019iclr-price,
title = {{K for the Price of 1: Parameter-Efficient Multi-Task and Transfer Learning}},
author = {Mudrakarta, Pramod Kaushik and Sandler, Mark and Zhmoginov, Andrey and Howard, Andrew},
booktitle = {International Conference on Learning Representations},
year = {2019},
url = {https://mlanthology.org/iclr/2019/mudrakarta2019iclr-price/}
}