The Advantage of Conditional Meta-Learning for Biased Regularization and Fine Tuning
Abstract
Biased regularization and fine tuning are two recent meta-learning approaches. They have been shown to be effective to tackle distributions of tasks, in which the tasks’ target vectors are all close to a common meta-parameter vector. However, these methods may perform poorly on heterogeneous environments of tasks, where the complexity of the tasks’ distribution cannot be captured by a single meta- parameter vector. We address this limitation by conditional meta-learning, inferring a conditioning function mapping task’s side information into a meta-parameter vector that is appropriate for that task at hand. We characterize properties of the environment under which the conditional approach brings a substantial advantage over standard meta-learning and we highlight examples of environments, such as those with multiple clusters, satisfying these properties. We then propose a convex meta-algorithm providing a comparable advantage also in practice. Numerical experiments confirm our theoretical findings.
Cite
Text
Denevi et al. "The Advantage of Conditional Meta-Learning for Biased Regularization and Fine Tuning." Neural Information Processing Systems, 2020.Markdown
[Denevi et al. "The Advantage of Conditional Meta-Learning for Biased Regularization and Fine Tuning." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/denevi2020neurips-advantage/)BibTeX
@inproceedings{denevi2020neurips-advantage,
title = {{The Advantage of Conditional Meta-Learning for Biased Regularization and Fine Tuning}},
author = {Denevi, Giulia and Pontil, Massimiliano and Ciliberto, Carlo},
booktitle = {Neural Information Processing Systems},
year = {2020},
url = {https://mlanthology.org/neurips/2020/denevi2020neurips-advantage/}
}