Bootstrapped Meta-Learning

Abstract

Meta-learning empowers artificial intelligence to increase its efficiency by learning how to learn. Unlocking this potential involves overcoming a challenging meta-optimisation problem. We propose an algorithm that tackles this problem by letting the meta-learner teach itself. The algorithm first bootstraps a target from the meta-learner, then optimises the meta-learner by minimising the distance to that target under a chosen (pseudo-)metric. Focusing on meta-learning with gradients, we establish conditions that guarantee performance improvements and show that metric can be used to control meta-optimisation. Meanwhile, the bootstrapping mechanism can extend the effective meta-learning horizon without requiring backpropagation through all updates. We achieve a new state-of-the art for model-free agents on the Atari ALE benchmark and demonstrate that it yields both performance and efficiency gains in multi-task meta-learning. Finally, we explore how bootstrapping opens up new possibilities and find that it can meta-learn efficient exploration in an epsilon-greedy Q-learning agent - without backpropagating through the update rule.

Cite

Text

Flennerhag et al. "Bootstrapped Meta-Learning." International Conference on Learning Representations, 2022.

Markdown

[Flennerhag et al. "Bootstrapped Meta-Learning." International Conference on Learning Representations, 2022.](https://mlanthology.org/iclr/2022/flennerhag2022iclr-bootstrapped/)

BibTeX

@inproceedings{flennerhag2022iclr-bootstrapped,
  title     = {{Bootstrapped Meta-Learning}},
  author    = {Flennerhag, Sebastian and Schroecker, Yannick and Zahavy, Tom and van Hasselt, Hado and Silver, David and Singh, Satinder},
  booktitle = {International Conference on Learning Representations},
  year      = {2022},
  url       = {https://mlanthology.org/iclr/2022/flennerhag2022iclr-bootstrapped/}
}