A Tutorial on Meta-Reinforcement Learning
Abstract
While deep reinforcement learning (RL) has fueled multiple high-profile successes in machine learning, it is held back from more widespread adoption by its often poor data efficiency and the limited generality of the policies it produces. A promising approach for alleviating these limitations is to cast the development of better RL algorithms as a machine learning problem itself in a process called meta-RL. Meta-RL is most commonly studied in a problem setting where, given a distribution of tasks, the goal is to learn a policy that is capable of adapting to any new task from the task distribution with as little data as possible. In this survey, we describe the meta-RL problem setting in detail as well as its major variations. We discuss how, at a high level, meta-RL research can be clustered based on the presence of a task distribution and the learning budget available for each individual task. Using these clusters, we then survey meta-RL algorithms and applications. We conclude by presenting the open problems on the path to making meta-RL part of the standard toolbox for a deep RL practitioner.
Cite
Text
Beck et al. "A Tutorial on Meta-Reinforcement Learning." Foundations and Trends in Machine Learning, 2025. doi:10.1561/2200000080Markdown
[Beck et al. "A Tutorial on Meta-Reinforcement Learning." Foundations and Trends in Machine Learning, 2025.](https://mlanthology.org/ftml/2025/beck2025ftml-tutorial/) doi:10.1561/2200000080BibTeX
@article{beck2025ftml-tutorial,
title = {{A Tutorial on Meta-Reinforcement Learning}},
author = {Beck, Jacob and Vuorio, Risto and Liu, Evan Zheran and Xiong, Zheng and Zintgraf, Luisa M. and Finn, Chelsea and Whiteson, Shimon},
journal = {Foundations and Trends in Machine Learning},
year = {2025},
pages = {224-384},
doi = {10.1561/2200000080},
volume = {18},
url = {https://mlanthology.org/ftml/2025/beck2025ftml-tutorial/}
}