Bayesian Exploration for Lifelong Reinforcement Learning

Abstract

A central question in reinforcement learning (RL) is how to leverage prior knowledge to accelerate learning in new tasks. We propose a Bayesian exploration method for lifelong reinforcement learning (BLRL) that aims to learn a Bayesian posterior that distills the common structure shared across different tasks. We further derive a sample complexity analysis of BLRL in the finite MDP setting. To scale our approach, we propose a variational Bayesian Lifelong Learning (VBLRL) algorithm that is based on Bayesian neural networks, can be combined with recent model-based RL methods, and exhibits backward transfer. Experimental results on three challenging domains show that our algorithms adapt to new tasks faster than state-of-the-art lifelong RL methods.

Cite

Text

Fu et al. "Bayesian Exploration for Lifelong Reinforcement Learning." NeurIPS 2021 Workshops: DeepRL, 2021.

Markdown

[Fu et al. "Bayesian Exploration for Lifelong Reinforcement Learning." NeurIPS 2021 Workshops: DeepRL, 2021.](https://mlanthology.org/neuripsw/2021/fu2021neuripsw-bayesian/)

BibTeX

@inproceedings{fu2021neuripsw-bayesian,
  title     = {{Bayesian Exploration for Lifelong Reinforcement Learning}},
  author    = {Fu, Haotian and Yu, Shangqun and Littman, Michael and Konidaris, George},
  booktitle = {NeurIPS 2021 Workshops: DeepRL},
  year      = {2021},
  url       = {https://mlanthology.org/neuripsw/2021/fu2021neuripsw-bayesian/}
}