Bayesian Reinforcement Learning: A Survey
Abstract
Bayesian methods for machine learning have been widely investigated, yielding principled methods for incorporating prior information into inference algorithms. In this survey, we provide an in-depth review of the role of Bayesian methods for the reinforcement learning (RL) paradigm. The major incentives for incorporating Bayesian reasoning in RL are: 1) it provides an elegant approach to action-selection (exploration/exploitation) as a function of the uncertainty in learning; and 2) it provides a machinery to incorporate prior knowledge into the algorithms. We first discuss models and methods for Bayesian inference in the simple single-step Bandit model. We then review the extensive recent literature on Bayesian methods for model-based RL, where prior information can be expressed on the parameters of the Markov model. We also present Bayesian methods for model-free RL, where priors are expressed over the value function or policy class. The objective of the paper is to provide a comprehensive survey on Bayesian RL algorithms and their theoretical and empirical properties.
Cite
Text
Ghavamzadeh et al. "Bayesian Reinforcement Learning: A Survey." Foundations and Trends in Machine Learning, 2015. doi:10.1561/2200000049Markdown
[Ghavamzadeh et al. "Bayesian Reinforcement Learning: A Survey." Foundations and Trends in Machine Learning, 2015.](https://mlanthology.org/ftml/2015/ghavamzadeh2015ftml-bayesian/) doi:10.1561/2200000049BibTeX
@article{ghavamzadeh2015ftml-bayesian,
title = {{Bayesian Reinforcement Learning: A Survey}},
author = {Ghavamzadeh, Mohammad and Mannor, Shie and Pineau, Joelle and Tamar, Aviv},
journal = {Foundations and Trends in Machine Learning},
year = {2015},
pages = {359-483},
doi = {10.1561/2200000049},
volume = {8},
url = {https://mlanthology.org/ftml/2015/ghavamzadeh2015ftml-bayesian/}
}