Bayesian Reinforcement Learning: A Survey

Abstract

Bayesian methods for machine learning have been widely investigated, yielding principled methods for incorporating prior information into inference algorithms. In this survey, we provide an in-depth review of the role of Bayesian methods for the reinforcement learning (RL) paradigm. The major incentives for incorporating Bayesian reasoning in RL are: 1) it provides an elegant approach to action-selection (exploration/exploitation) as a function of the uncertainty in learning; and 2) it provides a machinery to incorporate prior knowledge into the algorithms. We first discuss models and methods for Bayesian inference in the simple single-step Bandit model. We then review the extensive recent literature on Bayesian methods for model-based RL, where prior information can be expressed on the parameters of the Markov model. We also present Bayesian methods for model-free RL, where priors are expressed over the value function or policy class. The objective of the paper is to provide a comprehensive survey on Bayesian RL algorithms and their theoretical and empirical properties.

Cite

Text

Ghavamzadeh et al. "Bayesian Reinforcement Learning: A Survey." Foundations and Trends in Machine Learning, 2015. doi:10.1561/2200000049

Markdown

[Ghavamzadeh et al. "Bayesian Reinforcement Learning: A Survey." Foundations and Trends in Machine Learning, 2015.](https://mlanthology.org/ftml/2015/ghavamzadeh2015ftml-bayesian/) doi:10.1561/2200000049

BibTeX

@article{ghavamzadeh2015ftml-bayesian,
  title     = {{Bayesian Reinforcement Learning: A Survey}},
  author    = {Ghavamzadeh, Mohammad and Mannor, Shie and Pineau, Joelle and Tamar, Aviv},
  journal   = {Foundations and Trends in Machine Learning},
  year      = {2015},
  pages     = {359-483},
  doi       = {10.1561/2200000049},
  volume    = {8},
  url       = {https://mlanthology.org/ftml/2015/ghavamzadeh2015ftml-bayesian/}
}