Bayesian Reinforcement Learning via Deep, Sparse Sampling

Abstract

We address the problem of Bayesian reinforcement learning using efficient model-based online planning. We propose an optimism-free Bayes-adaptive algorithm to induce deeper and sparser exploration with a theoretical bound on its performance relative to the Bayes optimal as well as lower computational complexity. The main novelty is the use of a candidate policy generator, to generate long-term options in the planning tree (over beliefs), which allows us to create much sparser and deeper trees. Experimental results on different environments show that in comparison to the state-of-the-art, our algorithm is both computationally more efficient, and obtains significantly higher reward over time in discrete environments.

Cite

Text

Grover et al. "Bayesian Reinforcement Learning via Deep, Sparse Sampling." Artificial Intelligence and Statistics, 2020.

Markdown

[Grover et al. "Bayesian Reinforcement Learning via Deep, Sparse Sampling." Artificial Intelligence and Statistics, 2020.](https://mlanthology.org/aistats/2020/grover2020aistats-bayesian/)

BibTeX

@inproceedings{grover2020aistats-bayesian,
  title     = {{Bayesian Reinforcement Learning via Deep, Sparse Sampling}},
  author    = {Grover, Divya and Basu, Debabrota and Dimitrakakis, Christos},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2020},
  pages     = {3036-3045},
  volume    = {108},
  url       = {https://mlanthology.org/aistats/2020/grover2020aistats-bayesian/}
}