Bandit Algorithms Based on Thompson Sampling for Bounded Reward Distributions

Abstract

We focus on a classic reinforcement learning problem, called a multi-armed bandit, and more specifically in the stochastic setting with reward distributions bounded in $[0,1]$. For this model, an optimal problem-dependent asymptotic regret lower bound has been derived. However, the existing algorithms achieving this regret lower bound all require to solve an optimization problem at each step, inducing a large complexity. In this paper, we propose two new algorithms, which we prove to achieve the problem-dependent asymptotic regret lower bound. The first one, which we call Multinomial TS, is an adaptation of Thompson Sampling for Bernoulli rewards to multinomial reward distributions whose support is included in $\{0, \frac{1}{M}, …, 1\}$. This algorithm achieves the regret lower bound in the case of multinomial distributions with the aforementioned support, and it can be easily generalized to bounded reward distributions in $[0, 1]$ by randomly rounding the observed rewards. The second algorithm we introduce, which we call Non-parametric TS, is a randomized algorithm but not based on the posterior sampling in the strict sense. At each step, it computes an average of the observed rewards with random weight. Not only is it asymptotically optimal, but also it performs very well even for small horizons.

Cite

Text

Riou and Honda. "Bandit Algorithms Based on Thompson Sampling for Bounded Reward Distributions." Proceedings of the 31st International Conference  on Algorithmic Learning Theory, 2020.

Markdown

[Riou and Honda. "Bandit Algorithms Based on Thompson Sampling for Bounded Reward Distributions." Proceedings of the 31st International Conference  on Algorithmic Learning Theory, 2020.](https://mlanthology.org/alt/2020/riou2020alt-bandit/)

BibTeX

@inproceedings{riou2020alt-bandit,
  title     = {{Bandit Algorithms Based on Thompson Sampling for Bounded Reward Distributions}},
  author    = {Riou, Charles and Honda, Junya},
  booktitle = {Proceedings of the 31st International Conference  on Algorithmic Learning Theory},
  year      = {2020},
  pages     = {777-826},
  volume    = {117},
  url       = {https://mlanthology.org/alt/2020/riou2020alt-bandit/}
}