Optimal Regret Analysis of Thompson Sampling in Stochastic Multi-Armed Bandit Problem with Multiple Plays

Abstract

We discuss a multiple-play multi-armed bandit (MAB) problem in which several arms are selected at each round. Recently, Thompson sampling (TS), a randomized algorithm with a Bayesian spirit, has attracted much attention for its empirically excellent performance, and it is revealed to have an optimal regret bound in the standard single-play MAB problem. In this paper, we propose the multiple-play Thompson sampling (MP-TS) algorithm, an extension of TS to the multiple-play MAB problem, and discuss its regret analysis. We prove that MP-TS has the optimal regret upper bound that matches the regret lower bound provided by Anantharam et al.\,(1987). Therefore, MP-TS is the first computationally efficient algorithm with optimal regret. A set of computer simulations was also conducted, which compared MP-TS with state-of-the-art algorithms. We also propose a modification of MP-TS, which is shown to have better empirical performance.

Cite

Text

Komiyama et al. "Optimal Regret Analysis of Thompson Sampling in Stochastic Multi-Armed Bandit Problem with Multiple Plays." International Conference on Machine Learning, 2015.

Markdown

[Komiyama et al. "Optimal Regret Analysis of Thompson Sampling in Stochastic Multi-Armed Bandit Problem with Multiple Plays." International Conference on Machine Learning, 2015.](https://mlanthology.org/icml/2015/komiyama2015icml-optimal/)

BibTeX

@inproceedings{komiyama2015icml-optimal,
  title     = {{Optimal Regret Analysis of Thompson Sampling in Stochastic Multi-Armed Bandit Problem with Multiple Plays}},
  author    = {Komiyama, Junpei and Honda, Junya and Nakagawa, Hiroshi},
  booktitle = {International Conference on Machine Learning},
  year      = {2015},
  pages     = {1152-1161},
  volume    = {37},
  url       = {https://mlanthology.org/icml/2015/komiyama2015icml-optimal/}
}