On Multi-Armed Bandit Designs for Dose-Finding Trials

Abstract

We study the problem of finding the optimal dosage in early stage clinical trials through the multi-armed bandit lens. We advocate the use of the Thompson Sampling principle, a flexible algorithm that can accommodate different types of monotonicity assumptions on the toxicity and efficacy of the doses. For the simplest version of Thompson Sampling, based on a uniform prior distribution for each dose, we provide finite-time upper bounds on the number of sub-optimal dose selections, which is unprecedented for dose-finding algorithms. Through a large simulation study, we then show that variants of Thompson Sampling based on more sophisticated prior distributions outperform state-of-the-art dose identification algorithms in different types of dose-finding studies that occur in phase I or phase I/II trials.

Cite

Text

Aziz et al. "On Multi-Armed Bandit Designs for Dose-Finding Trials." Journal of Machine Learning Research, 2021.

Markdown

[Aziz et al. "On Multi-Armed Bandit Designs for Dose-Finding Trials." Journal of Machine Learning Research, 2021.](https://mlanthology.org/jmlr/2021/aziz2021jmlr-multiarmed/)

BibTeX

@article{aziz2021jmlr-multiarmed,
  title     = {{On Multi-Armed Bandit Designs for Dose-Finding Trials}},
  author    = {Aziz, Maryam and Kaufmann, Emilie and Riviere, Marie-Karelle},
  journal   = {Journal of Machine Learning Research},
  year      = {2021},
  pages     = {1-38},
  volume    = {22},
  url       = {https://mlanthology.org/jmlr/2021/aziz2021jmlr-multiarmed/}
}