On Multi-Armed Bandit Designs for Dose-Finding Trials
Abstract
We study the problem of finding the optimal dosage in early stage clinical trials through the multi-armed bandit lens. We advocate the use of the Thompson Sampling principle, a flexible algorithm that can accommodate different types of monotonicity assumptions on the toxicity and efficacy of the doses. For the simplest version of Thompson Sampling, based on a uniform prior distribution for each dose, we provide finite-time upper bounds on the number of sub-optimal dose selections, which is unprecedented for dose-finding algorithms. Through a large simulation study, we then show that variants of Thompson Sampling based on more sophisticated prior distributions outperform state-of-the-art dose identification algorithms in different types of dose-finding studies that occur in phase I or phase I/II trials.
Cite
Text
Aziz et al. "On Multi-Armed Bandit Designs for Dose-Finding Trials." Journal of Machine Learning Research, 2021.Markdown
[Aziz et al. "On Multi-Armed Bandit Designs for Dose-Finding Trials." Journal of Machine Learning Research, 2021.](https://mlanthology.org/jmlr/2021/aziz2021jmlr-multiarmed/)BibTeX
@article{aziz2021jmlr-multiarmed,
title = {{On Multi-Armed Bandit Designs for Dose-Finding Trials}},
author = {Aziz, Maryam and Kaufmann, Emilie and Riviere, Marie-Karelle},
journal = {Journal of Machine Learning Research},
year = {2021},
pages = {1-38},
volume = {22},
url = {https://mlanthology.org/jmlr/2021/aziz2021jmlr-multiarmed/}
}