Model Selection for Contextual Bandits
Abstract
We introduce the problem of model selection for contextual bandits, where a learner must adapt to the complexity of the optimal policy while balancing exploration and exploitation. Our main result is a new model selection guarantee for linear contextual bandits. We work in the stochastic realizable setting with a sequence of nested linear policy classes of dimension $d_1 < d_2 < \ldots$, where the $m^\star$-th class contains the optimal policy, and we design an algorithm that achieves $\tilde{O}l(T^{2/3}d^{1/3}_{m^\star})$ regret with no prior knowledge of the optimal dimension $d_{m^\star}$. The algorithm also achieves regret $\tilde{O}(T^{3/4} + \sqrt{Td_{m^\star}})$, which is optimal for $d_{m^{\star}}\geq{}\sqrt{T}$. This is the first model selection result for contextual bandits with non-vacuous regret for all values of $d_{m^\star}$, and to the best of our knowledge is the first positive result of this type for any online learning setting with partial information. The core of the algorithm is a new estimator for the gap in the best loss achievable by two linear policy classes, which we show admits a convergence rate faster than the rate required to learn the parameters for either class.
Cite
Text
Foster et al. "Model Selection for Contextual Bandits." Neural Information Processing Systems, 2019.Markdown
[Foster et al. "Model Selection for Contextual Bandits." Neural Information Processing Systems, 2019.](https://mlanthology.org/neurips/2019/foster2019neurips-model/)BibTeX
@inproceedings{foster2019neurips-model,
title = {{Model Selection for Contextual Bandits}},
author = {Foster, Dylan J and Krishnamurthy, Akshay and Luo, Haipeng},
booktitle = {Neural Information Processing Systems},
year = {2019},
pages = {14741-14752},
url = {https://mlanthology.org/neurips/2019/foster2019neurips-model/}
}