Planning Contextual Adaptive Experiments with Model Predictive Control

Abstract

Implementing adaptive experimentation methods in the real world often encounters a multitude of operational difficulties, including batched/delayed feedback, non-stationary environments, and constraints on treatment allocations. To improve the flexibility of adaptive experimentation, we propose a Bayesian, optimization-based framework founded on model-predictive control (MPC) for the linear contextual bandit setting. While we focus on simple regret minimization, the framework can flexibly incorporate multiple objectives along with constraints, batches, personalized and non-personalized policies, as well as predictions of future context arrivals. Most importantly, it maintains this flexibility while guaranteeing improvement over non-adaptive A/B testing across all time horizons, and empirically outperforms standard policies such as Thompson Sampling. Overall, this framework offers a way to guide adaptive designs across the varied demands of modern large-scale experiments.

Cite

Text

Che et al. "Planning Contextual Adaptive Experiments with Model Predictive Control." NeurIPS 2023 Workshops: ReALML, 2023.

Markdown

[Che et al. "Planning Contextual Adaptive Experiments with Model Predictive Control." NeurIPS 2023 Workshops: ReALML, 2023.](https://mlanthology.org/neuripsw/2023/che2023neuripsw-planning/)

BibTeX

@inproceedings{che2023neuripsw-planning,
  title     = {{Planning Contextual Adaptive Experiments with Model Predictive Control}},
  author    = {Che, Ethan and Wang, Jimmy and Namkoong, Hongseok},
  booktitle = {NeurIPS 2023 Workshops: ReALML},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/che2023neuripsw-planning/}
}