Beyond Central Limit Theorem for Higher-Order Inference in Batched Bandits
Abstract
Adaptive experiments have been gaining traction in a variety of domains, which stimulates a growing literature focusing on post-experimental statistical inference on data collected from such designs. Prior work constructs confidence intervals mainly based on two types of methods: (i) martingale concentration inequalities and (ii) asymptotic approximation to distribution of test statistics; this work contributes to the second kind. The current asymptotic approximation methods however mostly rely on first-order limit theorems, which can have a slow convergence in a data-poor regime. Besides, established results often rely on conditions that noises behave well, which can be problematic when the real-world instances are heavy-tailed or asymmetric. In this paper, we propose a higher-order asymptotic expansion formula for inference on adaptively collected data, which generalizes normal approximation to the distribution of standard test statistics. Our theorem relaxes assumptions on the noise distribution and benefits a higher-order approximation in the distributional distance to accommodate small sample sizes. We complement our results by promising empirical performances in simulations.
Cite
Text
Park et al. "Beyond Central Limit Theorem for Higher-Order Inference in Batched Bandits." NeurIPS 2022 Workshops: CML4Impact, 2022.Markdown
[Park et al. "Beyond Central Limit Theorem for Higher-Order Inference in Batched Bandits." NeurIPS 2022 Workshops: CML4Impact, 2022.](https://mlanthology.org/neuripsw/2022/park2022neuripsw-beyond/)BibTeX
@inproceedings{park2022neuripsw-beyond,
title = {{Beyond Central Limit Theorem for Higher-Order Inference in Batched Bandits}},
author = {Park, Yechan and Zhan, Ruohan and Yoshida, Nakahiro},
booktitle = {NeurIPS 2022 Workshops: CML4Impact},
year = {2022},
url = {https://mlanthology.org/neuripsw/2022/park2022neuripsw-beyond/}
}