Near-Optimal Confidence Sequences for Bounded Random Variables

Abstract

Many inference problems, such as sequential decision problems like A/B testing, adaptive sampling schemes like bandit selection, are often online in nature. The fundamental problem for online inference is to provide a sequence of confidence intervals that are valid uniformly over the growing-into-infinity sample sizes. To address this question, we provide a near-optimal confidence sequence for bounded random variables by utilizing Bentkus’ concentration results. We show that it improves on the existing approaches that use the Cram{é}r-Chernoff technique such as the Hoeffding, Bernstein, and Bennett inequalities. The resulting confidence sequence is confirmed to be favorable in synthetic coverage problems, adaptive stopping algorithms, and multi-armed bandit problems.

Cite

Text

Kuchibhotla and Zheng. "Near-Optimal Confidence Sequences for Bounded Random Variables." International Conference on Machine Learning, 2021.

Markdown

[Kuchibhotla and Zheng. "Near-Optimal Confidence Sequences for Bounded Random Variables." International Conference on Machine Learning, 2021.](https://mlanthology.org/icml/2021/kuchibhotla2021icml-nearoptimal/)

BibTeX

@inproceedings{kuchibhotla2021icml-nearoptimal,
  title     = {{Near-Optimal Confidence Sequences for Bounded Random Variables}},
  author    = {Kuchibhotla, Arun K and Zheng, Qinqing},
  booktitle = {International Conference on Machine Learning},
  year      = {2021},
  pages     = {5827-5837},
  volume    = {139},
  url       = {https://mlanthology.org/icml/2021/kuchibhotla2021icml-nearoptimal/}
}