Greedy Algorithms for Structured Bandits: A Sharp Characterization of Asymptotic Success / Failure

Abstract

We study the greedy (exploitation-only) algorithm in bandit problems with a known reward structure. We allow arbitrary finite reward structures, while prior work focused on a few specific ones. We fully characterize when the greedy algorithm asymptotically succeeds or fails, in the sense of sublinear vs. linear regret as a function of time. Our characterization identifies a partial identifiability property of the problem instance as the necessary and sufficient condition for the asymptotic success. Notably, once this property holds, the problem becomes easy—\emph{any} algorithm will succeed (in the same sense as above), provided it satisfies a mild non-degeneracy condition. Our characterization extends to contextual bandits and interactive decision-making with arbitrary feedback. Examples demonstrating broad applicability and extensions to infinite reward structures are provided.

Cite

Text

Slivkins et al. "Greedy Algorithms for Structured Bandits: A Sharp Characterization of Asymptotic Success / Failure." Advances in Neural Information Processing Systems, 2025.

Markdown

[Slivkins et al. "Greedy Algorithms for Structured Bandits: A Sharp Characterization of Asymptotic Success / Failure." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/slivkins2025neurips-greedy/)

BibTeX

@inproceedings{slivkins2025neurips-greedy,
  title     = {{Greedy Algorithms for Structured Bandits: A Sharp Characterization of Asymptotic Success / Failure}},
  author    = {Slivkins, Aleksandrs and Xu, Yunzong and Zuo, Shiliang},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/slivkins2025neurips-greedy/}
}