Disposable Linear Bandits for Online Recommendations

Abstract

We study the classic stochastic linear bandit problem under the restriction that each arm may be selected for limited number of times. This simple constraint, which we call disposability, captures a common restriction that occurs in recommendation problems from a diverse array of applications ranging from personalized styling services to dating platforms. We show that the regret for this problem is characterized by a previously-unstudied function of the reward distribution among optimal arms. Algorithmically, our upper bound relies on an optimism-based policy which, while computationally intractable, lends itself to approximation via a fast alternating heuristic initialized with a classic similarity score. Experiments show that our policy dominates a set of benchmarks which includes algorithms known to be optimal for the linear bandit without disposability, along with natural modifications to these algorithms for the disposable setting.

Cite

Text

Korkut and Li. "Disposable Linear Bandits for Online Recommendations." AAAI Conference on Artificial Intelligence, 2021. doi:10.1609/AAAI.V35I5.16540

Markdown

[Korkut and Li. "Disposable Linear Bandits for Online Recommendations." AAAI Conference on Artificial Intelligence, 2021.](https://mlanthology.org/aaai/2021/korkut2021aaai-disposable/) doi:10.1609/AAAI.V35I5.16540

BibTeX

@inproceedings{korkut2021aaai-disposable,
  title     = {{Disposable Linear Bandits for Online Recommendations}},
  author    = {Korkut, Melda and Li, Andrew},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2021},
  pages     = {4172-4180},
  doi       = {10.1609/AAAI.V35I5.16540},
  url       = {https://mlanthology.org/aaai/2021/korkut2021aaai-disposable/}
}