On Context-Dependent Clustering of Bandits

Abstract

We investigate a novel cluster-of-bandit algorithm CAB for collaborative recommendation tasks that implements the underlying feedback sharing mechanism by estimating user neighborhoods in a context-dependent manner. CAB makes sharp departures from the state of the art by incorporating collaborative effects into inference, as well as learning processes in a manner that seamlessly interleaves explore-exploit tradeoffs and collaborative steps. We prove regret bounds for CAB under various data-dependent assumptions which exhibit a crisp dependence on the expected number of clusters over the users, a natural measure of the statistical difficulty of the learning task. Experiments on production and real-world datasets show that CAB offers significantly increased prediction performance against a representative pool of state-of-the-art methods.

Cite

Text

Gentile et al. "On Context-Dependent Clustering of Bandits." International Conference on Machine Learning, 2017.

Markdown

[Gentile et al. "On Context-Dependent Clustering of Bandits." International Conference on Machine Learning, 2017.](https://mlanthology.org/icml/2017/gentile2017icml-contextdependent/)

BibTeX

@inproceedings{gentile2017icml-contextdependent,
  title     = {{On Context-Dependent Clustering of Bandits}},
  author    = {Gentile, Claudio and Li, Shuai and Kar, Purushottam and Karatzoglou, Alexandros and Zappella, Giovanni and Etrue, Evans},
  booktitle = {International Conference on Machine Learning},
  year      = {2017},
  pages     = {1253-1262},
  volume    = {70},
  url       = {https://mlanthology.org/icml/2017/gentile2017icml-contextdependent/}
}