Efficient Thompson Sampling for Online Matrix-Factorization Recommendation

Abstract

Matrix factorization (MF) collaborative filtering is an effective and widely used method in recommendation systems. However, the problem of finding an optimal trade-off between exploration and exploitation (otherwise known as the bandit problem), a crucial problem in collaborative filtering from cold-start, has not been previously addressed.In this paper, we present a novel algorithm for online MF recommendation that automatically combines finding the most relevantitems with exploring new or less-recommended items.Our approach, called Particle Thompson Sampling for Matrix-Factorization, is based on the general Thompson sampling framework, but augmented with a novel efficient online Bayesian probabilistic matrix factorization method based on the Rao-Blackwellized particle filter.Extensive experiments in collaborative filtering using several real-world datasets demonstrate that our proposed algorithm significantly outperforms the current state-of-the-arts.

Cite

Text

Kawale et al. "Efficient Thompson Sampling for Online Matrix-Factorization Recommendation." Neural Information Processing Systems, 2015.

Markdown

[Kawale et al. "Efficient Thompson Sampling for Online Matrix-Factorization Recommendation." Neural Information Processing Systems, 2015.](https://mlanthology.org/neurips/2015/kawale2015neurips-efficient/)

BibTeX

@inproceedings{kawale2015neurips-efficient,
  title     = {{Efficient Thompson Sampling for Online Matrix-Factorization Recommendation}},
  author    = {Kawale, Jaya and Bui, Hung H and Kveton, Branislav and Tran-Thanh, Long and Chawla, Sanjay},
  booktitle = {Neural Information Processing Systems},
  year      = {2015},
  pages     = {1297-1305},
  url       = {https://mlanthology.org/neurips/2015/kawale2015neurips-efficient/}
}