Provably Efficient Multi-Task Meta Bandit Learning via Shared Representations

Abstract

Learning-to-learn or meta-learning focuses on developing algorithms that leverage prior experience to quickly acquire new skills or adapt to novel environments. A crucial component of meta-learning is representation learning, which aims to construct data representations capable of transferring knowledge across multiple tasks—a critical advantage in data-scarce settings. We study how representation learning can improve the efficiency of bandit problems. We consider $T$ $d$-dimensional linear bandits that share a common low-dimensional linear representation. We provide provably fast, sample-efficient algorithms to address the two key problems in meta-learning: (1) learning a common set of features from multiple related bandit tasks and (2) transferring this knowledge to new, unseen bandit tasks. We validated the theoretical results through numerical experiments using real-world and synthetic datasets, comparing them against benchmark algorithms.

Cite

Text

Lin and Moothedath. "Provably Efficient Multi-Task Meta Bandit Learning via Shared Representations." Advances in Neural Information Processing Systems, 2025.

Markdown

[Lin and Moothedath. "Provably Efficient Multi-Task Meta Bandit Learning via Shared Representations." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/lin2025neurips-provably/)

BibTeX

@inproceedings{lin2025neurips-provably,
  title     = {{Provably Efficient Multi-Task Meta Bandit Learning via Shared Representations}},
  author    = {Lin, Jiabin and Moothedath, Shana},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/lin2025neurips-provably/}
}