Meta-Learning with Neural Bandit Scheduler

Abstract

Meta-learning has been proven an effective learning paradigm for training machine learning models with good generalization ability. Apart from the common practice of uniformly sampling the meta-training tasks, existing methods working on task scheduling strategies are mainly based on pre-defined sampling protocols or the assumed task-model correlations, and greedily make scheduling decisions, which can lead to sub-optimal performance bottlenecks of the meta-model. In this paper, we propose a novel task scheduling framework under Contextual Bandits settings, named BASS, which directly optimizes the task scheduling strategy based on the status of the meta-model. By balancing the exploitation and exploration in meta-learning task scheduling, BASS can help tackle the challenge of limited knowledge about the task distribution during the early stage of meta-training, while simultaneously exploring potential benefits for forthcoming meta-training iterations through an adaptive exploration strategy. Theoretical analysis and extensive experiments are presented to show the effectiveness of our proposed framework.

Cite

Text

Qi et al. "Meta-Learning with Neural Bandit Scheduler." Neural Information Processing Systems, 2023.

Markdown

[Qi et al. "Meta-Learning with Neural Bandit Scheduler." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/qi2023neurips-metalearning/)

BibTeX

@inproceedings{qi2023neurips-metalearning,
  title     = {{Meta-Learning with Neural Bandit Scheduler}},
  author    = {Qi, Yunzhe and Ban, Yikun and Wei, Tianxin and Zou, Jiaru and Yao, Huaxiu and He, Jingrui},
  booktitle = {Neural Information Processing Systems},
  year      = {2023},
  url       = {https://mlanthology.org/neurips/2023/qi2023neurips-metalearning/}
}