Boosted Curriculum Reinforcement Learning

Abstract

Curriculum value-based reinforcement learning (RL) solves a complex target task by reusing action-values across a tailored sequence of related tasks of increasing difficulty. However, finding an exact way of reusing action-values in this setting is still a poorly understood problem. In this paper, we introduce the concept of boosting to curriculum value-based RL, by approximating the action-value function as a sum of residuals trained on each task. This approach, which we refer to as boosted curriculum reinforcement learning (BCRL), has the benefit of naturally increasing the representativeness of the functional space by adding a new residual each time a new task is presented. This procedure allows reusing previous action-values while promoting expressiveness of the action-value function. We theoretically study BCRL as an approximate value iteration algorithm, discussing advantages over regular curriculum RL in terms of approximation accuracy and convergence to the optimal action-value function. Finally, we provide detailed empirical evidence of the benefits of BCRL in problems requiring curricula for accurate action-value estimation and targeted exploration.

Cite

Text

Klink et al. "Boosted Curriculum Reinforcement Learning." International Conference on Learning Representations, 2022.

Markdown

[Klink et al. "Boosted Curriculum Reinforcement Learning." International Conference on Learning Representations, 2022.](https://mlanthology.org/iclr/2022/klink2022iclr-boosted/)

BibTeX

@inproceedings{klink2022iclr-boosted,
  title     = {{Boosted Curriculum Reinforcement Learning}},
  author    = {Klink, Pascal and D'Eramo, Carlo and Peters, Jan and Pajarinen, Joni},
  booktitle = {International Conference on Learning Representations},
  year      = {2022},
  url       = {https://mlanthology.org/iclr/2022/klink2022iclr-boosted/}
}