Learning to Modulate Pre-Trained Models in RL

Abstract

Reinforcement Learning (RL) has experienced great success in complex games and simulations. However, RL agents are often highly specialized for a particular task, and it is difficult to adapt a trained agent to a new task. In supervised learning, an established paradigm is multi-task pre-training followed by fine-tuning. A similar trend is emerging in RL, where agents are pre-trained on data collections that comprise a multitude of tasks. Despite these developments, it remains an open challenge how to adapt such pre-trained agents to novel tasks while retaining performance on the pre-training tasks. In this regard, we pre-train an agent on a set of tasks from the Meta-World benchmark suite and adapt it to tasks from Continual-World. We conduct a comprehensive comparison of fine-tuning methods originating from supervised learning in our setup. Our findings show that fine-tuning is feasible, but for existing methods, performance on previously learned tasks often deteriorates. Therefore, we propose a novel approach that avoids forgetting by modulating the information flow of the pre-trained model. Our method outperforms existing fine-tuning approaches, and achieves state-of-the-art performance on the Continual-World benchmark. To facilitate future research in this direction, we collect datasets for all Meta-World tasks and make them publicly available.

Cite

Text

Schmied et al. "Learning to Modulate Pre-Trained Models in RL." ICLR 2023 Workshops: RRL, 2023.

Markdown

[Schmied et al. "Learning to Modulate Pre-Trained Models in RL." ICLR 2023 Workshops: RRL, 2023.](https://mlanthology.org/iclrw/2023/schmied2023iclrw-learning/)

BibTeX

@inproceedings{schmied2023iclrw-learning,
  title     = {{Learning to Modulate Pre-Trained Models in RL}},
  author    = {Schmied, Thomas and Hofmarcher, Markus and Paischer, Fabian and Pascanu, Razvan and Hochreiter, Sepp},
  booktitle = {ICLR 2023 Workshops: RRL},
  year      = {2023},
  url       = {https://mlanthology.org/iclrw/2023/schmied2023iclrw-learning/}
}