Model-Agnostic Meta-Learning with Open-Ended Reinforcement Learning

Abstract

This paper is an in-progress research that builds on the Open-Ended Reinforcement Learning with Neural Reward Functions proposed by Meier and Mujika [1] which use reward functions encoded by neural networks. One key limitation of their paper is the necessity of re-learning for each new skill learned by the agent. Consequently, we propose integrating meta-learning algorithms to tackle this problem. We, therefore, study the use of MAML, Model-Agnostic Meta Learning that we believe could make policy learning more efficient. MAML operates by learning an initialization of the model parameters that can be fine-tuned with a small number of examples from a new task which allows for rapid adaptation to new tasks.

Cite

Text

Shabbar. "Model-Agnostic Meta-Learning with Open-Ended Reinforcement Learning." NeurIPS 2024 Workshops: IMOL, 2024.

Markdown

[Shabbar. "Model-Agnostic Meta-Learning with Open-Ended Reinforcement Learning." NeurIPS 2024 Workshops: IMOL, 2024.](https://mlanthology.org/neuripsw/2024/shabbar2024neuripsw-modelagnostic/)

BibTeX

@inproceedings{shabbar2024neuripsw-modelagnostic,
  title     = {{Model-Agnostic Meta-Learning with Open-Ended Reinforcement Learning}},
  author    = {Shabbar, Aya},
  booktitle = {NeurIPS 2024 Workshops: IMOL},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/shabbar2024neuripsw-modelagnostic/}
}