A Game Theoretic Framework for Model Based Reinforcement Learning

Abstract

Designing stable and efficient algorithms for model-based reinforcement learning (MBRL) with function approximation has remained challenging despite growing interest in the field. To help expose the practical challenges in MBRL and simplify algorithm design from the lens of abstraction, we develop a new framework that casts MBRL as a game between: (1) a policy player, which attempts to maximize rewards under the learned model; (2) a model player, which attempts to fit the real-world data collected by the policy player. We show that a near-optimal policy for the environment can be obtained by finding an approximate equilibrium for aforementioned game, and we develop two families of algorithms to find the game equilibrium by drawing upon ideas from Stackelberg games. Experimental studies suggest that the proposed algorithms achieve state of the art sample efficiency, match the asymptotic performance of model-free policy gradient, and scale gracefully to high-dimensional tasks like dexterous hand manipulation. Project page: \url{https://sites.google.com/view/mbrl-game}.

Cite

Text

Rajeswaran et al. "A Game Theoretic Framework for Model Based Reinforcement Learning." International Conference on Machine Learning, 2020.

Markdown

[Rajeswaran et al. "A Game Theoretic Framework for Model Based Reinforcement Learning." International Conference on Machine Learning, 2020.](https://mlanthology.org/icml/2020/rajeswaran2020icml-game/)

BibTeX

@inproceedings{rajeswaran2020icml-game,
  title     = {{A Game Theoretic Framework for Model Based Reinforcement Learning}},
  author    = {Rajeswaran, Aravind and Mordatch, Igor and Kumar, Vikash},
  booktitle = {International Conference on Machine Learning},
  year      = {2020},
  pages     = {7953-7963},
  volume    = {119},
  url       = {https://mlanthology.org/icml/2020/rajeswaran2020icml-game/}
}