Multi-Agent Generative Adversarial Imitation Learning
Abstract
Imitation learning algorithms can be used to learn a policy from expert demonstrations without access to a reward signal. However, most existing approaches are not applicable in multi-agent settings due to the existence of multiple (Nash) equilibria and non-stationary environments. We propose a new framework for multi-agent imitation learning for general Markov games, where we build upon a generalized notion of inverse reinforcement learning. We further introduce a practical multi-agent actor-critic algorithm with good empirical performance. Our method can be used to imitate complex behaviors in high-dimensional environments with multiple cooperative or competing agents.
Cite
Text
Song et al. "Multi-Agent Generative Adversarial Imitation Learning." Neural Information Processing Systems, 2018.Markdown
[Song et al. "Multi-Agent Generative Adversarial Imitation Learning." Neural Information Processing Systems, 2018.](https://mlanthology.org/neurips/2018/song2018neurips-multiagent/)BibTeX
@inproceedings{song2018neurips-multiagent,
title = {{Multi-Agent Generative Adversarial Imitation Learning}},
author = {Song, Jiaming and Ren, Hongyu and Sadigh, Dorsa and Ermon, Stefano},
booktitle = {Neural Information Processing Systems},
year = {2018},
pages = {7461-7472},
url = {https://mlanthology.org/neurips/2018/song2018neurips-multiagent/}
}