Bayesian Multi-Type Mean Field Multi-Agent Imitation Learning

Abstract

Multi-agent Imitation learning (MAIL) refers to the problem that agents learn to perform a task interactively in a multi-agent system through observing and mimicking expert demonstrations, without any knowledge of a reward function from the environment. MAIL has received a lot of attention due to promising results achieved on synthesized tasks, with the potential to be applied to complex real-world multi-agent tasks. Key challenges for MAIL include sample efficiency and scalability. In this paper, we proposed Bayesian multi-type mean field multi-agent imitation learning (BM3IL). Our method improves sample efficiency through establishing a Bayesian formulation for MAIL, and enhances scalability through introducing a new multi-type mean field approximation. We demonstrate the performance of our algorithm through benchmarking with three state-of-the-art multi-agent imitation learning algorithms on several tasks, including solving a multi-agent traffic optimization problem in a real-world transportation network. Experimental results indicate that our algorithm significantly outperforms all other algorithms in all scenarios.

Cite

Text

Yang et al. "Bayesian Multi-Type Mean Field Multi-Agent Imitation Learning." Neural Information Processing Systems, 2020.

Markdown

[Yang et al. "Bayesian Multi-Type Mean Field Multi-Agent Imitation Learning." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/yang2020neurips-bayesian/)

BibTeX

@inproceedings{yang2020neurips-bayesian,
  title     = {{Bayesian Multi-Type Mean Field Multi-Agent Imitation Learning}},
  author    = {Yang, Fan and Vereshchaka, Alina and Chen, Changyou and Dong, Wen},
  booktitle = {Neural Information Processing Systems},
  year      = {2020},
  url       = {https://mlanthology.org/neurips/2020/yang2020neurips-bayesian/}
}