Deep Coordination Graphs
Abstract
This paper introduces the deep coordination graph (DCG) for collaborative multi-agent reinforcement learning. DCG strikes a flexible trade-off between representational capacity and generalization by factoring the joint value function of all agents according to a coordination graph into payoffs between pairs of agents. The value can be maximized by local message passing along the graph, which allows training of the value function end-to-end with Q-learning. Payoff functions are approximated with deep neural networks that employ parameter sharing and low-rank approximations to significantly improve sample efficiency. We show that DCG can solve predator-prey tasks that highlight the relative overgeneralization pathology, as well as challenging StarCraft II micromanagement tasks.
Cite
Text
Boehmer et al. "Deep Coordination Graphs." International Conference on Machine Learning, 2020.Markdown
[Boehmer et al. "Deep Coordination Graphs." International Conference on Machine Learning, 2020.](https://mlanthology.org/icml/2020/boehmer2020icml-deep/)BibTeX
@inproceedings{boehmer2020icml-deep,
title = {{Deep Coordination Graphs}},
author = {Boehmer, Wendelin and Kurin, Vitaly and Whiteson, Shimon},
booktitle = {International Conference on Machine Learning},
year = {2020},
pages = {980-991},
volume = {119},
url = {https://mlanthology.org/icml/2020/boehmer2020icml-deep/}
}