Consensus Learning for Cooperative Multi-Agent Reinforcement Learning

Abstract

Almost all multi-agent reinforcement learning algorithms without communication follow the principle of centralized training with decentralized execution. During the centralized training, agents can be guided by the same signals, such as the global state. However, agents lack the shared signal and choose actions given local observations during execution. Inspired by viewpoint invariance and contrastive learning, we propose consensus learning for cooperative multi-agent reinforcement learning in this study. Although based on local observations, different agents can infer the same consensus in discrete spaces without communication. We feed the inferred one-hot consensus to the network of agents as an explicit input in a decentralized way, thereby fostering their cooperative spirit. With minor model modifications, our suggested framework can be extended to a variety of multi-agent reinforcement learning algorithms. Moreover, we carry out these variants on some fully cooperative tasks and get convincing results.

Cite

Text

Xu et al. "Consensus Learning for Cooperative Multi-Agent Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I10.26385

Markdown

[Xu et al. "Consensus Learning for Cooperative Multi-Agent Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/xu2023aaai-consensus/) doi:10.1609/AAAI.V37I10.26385

BibTeX

@inproceedings{xu2023aaai-consensus,
  title     = {{Consensus Learning for Cooperative Multi-Agent Reinforcement Learning}},
  author    = {Xu, Zhiwei and Zhang, Bin and Li, Dapeng and Zhang, Zeren and Zhou, Guangchong and Chen, Hao and Fan, Guoliang},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2023},
  pages     = {11726-11734},
  doi       = {10.1609/AAAI.V37I10.26385},
  url       = {https://mlanthology.org/aaai/2023/xu2023aaai-consensus/}
}