Selectively Sharing Experiences Improves Multi-Agent Reinforcement Learning
Abstract
We present a novel multi-agent RL approach, Selective Multi-Agent PER, in which agents share with other agents a limited number of transitions they observe during training. They follow a similar heuristic as is used in (single-agent) Prioritized Experience Replay, and choose those transitions based on their td-error. The intuition behind this is that even a small number of relevant experiences from other agents could help each agent learn. Unlike many other multi-agent RL algorithms, this approach allows for largely decentralized training, requiring only a limited communication channel between agents. We show that our approach outperforms baseline no-sharing decentralized training and state-of-the art multi-agent RL algorithms. Further, sharing only a small number of experiences outperforms sharing all experiences between agents, and the performance uplift from selective experience sharing is robust across a range of hyperparameters and DQN variants.
Cite
Text
Gerstgrasser et al. "Selectively Sharing Experiences Improves Multi-Agent Reinforcement Learning." NeurIPS 2022 Workshops: DeepRL, 2022.Markdown
[Gerstgrasser et al. "Selectively Sharing Experiences Improves Multi-Agent Reinforcement Learning." NeurIPS 2022 Workshops: DeepRL, 2022.](https://mlanthology.org/neuripsw/2022/gerstgrasser2022neuripsw-selectively/)BibTeX
@inproceedings{gerstgrasser2022neuripsw-selectively,
title = {{Selectively Sharing Experiences Improves Multi-Agent Reinforcement Learning}},
author = {Gerstgrasser, Matthias and Danino, Tom and Keren, Sarah},
booktitle = {NeurIPS 2022 Workshops: DeepRL},
year = {2022},
url = {https://mlanthology.org/neuripsw/2022/gerstgrasser2022neuripsw-selectively/}
}