Hierarchical Multi-Agent Skill Discovery
Abstract
Skill discovery has shown significant progress in unsupervised reinforcement learning. This approach enables the discovery of a wide range of skills without any extrinsic reward, which can be effectively combined to tackle complex tasks. However, such unsupervised skill learning has not been well applied to multi-agent reinforcement learning (MARL) due to two primary challenges. One is how to learn skills not only for the individual agents but also for the entire team, and the other is how to coordinate the skills of different agents to accomplish multi-agent tasks. To address these challenges, we present Hierarchical Multi-Agent Skill Discovery (HMASD), a two-level hierarchical algorithm for discovering both team and individual skills in MARL. The high-level policy employs a transformer structure to realize sequential skill assignment, while the low-level policy learns to discover valuable team and individual skills. We evaluate HMASD on sparse reward multi-agent benchmarks, and the results show that HMASD achieves significant performance improvements compared to strong MARL baselines.
Cite
Text
Yang et al. "Hierarchical Multi-Agent Skill Discovery." Neural Information Processing Systems, 2023.Markdown
[Yang et al. "Hierarchical Multi-Agent Skill Discovery." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/yang2023neurips-hierarchical/)BibTeX
@inproceedings{yang2023neurips-hierarchical,
title = {{Hierarchical Multi-Agent Skill Discovery}},
author = {Yang, Mingyu and Yang, Yaodong and Lu, Zhenbo and Zhou, Wengang and Li, Houqiang},
booktitle = {Neural Information Processing Systems},
year = {2023},
url = {https://mlanthology.org/neurips/2023/yang2023neurips-hierarchical/}
}