LAGMA: LAtent Goal-Guided Multi-Agent Reinforcement Learning
Abstract
In cooperative multi-agent reinforcement learning (MARL), agents collaborate to achieve common goals, such as defeating enemies and scoring a goal. However, learning goal-reaching paths toward such a semantic goal takes a considerable amount of time in complex tasks and the trained model often fails to find such paths. To address this, we present LAtent Goal-guided Multi-Agent reinforcement learning (LAGMA), which generates a goal-reaching trajectory in latent space and provides a latent goal-guided incentive to transitions toward this reference trajectory. LAGMA consists of three major components: (a) quantized latent space constructed via a modified VQ-VAE for efficient sample utilization, (b) goal-reaching trajectory generation via extended VQ codebook, and (c) latent goal-guided intrinsic reward generation to encourage transitions towards the sampled goal-reaching path. The proposed method is evaluated by StarCraft II with both dense and sparse reward settings and Google Research Football. Empirical results show further performance improvement over state-of-the-art baselines.
Cite
Text
Na and Moon. "LAGMA: LAtent Goal-Guided Multi-Agent Reinforcement Learning." International Conference on Machine Learning, 2024.Markdown
[Na and Moon. "LAGMA: LAtent Goal-Guided Multi-Agent Reinforcement Learning." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/na2024icml-lagma/)BibTeX
@inproceedings{na2024icml-lagma,
title = {{LAGMA: LAtent Goal-Guided Multi-Agent Reinforcement Learning}},
author = {Na, Hyungho and Moon, Il-Chul},
booktitle = {International Conference on Machine Learning},
year = {2024},
pages = {37122-37140},
volume = {235},
url = {https://mlanthology.org/icml/2024/na2024icml-lagma/}
}