Local Optimization Achieves Global Optimality in Multi-Agent Reinforcement Learning
Abstract
Policy optimization methods with function approximation are widely used in multi-agent reinforcement learning. However, it remains elusive how to design such algorithms with statistical guarantees. Leveraging a multi-agent performance difference lemma that characterizes the landscape of multi-agent policy optimization, we find that the localized action value function serves as an ideal descent direction for each local policy. Motivated by the observation, we present a multi-agent PPO algorithm in which the local policy of each agent is updated similarly to vanilla PPO. We prove that with standard regularity conditions on the Markov game and problem-dependent quantities, our algorithm converges to the globally optimal policy at a sublinear rate. We extend our algorithm to the off-policy setting and introduce pessimism to policy evaluation, which aligns with experiments. To our knowledge, this is the first provably convergent multi-agent PPO algorithm in cooperative Markov games.
Cite
Text
Zhao et al. "Local Optimization Achieves Global Optimality in Multi-Agent Reinforcement Learning." International Conference on Machine Learning, 2023.Markdown
[Zhao et al. "Local Optimization Achieves Global Optimality in Multi-Agent Reinforcement Learning." International Conference on Machine Learning, 2023.](https://mlanthology.org/icml/2023/zhao2023icml-local/)BibTeX
@inproceedings{zhao2023icml-local,
title = {{Local Optimization Achieves Global Optimality in Multi-Agent Reinforcement Learning}},
author = {Zhao, Yulai and Yang, Zhuoran and Wang, Zhaoran and Lee, Jason D.},
booktitle = {International Conference on Machine Learning},
year = {2023},
pages = {42200-42226},
volume = {202},
url = {https://mlanthology.org/icml/2023/zhao2023icml-local/}
}