Learning Mean-Field Games
Abstract
This paper presents a general mean-field game (GMFG) framework for simultaneous learning and decision-making in stochastic games with a large population. It first establishes the existence of a unique Nash Equilibrium to this GMFG, and explains that naively combining Q-learning with the fixed-point approach in classical MFGs yields unstable algorithms. It then proposes a Q-learning algorithm with Boltzmann policy (GMF-Q), with analysis of convergence property and computational complexity. The experiments on repeated Ad auction problems demonstrate that this GMF-Q algorithm is efficient and robust in terms of convergence and learning accuracy. Moreover, its performance is superior in convergence, stability, and learning ability, when compared with existing algorithms for multi-agent reinforcement learning.
Cite
Text
Guo et al. "Learning Mean-Field Games." Neural Information Processing Systems, 2019.Markdown
[Guo et al. "Learning Mean-Field Games." Neural Information Processing Systems, 2019.](https://mlanthology.org/neurips/2019/guo2019neurips-learning/)BibTeX
@inproceedings{guo2019neurips-learning,
title = {{Learning Mean-Field Games}},
author = {Guo, Xin and Hu, Anran and Xu, Renyuan and Zhang, Junzi},
booktitle = {Neural Information Processing Systems},
year = {2019},
pages = {4966-4976},
url = {https://mlanthology.org/neurips/2019/guo2019neurips-learning/}
}