Learning Mean-Field Games with Discounted and Average Costs

Abstract

We consider learning approximate Nash equilibria for discrete-time mean-field games with stochastic nonlinear state dynamics subject to both average and discounted costs. To this end, we introduce a mean-field equilibrium (MFE) operator, whose fixed point is a mean-field equilibrium, i.e., equilibrium in the infinite population limit. We first prove that this operator is a contraction, and propose a learning algorithm to compute an approximate mean-field equilibrium by approximating the MFE operator with a random one. Moreover, using the contraction property of the MFE operator, we establish the error analysis of the proposed learning algorithm. We then show that the learned mean-field equilibrium constitutes an approximate Nash equilibrium for finite-agent games.

Cite

Text

Anahtarci et al. "Learning Mean-Field Games with Discounted and Average Costs." Journal of Machine Learning Research, 2023.

Markdown

[Anahtarci et al. "Learning Mean-Field Games with Discounted and Average Costs." Journal of Machine Learning Research, 2023.](https://mlanthology.org/jmlr/2023/anahtarci2023jmlr-learning/)

BibTeX

@article{anahtarci2023jmlr-learning,
  title     = {{Learning Mean-Field Games with Discounted and Average Costs}},
  author    = {Anahtarci, Berkay and Kariksiz, Can Deha and Saldi, Naci},
  journal   = {Journal of Machine Learning Research},
  year      = {2023},
  pages     = {1-59},
  volume    = {24},
  url       = {https://mlanthology.org/jmlr/2023/anahtarci2023jmlr-learning/}
}