On the Statistical Efficiency of Mean-Field Reinforcement Learning with General Function Approximation

Abstract

In this paper, we study the fundamental statistical efficiency of Reinforcement Learning in Mean-Field Control (MFC) and Mean-Field Game (MFG) with general model-based function approximation. We introduce a new concept called Mean-Field Model-Based Eluder Dimension (MF-MBED), which characterizes the inherent complexity of mean-field model classes. We show that a rich family of Mean-Field RL problems exhibits low MF-MBED. Additionally, we propose algorithms based on maximal likelihood estimation, which can return an $\epsilon$-optimal policy for MFC or an $\epsilon$-Nash Equilibrium policy for MFG. The overall sample complexity depends only polynomially on MF-MBED, which is potentially much lower than the size of state-action space. Compared with previous works, our results only require the minimal assumptions including realizability and Lipschitz continuity.

Cite

Text

Huang et al. "On the Statistical Efficiency of Mean-Field Reinforcement Learning with General Function Approximation." Artificial Intelligence and Statistics, 2024.

Markdown

[Huang et al. "On the Statistical Efficiency of Mean-Field Reinforcement Learning with General Function Approximation." Artificial Intelligence and Statistics, 2024.](https://mlanthology.org/aistats/2024/huang2024aistats-statistical/)

BibTeX

@inproceedings{huang2024aistats-statistical,
  title     = {{On the Statistical Efficiency of Mean-Field Reinforcement Learning with General Function Approximation}},
  author    = {Huang, Jiawei and Yardim, Batuhan and He, Niao},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2024},
  pages     = {289-297},
  volume    = {238},
  url       = {https://mlanthology.org/aistats/2024/huang2024aistats-statistical/}
}