On the Statistical Efficiency of Mean-Field Reinforcement Learning with General Function Approximation
Abstract
In this paper, we study the fundamental statistical efficiency of Reinforcement Learning in Mean-Field Control (MFC) and Mean-Field Game (MFG) with general model-based function approximation. We introduce a new concept called Mean-Field Model-Based Eluder Dimension (MF-MBED), which characterizes the inherent complexity of mean-field model classes. We show that a rich family of Mean-Field RL problems exhibits low MF-MBED. Additionally, we propose algorithms based on maximal likelihood estimation, which can return an $\epsilon$-optimal policy for MFC or an $\epsilon$-Nash Equilibrium policy for MFG. The overall sample complexity depends only polynomially on MF-MBED, which is potentially much lower than the size of state-action space. Compared with previous works, our results only require the minimal assumptions including realizability and Lipschitz continuity.
Cite
Text
Huang et al. "On the Statistical Efficiency of Mean-Field Reinforcement Learning with General Function Approximation." Artificial Intelligence and Statistics, 2024.Markdown
[Huang et al. "On the Statistical Efficiency of Mean-Field Reinforcement Learning with General Function Approximation." Artificial Intelligence and Statistics, 2024.](https://mlanthology.org/aistats/2024/huang2024aistats-statistical/)BibTeX
@inproceedings{huang2024aistats-statistical,
title = {{On the Statistical Efficiency of Mean-Field Reinforcement Learning with General Function Approximation}},
author = {Huang, Jiawei and Yardim, Batuhan and He, Niao},
booktitle = {Artificial Intelligence and Statistics},
year = {2024},
pages = {289-297},
volume = {238},
url = {https://mlanthology.org/aistats/2024/huang2024aistats-statistical/}
}