Convergence of Actor-Critic with Multi-Layer Neural Networks
Abstract
The early theory of actor-critic methods considered convergence using linear function approximators for the policy and value functions. Recent work has established convergence using neural network approximators with a single hidden layer. In this work we are taking the natural next step and establish convergence using deep neural networks with an arbitrary number of hidden layers, thus closing a gap between theory and practice. We show that actor-critic updates projected on a ball around the initial condition will converge to a neighborhood where the average of the squared gradients is $\tilde{O} \left( 1/\sqrt{m} \right) + O \left( \epsilon \right)$, with $m$ being the width of the neural network and $\epsilon$ the approximation quality of the best critic neural network over the projected set.
Cite
Text
Tian et al. "Convergence of Actor-Critic with Multi-Layer Neural Networks." Neural Information Processing Systems, 2023.Markdown
[Tian et al. "Convergence of Actor-Critic with Multi-Layer Neural Networks." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/tian2023neurips-convergence/)BibTeX
@inproceedings{tian2023neurips-convergence,
title = {{Convergence of Actor-Critic with Multi-Layer Neural Networks}},
author = {Tian, Haoxing and Olshevsky, Alex and Paschalidis, Yannis},
booktitle = {Neural Information Processing Systems},
year = {2023},
url = {https://mlanthology.org/neurips/2023/tian2023neurips-convergence/}
}