On Representation Complexity of Model-Based and Model-Free Reinforcement Learning
Abstract
We study the representation complexity of model-based and model-free reinforcement learning (RL) in the context of circuit complexity. We prove theoretically that there exists a broad class of MDPs such that their underlying transition and reward functions can be represented by constant depth circuits with polynomial size, while the optimal $Q$-function suffers an exponential circuit complexity in constant-depth circuits. By drawing attention to the approximation errors and building connections to complexity theory, our theory provides unique insights into why model-based algorithms usually enjoy better sample complexity than model-free algorithms from a novel representation complexity perspective: in some cases, the ground-truth rule (model) of the environment is simple to represent, while other quantities, such as $Q$-function, appear complex. We empirically corroborate our theory by comparing the approximation error of the transition kernel, reward function, and optimal $Q$-function in various Mujoco environments, which demonstrates that the approximation errors of the transition kernel and reward function are consistently lower than those of the optimal $Q$-function. To the best of our knowledge, this work is the first to study the circuit complexity of RL, which also provides a rigorous framework for future research.
Cite
Text
Zhu et al. "On Representation Complexity of Model-Based and Model-Free Reinforcement Learning." International Conference on Learning Representations, 2024.Markdown
[Zhu et al. "On Representation Complexity of Model-Based and Model-Free Reinforcement Learning." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/zhu2024iclr-representation/)BibTeX
@inproceedings{zhu2024iclr-representation,
title = {{On Representation Complexity of Model-Based and Model-Free Reinforcement Learning}},
author = {Zhu, Hanlin and Huang, Baihe and Russell, Stuart},
booktitle = {International Conference on Learning Representations},
year = {2024},
url = {https://mlanthology.org/iclr/2024/zhu2024iclr-representation/}
}