On-Policy Deep Reinforcement Learning for the Average-Reward Criterion

Abstract

We develop theory and algorithms for average-reward on-policy Reinforcement Learning (RL). We first consider bounding the difference of the long-term average reward for two policies. We show that previous work based on the discounted return (Schulman et al. 2015, Achiam et al. 2017) results in a non-meaningful lower bound in the average reward setting. By addressing the average-reward criterion directly, we then derive a novel bound which depends on the average divergence between the policies and on Kemeny’s constant. Based on this bound, we develop an iterative procedure which produces a sequence of monotonically improved policies for the average reward criterion. This iterative procedure can then be combined with classic Deep Reinforcement Learning (DRL) methods, resulting in practical DRL algorithms that target the long-run average reward criterion. In particular, we demonstrate that Average-Reward TRPO (ATRPO), which adapts the on-policy TRPO algorithm to the average-reward criterion, significantly outperforms TRPO in the most challenging MuJuCo environments.

Cite

Text

Zhang and Ross. "On-Policy Deep Reinforcement Learning for the Average-Reward Criterion." International Conference on Machine Learning, 2021.

Markdown

[Zhang and Ross. "On-Policy Deep Reinforcement Learning for the Average-Reward Criterion." International Conference on Machine Learning, 2021.](https://mlanthology.org/icml/2021/zhang2021icml-onpolicy/)

BibTeX

@inproceedings{zhang2021icml-onpolicy,
  title     = {{On-Policy Deep Reinforcement Learning for the Average-Reward Criterion}},
  author    = {Zhang, Yiming and Ross, Keith W},
  booktitle = {International Conference on Machine Learning},
  year      = {2021},
  pages     = {12535-12545},
  volume    = {139},
  url       = {https://mlanthology.org/icml/2021/zhang2021icml-onpolicy/}
}