Steady State Analysis of Episodic Reinforcement Learning

Abstract

Reinforcement Learning (RL) tasks generally divide into two kinds: continual learning and episodic learning. The concept of steady state has played a foundational role in the continual setting, where unique steady-state distribution is typically presumed to exist in the task being studied, which enables principled conceptual framework as well as efficient data collection method for continual RL algorithms. On the other hand, the concept of steady state has been widely considered irrelevant for episodic RL tasks, in which the decision process terminates in finite time. Alternative concepts, such as episode-wise visitation frequency, are used in episodic RL algorithms, which are not only inconsistent with their counterparts in continual RL, and also make it harder to design and analyze RL algorithms in the episodic setting.

Cite

Text

Bojun. "Steady State Analysis of Episodic Reinforcement Learning." Neural Information Processing Systems, 2020.

Markdown

[Bojun. "Steady State Analysis of Episodic Reinforcement Learning." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/bojun2020neurips-steady/)

BibTeX

@inproceedings{bojun2020neurips-steady,
  title     = {{Steady State Analysis of Episodic Reinforcement Learning}},
  author    = {Bojun, Huang},
  booktitle = {Neural Information Processing Systems},
  year      = {2020},
  url       = {https://mlanthology.org/neurips/2020/bojun2020neurips-steady/}
}