Value Preserving State-Action Abstractions

Abstract

Abstraction can improve the sample efficiency of reinforcement learning. However, the process of abstraction inherently discards information, potentially compromising an agent’s ability to represent high-value policies. To mitigate this, we here introduce combinations of state abstractions and options that are guaranteed to preserve representation of near-optimal policies. We first define $\phi$-relative options, a general formalism for analyzing the value loss of options paired with a state abstraction, and present necessary and sufficient conditions for $\phi$-relative options to preserve near-optimal behavior in any finite Markov Decision Process. We further show that, under appropriate assumptions, $\phi$-relative options can be composed to induce hierarchical abstractions that are also guaranteed to represent high-value policies.

Cite

Text

Abel et al. "Value Preserving State-Action Abstractions." Artificial Intelligence and Statistics, 2020.

Markdown

[Abel et al. "Value Preserving State-Action Abstractions." Artificial Intelligence and Statistics, 2020.](https://mlanthology.org/aistats/2020/abel2020aistats-value/)

BibTeX

@inproceedings{abel2020aistats-value,
  title     = {{Value Preserving State-Action Abstractions}},
  author    = {Abel, David and Umbanhowar, Nate and Khetarpal, Khimya and Arumugam, Dilip and Precup, Doina and Littman, Michael},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2020},
  pages     = {1639-1650},
  volume    = {108},
  url       = {https://mlanthology.org/aistats/2020/abel2020aistats-value/}
}