A Geometric Perspective on Optimal Representations for Reinforcement Learning

Abstract

We propose a new perspective on representation learning in reinforcement learning based on geometric properties of the space of value functions. From there, we provide formal evidence regarding the usefulness of value functions as auxiliary tasks in reinforcement learning. Our formulation considers adapting the representation to minimize the (linear) approximation of the value function of all stationary policies for a given environment. We show that this optimization reduces to making accurate predictions regarding a special class of value functions which we call adversarial value functions (AVFs). We demonstrate that using value functions as auxiliary tasks corresponds to an expected-error relaxation of our formulation, with AVFs a natural candidate, and identify a close relationship with proto-value functions (Mahadevan, 2005). We highlight characteristics of AVFs and their usefulness as auxiliary tasks in a series of experiments on the four-room domain.

Cite

Text

Bellemare et al. "A Geometric Perspective on Optimal Representations for Reinforcement Learning." Neural Information Processing Systems, 2019.

Markdown

[Bellemare et al. "A Geometric Perspective on Optimal Representations for Reinforcement Learning." Neural Information Processing Systems, 2019.](https://mlanthology.org/neurips/2019/bellemare2019neurips-geometric/)

BibTeX

@inproceedings{bellemare2019neurips-geometric,
  title     = {{A Geometric Perspective on Optimal Representations for Reinforcement Learning}},
  author    = {Bellemare, Marc and Dabney, Will and Dadashi, Robert and Taiga, Adrien Ali and Castro, Pablo Samuel and Le Roux, Nicolas and Schuurmans, Dale and Lattimore, Tor and Lyle, Clare},
  booktitle = {Neural Information Processing Systems},
  year      = {2019},
  pages     = {4358-4369},
  url       = {https://mlanthology.org/neurips/2019/bellemare2019neurips-geometric/}
}