A Distributional Analogue to the Successor Representation

Abstract

This paper contributes a new approach for distributional reinforcement learning which elucidates a clean separation of transition structure and reward in the learning process. Analogous to how the successor representation (SR) describes the expected consequences of behaving according to a given policy, our distributional successor measure (SM) describes the distributional consequences of this behaviour. We formulate the distributional SM as a distribution over distributions and provide theory connecting it with distributional and model-based reinforcement learning. Moreover, we propose an algorithm that learns the distributional SM from data by minimizing a two-level maximum mean discrepancy. Key to our method are a number of algorithmic techniques that are independently valuable for learning generative models of state. As an illustration of the usefulness of the distributional SM, we show that it enables zero-shot risk-sensitive policy evaluation in a way that was not previously possible.

Cite

Text

Wiltzer et al. "A Distributional Analogue to the Successor Representation." International Conference on Machine Learning, 2024.

Markdown

[Wiltzer et al. "A Distributional Analogue to the Successor Representation." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/wiltzer2024icml-distributional/)

BibTeX

@inproceedings{wiltzer2024icml-distributional,
  title     = {{A Distributional Analogue to the Successor Representation}},
  author    = {Wiltzer, Harley and Farebrother, Jesse and Gretton, Arthur and Tang, Yunhao and Barreto, Andre and Dabney, Will and Bellemare, Marc G and Rowland, Mark},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {52994-53016},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/wiltzer2024icml-distributional/}
}