Enhanced Exploration via Variational Learned Priors
Abstract
Efficient exploration in reinforcement learning is challenging, especially in sparse-reward environments. Intrinsic motivation, such as rewarding state novelty, can enhance exploration. We propose an intrinsic motivation approach, called Variational Learned Priors, that uses variational state encoding to estimate novelty via the Kullback-Leibler divergence between the posterior distribution and a learned prior of a Variational Autoencoder. We assess this intrinsic reward with four different learned priors. Our results show that this method improves exploration efficiency and accelerates extrinsic reward accumulation across various domains.
Cite
Text
Nicholson et al. "Enhanced Exploration via Variational Learned Priors." NeurIPS 2024 Workshops: IMOL, 2024.Markdown
[Nicholson et al. "Enhanced Exploration via Variational Learned Priors." NeurIPS 2024 Workshops: IMOL, 2024.](https://mlanthology.org/neuripsw/2024/nicholson2024neuripsw-enhanced/)BibTeX
@inproceedings{nicholson2024neuripsw-enhanced,
title = {{Enhanced Exploration via Variational Learned Priors}},
author = {Nicholson, Jessica and Goodier, Joseph S and Patel, Akshil and Şimşek, Özgür},
booktitle = {NeurIPS 2024 Workshops: IMOL},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/nicholson2024neuripsw-enhanced/}
}