DARL: Distance-Aware Uncertainty Estimation for Offline Reinforcement Learning
Abstract
To facilitate offline reinforcement learning, uncertainty estimation is commonly used to detect out-of-distribution data. By inspecting, we show that current explicit uncertainty estimators such as Monte Carlo Dropout and model ensemble are not competent to provide trustworthy uncertainty estimation in offline reinforcement learning. Accordingly, we propose a non-parametric distance-aware uncertainty estimator which is sensitive to the change in the input space for offline reinforcement learning. Based on our new estimator, adaptive truncated quantile critics are proposed to underestimate the out-of-distribution samples. We show that the proposed distance-aware uncertainty estimator is able to offer better uncertainty estimation compared to previous methods. Experimental results demonstrate that our proposed DARL method is competitive to the state-of-the-art methods in offline evaluation tasks.
Cite
Text
Zhang et al. "DARL: Distance-Aware Uncertainty Estimation for Offline Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I9.26327Markdown
[Zhang et al. "DARL: Distance-Aware Uncertainty Estimation for Offline Reinforcement Learning." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/zhang2023aaai-darl/) doi:10.1609/AAAI.V37I9.26327BibTeX
@inproceedings{zhang2023aaai-darl,
title = {{DARL: Distance-Aware Uncertainty Estimation for Offline Reinforcement Learning}},
author = {Zhang, Hongchang and Shao, Jianzhun and He, Shuncheng and Jiang, Yuhang and Ji, Xiangyang},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {11210-11218},
doi = {10.1609/AAAI.V37I9.26327},
url = {https://mlanthology.org/aaai/2023/zhang2023aaai-darl/}
}