Revisiting Behavior Regularized Actor-Critic
Abstract
In recent years, significant advancements have been made in offline reinforcement learning, with a growing number of novel algorithms of varying degrees of complexity. Despite this progress, the significance of specific design choices and the application of common deep learning techniques remains unexplored. In this work, we demonstrate that it is possible to achieve state-of-the-art performance on the D4RL benchmark through a simple set of modifications to the minimalist offline RL approach and careful hyperparameter search. Furthermore, our ablations emphasize the importance of minor design choices and hyperparameter tuning while highlighting the untapped potential of using deep learning techniques in offline reinforcement learning.
Cite
Text
Tarasov et al. "Revisiting Behavior Regularized Actor-Critic." ICLR 2023 Workshops: RRL, 2023.Markdown
[Tarasov et al. "Revisiting Behavior Regularized Actor-Critic." ICLR 2023 Workshops: RRL, 2023.](https://mlanthology.org/iclrw/2023/tarasov2023iclrw-revisiting/)BibTeX
@inproceedings{tarasov2023iclrw-revisiting,
title = {{Revisiting Behavior Regularized Actor-Critic}},
author = {Tarasov, Denis and Kurenkov, Vladislav and Nikulin, Alexander and Kolesnikov, Sergey},
booktitle = {ICLR 2023 Workshops: RRL},
year = {2023},
url = {https://mlanthology.org/iclrw/2023/tarasov2023iclrw-revisiting/}
}