Q-Ensemble for Offline RL: Don't Scale the Ensemble, Scale the Batch Size

Abstract

Training large neural networks is known to be time-consuming, with the learning duration taking days or even weeks. To address this problem, large-batch optimization was introduced. This approach demonstrated that scaling mini-batch sizes with appropriate learning rate adjustments can speed up the training process by orders of magnitude. While long training time was not typically a major issue for model-free deep offline RL algorithms, recently introduced Q-ensemble methods achieving state-of-the-art performance made this issue more relevant, notably extending the training duration. In this work, we demonstrate how this class of methods can benefit from large-batch optimization, which is commonly overlooked by the deep offline RL community. We show that scaling the mini-batch size and naively adjusting the learning rate allows for (1) a reduced size of the Q-ensemble, (2) stronger penalization of out-of-distribution actions, and (3) improved convergence time, effectively shortening training duration by 3x-4x times on average.

Cite

Text

Nikulin et al. "Q-Ensemble for Offline RL: Don't Scale the Ensemble, Scale the Batch Size." NeurIPS 2022 Workshops: Offline_RL, 2022.

Markdown

[Nikulin et al. "Q-Ensemble for Offline RL: Don't Scale the Ensemble, Scale the Batch Size." NeurIPS 2022 Workshops: Offline_RL, 2022.](https://mlanthology.org/neuripsw/2022/nikulin2022neuripsw-qensemble/)

BibTeX

@inproceedings{nikulin2022neuripsw-qensemble,
  title     = {{Q-Ensemble for Offline RL: Don't Scale the Ensemble, Scale the Batch Size}},
  author    = {Nikulin, Alexander and Kurenkov, Vladislav and Tarasov, Denis and Akimov, Dmitry and Kolesnikov, Sergey},
  booktitle = {NeurIPS 2022 Workshops: Offline_RL},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/nikulin2022neuripsw-qensemble/}
}