Deep Policy Gradient Methods Without Batch Updates, Target Networks, or Replay Buffers

Abstract

Modern deep policy gradient methods achieve effective performance on simulated robotic tasks, but they all require large replay buffers or expensive batch updates, or both, making them incompatible for real systems with resource-limited computers. We show that these methods fail catastrophically when limited to small replay buffers or during incremental learning, where updates only use the most recent sample without batch updates or a replay buffer. We propose a novel incremental deep policy gradient method --- Action Value Gradient (AVG) and a set of normalization and scaling techniques to address the challenges of instability in incremental learning. On robotic simulation benchmarks, we show that AVG is the only incremental method that learns effectively, often achieving final performance comparable to batch policy gradient methods. This advancement enabled us to show for the first time effective deep reinforcement learning with real robots using only incremental updates, employing a robotic manipulator and a mobile robot.

Cite

Text

Vasan et al. "Deep Policy Gradient Methods Without Batch Updates, Target Networks, or Replay Buffers." Neural Information Processing Systems, 2024. doi:10.52202/079017-0026

Markdown

[Vasan et al. "Deep Policy Gradient Methods Without Batch Updates, Target Networks, or Replay Buffers." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/vasan2024neurips-deep/) doi:10.52202/079017-0026

BibTeX

@inproceedings{vasan2024neurips-deep,
  title     = {{Deep Policy Gradient Methods Without Batch Updates, Target Networks, or Replay Buffers}},
  author    = {Vasan, Gautham and Elsayed, Mohamed and Azimi, Alireza and He, Jiamin and Shariar, Fahim and Bellinger, Colin and White, Martha and Mahmood, A. Rupam},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-0026},
  url       = {https://mlanthology.org/neurips/2024/vasan2024neurips-deep/}
}