Collect & Infer - A Fresh Look at Data-Efficient Reinforcement Learning

Abstract

This position paper proposes a fresh look at Reinforcement Learning (RL) from the perspective of data-efficiency. RL has gone through three major stages: pure on-line RL where every data-point is considered only once, RL with a replay buffer where additional learning is done on a portion of the experience, and finally transition memory based RL, where, conceptually, all transitions are stored, and flexibly re-used in every update step. While inferring knowledge from all stored experience has led to a tremendous gain in data-efficiency, the question of how this data is collected has been vastly understudied. We argue that data-efficiency can only be achieved through careful consideration of both aspects. We propose to make this insight explicit via a paradigm that we call ’Collect and Infer’, which explicitly models RL as two separate but interconnected processes, concerned with data collection and knowledge inference respectively.

Cite

Text

Riedmiller et al. "Collect & Infer - A Fresh Look at Data-Efficient Reinforcement Learning." Conference on Robot Learning, 2021.

Markdown

[Riedmiller et al. "Collect & Infer - A Fresh Look at Data-Efficient Reinforcement Learning." Conference on Robot Learning, 2021.](https://mlanthology.org/corl/2021/riedmiller2021corl-collect/)

BibTeX

@inproceedings{riedmiller2021corl-collect,
  title     = {{Collect & Infer - A Fresh Look at Data-Efficient Reinforcement Learning}},
  author    = {Riedmiller, Martin and Springenberg, Jost Tobias and Hafner, Roland and Heess, Nicolas},
  booktitle = {Conference on Robot Learning},
  year      = {2021},
  pages     = {1736-1744},
  volume    = {164},
  url       = {https://mlanthology.org/corl/2021/riedmiller2021corl-collect/}
}