Fewer May Be Better: Enhancing Offline Reinforcement Learning with Reduced Dataset

Abstract

Research in offline reinforcement learning (RL) marks a paradigm shift in RL. However, a critical yet under-investigated aspect of offline RL is determining the subset of the offline dataset, which is used to improve algorithm performance while accelerating algorithm training. Moreover, the size of reduced datasets can uncover the requisite offline data volume essential for addressing analogous challenges. Based on the above considerations, we propose identifying Reduced Datasets for Offline RL (ReDOR) by formulating it as a gradient approximation optimization problem. We prove that the common actor-critic framework in reinforcement learning can be transformed into a submodular objective. This insight enables us to construct a subset by adopting the orthogonal matching pursuit (OMP). Specifically, we have made several critical modifications to OMP to enable successful adaptation with Offline RL algorithms. The experimental results indicate that the data subsets constructed by the ReDOR can significantly improve algorithm performance with low computational complexity.

Cite

Text

Yang et al. "Fewer May Be Better: Enhancing Offline Reinforcement Learning with Reduced Dataset." International Conference on Learning Representations, 2025.

Markdown

[Yang et al. "Fewer May Be Better: Enhancing Offline Reinforcement Learning with Reduced Dataset." International Conference on Learning Representations, 2025.](https://mlanthology.org/iclr/2025/yang2025iclr-fewer/)

BibTeX

@inproceedings{yang2025iclr-fewer,
  title     = {{Fewer May Be Better: Enhancing Offline Reinforcement Learning with Reduced Dataset}},
  author    = {Yang, Yiqin and Wang, Quanwei and Li, Chenghao and Hu, Hao and Wu, Chengjie and Jiang, Yuhua and Zhong, Dianyu and Zhang, Ziyou and Zhao, Qianchuan and Zhang, Chongjie and Xu, Bo},
  booktitle = {International Conference on Learning Representations},
  year      = {2025},
  url       = {https://mlanthology.org/iclr/2025/yang2025iclr-fewer/}
}