RIME: Robust Preference-Based Reinforcement Learning with Noisy Preferences

Abstract

Preference-based Reinforcement Learning (PbRL) circumvents the need for reward engineering by harnessing human preferences as the reward signal. However, current PbRL methods excessively depend on high-quality feedback from domain experts, which results in a lack of robustness. In this paper, we present RIME, a robust PbRL algorithm for effective reward learning from noisy preferences. Our method utilizes a sample selection-based discriminator to dynamically filter out noise and ensure robust training. To counteract the cumulative error stemming from incorrect selection, we suggest a warm start for the reward model, which additionally bridges the performance gap during the transition from pre-training to online training in PbRL. Our experiments on robotic manipulation and locomotion tasks demonstrate that RIME significantly enhances the robustness of the state-of-the-art PbRL method. Code is available at https://github.com/CJReinforce/RIME_ICML2024.

Cite

Text

Cheng et al. "RIME: Robust Preference-Based Reinforcement Learning with Noisy Preferences." International Conference on Machine Learning, 2024.

Markdown

[Cheng et al. "RIME: Robust Preference-Based Reinforcement Learning with Noisy Preferences." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/cheng2024icml-rime/)

BibTeX

@inproceedings{cheng2024icml-rime,
  title     = {{RIME: Robust Preference-Based Reinforcement Learning with Noisy Preferences}},
  author    = {Cheng, Jie and Xiong, Gang and Dai, Xingyuan and Miao, Qinghai and Lv, Yisheng and Wang, Fei-Yue},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {8229-8247},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/cheng2024icml-rime/}
}