Analyzing Reward Functions via Trajectory Alignment
Abstract
Reward design in reinforcement learning (RL) is often overlooked, with the assumption that a well-defined reward is readily available. However, reward functions can be challenging to design and prone to reward hacking, potentially leading to unintended or dangerous consequences in real-world applications. To create safe RL agents, reward alignment is crucial. We define reward alignment as the process of designing reward functions that preserve the preferences of a human stakeholder. In practice, reward functions are designed with training performance as the primary measure of success; this measure, however, may not reflect alignment. This work studies the practical implications of reward design on alignment. Specifically, we (1) propose a reward alignment metric, the Trajectory Alignment coefficient, that measures the similarity between the preference orderings of a human stakeholder and the preference orderings induced by a reward function, (2) use this metric to quantify the prevalence and extent of misalignment in human-designed reward functions, and (3) examine how misalignment affects the efficacy of these human-designed reward functions in terms of training performance.
Cite
Text
Muslimani et al. "Analyzing Reward Functions via Trajectory Alignment." NeurIPS 2024 Workshops: Behavioral_ML, 2024.Markdown
[Muslimani et al. "Analyzing Reward Functions via Trajectory Alignment." NeurIPS 2024 Workshops: Behavioral_ML, 2024.](https://mlanthology.org/neuripsw/2024/muslimani2024neuripsw-analyzing/)BibTeX
@inproceedings{muslimani2024neuripsw-analyzing,
title = {{Analyzing Reward Functions via Trajectory Alignment}},
author = {Muslimani, Calarina and Chandramouli, Suyog and Booth, Serena and Knox, W. Bradley and Taylor, Matthew E.},
booktitle = {NeurIPS 2024 Workshops: Behavioral_ML},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/muslimani2024neuripsw-analyzing/}
}