Group Fairness in Reinforcement Learning

Abstract

We pose and study the problem of satisfying fairness in the online Reinforcement Learning (RL) setting. We focus on the group notions of fairness, according to which agents belonging to different groups should have similar performance based on some given measure. We consider the setting of maximizing return in an unknown environment (unknown transition and reward function) and show that it is possible to have RL algorithms that learn the best fair policies without violating the fairness requirements at any point in time during the learning process. In the tabular finite-horizon episodic setting, we provide an algorithm that combines the principle of optimism and pessimism under uncertainty to achieve zero fairness violation with arbitrarily high probability while also maintaining sub-linear regret guarantees. For the high-dimensional Deep-RL setting, we present algorithms based on the performance-difference style approximate policy improvement update step and we report encouraging empirical results on various traditional RL-inspired benchmarks showing that our algorithms display the desired behavior of learning the optimal policy while performing a fair learning process.

Cite

Text

Satija et al. "Group Fairness in Reinforcement Learning." Transactions on Machine Learning Research, 2023.

Markdown

[Satija et al. "Group Fairness in Reinforcement Learning." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/satija2023tmlr-group/)

BibTeX

@article{satija2023tmlr-group,
  title     = {{Group Fairness in Reinforcement Learning}},
  author    = {Satija, Harsh and Lazaric, Alessandro and Pirotta, Matteo and Pineau, Joelle},
  journal   = {Transactions on Machine Learning Research},
  year      = {2023},
  url       = {https://mlanthology.org/tmlr/2023/satija2023tmlr-group/}
}