FairRR: Pre-Processing for Group Fairness Through Randomized Response
Abstract
The increasing usage of machine learning models in consequential decision-making processes has spurred research into the fairness of these systems. While significant work has been done to study group fairness in the in-processing and post-processing setting, there has been little that theoretically connects these results to the pre-processing domain. This paper extends recent fair statistical learning results and proposes that achieving group fairness in downstream models can be formulated as finding the optimal design matrix in which to modify a response variable in a Randomized Response framework. We show that measures of group fairness can be directly controlled for with optimal model utility, proposing a pre-processing algorithm called FairRR that yields excellent downstream model utility and fairness.
Cite
Text
John Ward et al. "FairRR: Pre-Processing for Group Fairness Through Randomized Response." Artificial Intelligence and Statistics, 2024.Markdown
[John Ward et al. "FairRR: Pre-Processing for Group Fairness Through Randomized Response." Artificial Intelligence and Statistics, 2024.](https://mlanthology.org/aistats/2024/johnward2024aistats-fairrr/)BibTeX
@inproceedings{johnward2024aistats-fairrr,
title = {{FairRR: Pre-Processing for Group Fairness Through Randomized Response}},
author = {John Ward, Joshua and Zeng, Xianli and Cheng, Guang},
booktitle = {Artificial Intelligence and Statistics},
year = {2024},
pages = {3826-3834},
volume = {238},
url = {https://mlanthology.org/aistats/2024/johnward2024aistats-fairrr/}
}