Fairness Without Demographics Through Adversarially Reweighted Learning

Abstract

Much of the previous machine learning (ML) fairness literature assumes that protected features such as race and sex are present in the dataset, and relies upon them to mitigate fairness concerns. However, in practice factors like privacy and regulation often preclude the collection of protected features, or their use for training or inference, severely limiting the applicability of traditional fairness research. Therefore, we ask: How can we train a ML model to improve fairness when we do not even know the protected group memberships? In this work we address this problem by proposing Adversarially Reweighted Learning (ARL). In particular, we hypothesize that non-protected features and task labels are valuable for identifying fairness issues, and can be used to co-train an adversarial reweighting approach for improving fairness. Our results show that ARL improves Rawlsian Max-Min fairness, with notable AUC improvements for worst-case protected groups in multiple datasets, outperforming state-of-the-art alternatives.

Cite

Text

Lahoti et al. "Fairness Without Demographics Through Adversarially Reweighted Learning." Neural Information Processing Systems, 2020.

Markdown

[Lahoti et al. "Fairness Without Demographics Through Adversarially Reweighted Learning." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/lahoti2020neurips-fairness/)

BibTeX

@inproceedings{lahoti2020neurips-fairness,
  title     = {{Fairness Without Demographics Through Adversarially Reweighted Learning}},
  author    = {Lahoti, Preethi and Beutel, Alex and Chen, Jilin and Lee, Kang Wook and Prost, Flavien and Thain, Nithum and Wang, Xuezhi and Chi, Ed},
  booktitle = {Neural Information Processing Systems},
  year      = {2020},
  url       = {https://mlanthology.org/neurips/2020/lahoti2020neurips-fairness/}
}