Pushing the Accuracy-Fairness Tradeoff Frontier with Introspective Self-Play
Abstract
Improving the accuracy-fairness frontier of deep neural network (DNN) models is an important problem. Uncertainty-based active learning (AL) can potentially improve the frontier by preferentially sampling underrepresented subgroups to create a more balanced training dataset. However, the quality of uncertainty estimates from modern DNNs tend to degrade in the presence of spurious correlations and dataset bias, compromising the effectiveness of AL for sampling tail groups. In this work, we propose $Introspective Self-play$ (ISP), a simple approach to improve the uncertainty estimation of a deep neural network under dataset bias, by adding an auxiliary $Introspection$ task requiring a model to predict the bias for each data point in addition to the label. We show that ISP provably improves the bias-awareness of the model representation and the resulting uncertainty estimates. On two real-world tabular and language tasks,ISP serves as a simple “plug-in” for AL model training, consistently improving both the tail-group sampling rate and the final accuracy-fairness trade-off frontier of popular AL methods.
Cite
Text
Liu et al. "Pushing the Accuracy-Fairness Tradeoff Frontier with Introspective Self-Play." NeurIPS 2022 Workshops: DistShift, 2022.Markdown
[Liu et al. "Pushing the Accuracy-Fairness Tradeoff Frontier with Introspective Self-Play." NeurIPS 2022 Workshops: DistShift, 2022.](https://mlanthology.org/neuripsw/2022/liu2022neuripsw-pushing/)BibTeX
@inproceedings{liu2022neuripsw-pushing,
title = {{Pushing the Accuracy-Fairness Tradeoff Frontier with Introspective Self-Play}},
author = {Liu, Jeremiah Zhe and Dvijotham, Krishnamurthy Dj and Lee, Jihyeon and Yuan, Quan and Strobel, Martin and Lakshminarayanan, Balaji and Ramachandran, Deepak},
booktitle = {NeurIPS 2022 Workshops: DistShift},
year = {2022},
url = {https://mlanthology.org/neuripsw/2022/liu2022neuripsw-pushing/}
}