Understanding Instance-Level Impact of Fairness Constraints

Abstract

A variety of fairness constraints have been proposed in the literature to mitigate group-level statistical bias. Their impacts have been largely evaluated for different groups of populations corresponding to a set of sensitive attributes, such as race or gender. Nonetheless, the community has not observed sufficient explorations for how imposing fairness constraints fare at an instance level. Building on the concept of influence function, a measure that characterizes the impact of a training example on the target model and its predictive performance, this work studies the influence of training examples when fairness constraints are imposed. We find out that under certain assumptions, the influence function with respect to fairness constraints can be decomposed into a kernelized combination of training examples. One promising application of the proposed fairness influence function is to identify suspicious training examples that may cause model discrimination by ranking their influence scores. We demonstrate with extensive experiments that training on a subset of weighty data examples leads to lower fairness violations with a trade-off of accuracy.

Cite

Text

Wang et al. "Understanding Instance-Level Impact of Fairness Constraints." International Conference on Machine Learning, 2022.

Markdown

[Wang et al. "Understanding Instance-Level Impact of Fairness Constraints." International Conference on Machine Learning, 2022.](https://mlanthology.org/icml/2022/wang2022icml-understanding/)

BibTeX

@inproceedings{wang2022icml-understanding,
  title     = {{Understanding Instance-Level Impact of Fairness Constraints}},
  author    = {Wang, Jialu and Wang, Xin Eric and Liu, Yang},
  booktitle = {International Conference on Machine Learning},
  year      = {2022},
  pages     = {23114-23130},
  volume    = {162},
  url       = {https://mlanthology.org/icml/2022/wang2022icml-understanding/}
}