Adjusting Machine Learning Decisions for Equal Opportunity and Counterfactual Fairness

Abstract

Machine learning (ML) methods have the potential to automate high-stakes decisions, such as bail admissions or credit lending, by analyzing and learning from historical data. But these algorithmic decisions may be unfair: in learning from historical data, they may replicate discriminatory practices from the past. In this paper, we propose two algorithms that adjust fitted ML predictors to produce decisions that are fair. Our methods provide post-hoc adjustments to the predictors, without requiring that they be retrained. We consider a causal model of the ML decisions, define fairness through counterfactual decisions within the model, and then form algorithmic decisions that capture the historical data as well as possible but are provably fair. In particular, we consider two definitions of fairness. The first is ``equal counterfactual opportunity,'' where the counterfactual distribution of the decision is the same regardless of the protected attribute; the second is counterfactual fairness. We evaluate the algorithms, and the trade-off between accuracy and fairness, on datasets about admissions, income, credit, and recidivism.

Cite

Text

Wang et al. "Adjusting Machine Learning Decisions for Equal Opportunity and Counterfactual Fairness." Transactions on Machine Learning Research, 2023.

Markdown

[Wang et al. "Adjusting Machine Learning Decisions for Equal Opportunity and Counterfactual Fairness." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/wang2023tmlr-adjusting/)

BibTeX

@article{wang2023tmlr-adjusting,
  title     = {{Adjusting Machine Learning Decisions for Equal Opportunity and Counterfactual Fairness}},
  author    = {Wang, Yixin and Sridhar, Dhanya and Blei, David},
  journal   = {Transactions on Machine Learning Research},
  year      = {2023},
  url       = {https://mlanthology.org/tmlr/2023/wang2023tmlr-adjusting/}
}