Blackbox Post-Processing for Multiclass Fairness

Abstract

Applying standard machine learning approaches for classification can produce unequal results across different demographic groups. When then used in real-world settings, these inequities can have negative societal impacts. This has motivated the development of various approaches to fair classification with machine learning models in recent years. In this paper, we consider the problem of modifying the predictions of a blackbox machine learning classifier in order to achieve fairness in a multiclass setting. To accomplish this, we extend the 'post-processing' approach in Hardt et al. 2016, which focuses on fairness for binary classification, to the setting of fair multiclass classification. We explore when our approach produces both fair and accurate predictions through systematic synthetic experiments and also evaluate discrimination-fairness tradeoffs on several publicly available real-world application datasets. We find that overall, our approach produces minor drops in accuracy and enforces fairness when the number of individuals in the dataset is high relative to the number of classes and protected groups.

Cite

Text

Putzel and Lee. "Blackbox Post-Processing for Multiclass Fairness." AAAI Conference on Artificial Intelligence, 2022. doi:10.48550/arxiv.2201.04461

Markdown

[Putzel and Lee. "Blackbox Post-Processing for Multiclass Fairness." AAAI Conference on Artificial Intelligence, 2022.](https://mlanthology.org/aaai/2022/putzel2022aaai-blackbox/) doi:10.48550/arxiv.2201.04461

BibTeX

@inproceedings{putzel2022aaai-blackbox,
  title     = {{Blackbox Post-Processing for Multiclass Fairness}},
  author    = {Putzel, Preston and Lee, Scott},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2022},
  doi       = {10.48550/arxiv.2201.04461},
  url       = {https://mlanthology.org/aaai/2022/putzel2022aaai-blackbox/}
}