The Matrix Reloaded: Towards Counterfactual Group Fairness in Machine Learning
Abstract
In today’s data-driven world, addressing bias is essential to minimize discriminatory outcomes and work toward fairness in machine learning models. This paper presents a novel data-centric framework for bias analysis, harnessing the power of counterfactual reasoning. We detail a process for generating plausible counterfactuals suited for group evaluation, using probabilistic distributions and optionally incorporating domain knowledge, as a more efficient alternative to computationally intensive generative models.Additionally, we introduce the Counterfactual Confusion Matrix, from which we derive a suite of metrics that provide a comprehensive view of a model’s behaviour under counterfactual conditions. These metrics offer unique insights into the model’s resilience and susceptibility to changes in sensitive attributes, such as sex or race. We demonstrate their utility and complementarity with standard group fairness metrics through experiments on real-world datasets. Our results show that domain knowledge is key, and that our metrics can reveal subtle biases that traditional bias evaluation strategies may overlook, providing a more nuanced understanding of potential model bias.
Cite
Text
Pinto et al. "The Matrix Reloaded: Towards Counterfactual Group Fairness in Machine Learning." Data-centric Machine Learning Research, 2024.Markdown
[Pinto et al. "The Matrix Reloaded: Towards Counterfactual Group Fairness in Machine Learning." Data-centric Machine Learning Research, 2024.](https://mlanthology.org/dmlr/2024/pinto2024dmlr-matrix/)BibTeX
@article{pinto2024dmlr-matrix,
title = {{The Matrix Reloaded: Towards Counterfactual Group Fairness in Machine Learning}},
author = {Pinto, Mariana and Carreiro, Andre V and Madeira, Pedro and Lopez, Alberto and Gamboa, Hugo},
journal = {Data-centric Machine Learning Research},
year = {2024},
pages = {1-55},
volume = {2},
url = {https://mlanthology.org/dmlr/2024/pinto2024dmlr-matrix/}
}