Searching for Unfairness in Algorithms' Outputs: Novel Tests and Insights

Abstract

As AI algorithms are deployed extensively, the need to ensure the fairness of their outputs is critical. Most existing work is on “fairness by design” approaches that incorporate limited tests for fairness into a limited number of algorithms. Here, we explore a framework that removes these limitations and can be used with any algorithm’s output that allocates instances to one of K categories/classes such as outlier detection (OD), clustering and classification. The framework can encode standard and novel fairness types beyond simple counting, and importantly, it can detect intersectional unfairness without being specifically told what to look for. Our experimental results show that both standard and novel types of unfairness exist extensively in the outputs of fair-by-design algorithms and the counter-intuitive result that they can actually increase intersectional unfairness.

Cite

Text

Davidson and Ravi. "Searching for Unfairness in Algorithms' Outputs: Novel Tests and Insights." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I26.34933

Markdown

[Davidson and Ravi. "Searching for Unfairness in Algorithms' Outputs: Novel Tests and Insights." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/davidson2025aaai-searching/) doi:10.1609/AAAI.V39I26.34933

BibTeX

@inproceedings{davidson2025aaai-searching,
  title     = {{Searching for Unfairness in Algorithms' Outputs: Novel Tests and Insights}},
  author    = {Davidson, Ian and Ravi, S. S.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {27242-27249},
  doi       = {10.1609/AAAI.V39I26.34933},
  url       = {https://mlanthology.org/aaai/2025/davidson2025aaai-searching/}
}