Trustworthy AI Must Account for Intersectionality
Abstract
Trustworthy AI encompasses many aspirational aspects for aligning AI systems with human values, including fairness, privacy, robustness, explainability, and uncertainty quantification. However, efforts to enhance one aspect often introduce unintended trade-offs that negatively impact others, making it challenging to improve all aspects simultaneously. In this paper, we review notable approaches to five aspects and systematically consider every pair, detailing the negative interactions that can arise. For example, applying differential privacy to model training can amplify biases in the data, undermining fairness. Drawing on these findings, we take the position that addressing trustworthiness along each axis in isolation is insufficient. Instead, to achieve better alignment between humans and AI, efforts in Trustworthy AI must account for intersectionality between aspects and adopt a holistic view across all relevant axes at once. To illustrate our perspective, we provide guidance on how researchers can work towards integrated trustworthiness, and a case study on how intersectionality applies to the financial industry.
Cite
Text
Cresswell. "Trustworthy AI Must Account for Intersectionality." ICLR 2025 Workshops: Bi-Align, 2025.Markdown
[Cresswell. "Trustworthy AI Must Account for Intersectionality." ICLR 2025 Workshops: Bi-Align, 2025.](https://mlanthology.org/iclrw/2025/cresswell2025iclrw-trustworthy/)BibTeX
@inproceedings{cresswell2025iclrw-trustworthy,
title = {{Trustworthy AI Must Account for Intersectionality}},
author = {Cresswell, Jesse C.},
booktitle = {ICLR 2025 Workshops: Bi-Align},
year = {2025},
url = {https://mlanthology.org/iclrw/2025/cresswell2025iclrw-trustworthy/}
}