One-Network Adversarial Fairness
Abstract
There is currently a great expansion of the impact of machine learning algorithms on our lives, prompting the need for objectives other than pure performance, including fairness. Fairness here means that the outcome of an automated decisionmaking system should not discriminate between subgroups characterized by sensitive attributes such as gender or race. Given any existing differentiable classifier, we make only slight adjustments to the architecture including adding a new hidden layer, in order to enable the concurrent adversarial optimization for fairness and accuracy. Our framework provides one way to quantify the tradeoff between fairness and accuracy, while also leading to strong empirical performance.
Cite
Text
Adel et al. "One-Network Adversarial Fairness." AAAI Conference on Artificial Intelligence, 2019. doi:10.1609/AAAI.V33I01.33012412Markdown
[Adel et al. "One-Network Adversarial Fairness." AAAI Conference on Artificial Intelligence, 2019.](https://mlanthology.org/aaai/2019/adel2019aaai-one/) doi:10.1609/AAAI.V33I01.33012412BibTeX
@inproceedings{adel2019aaai-one,
title = {{One-Network Adversarial Fairness}},
author = {Adel, Tameem and Valera, Isabel and Ghahramani, Zoubin and Weller, Adrian},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2019},
pages = {2412-2420},
doi = {10.1609/AAAI.V33I01.33012412},
url = {https://mlanthology.org/aaai/2019/adel2019aaai-one/}
}