One-Network Adversarial Fairness

Abstract

There is currently a great expansion of the impact of machine learning algorithms on our lives, prompting the need for objectives other than pure performance, including fairness. Fairness here means that the outcome of an automated decisionmaking system should not discriminate between subgroups characterized by sensitive attributes such as gender or race. Given any existing differentiable classifier, we make only slight adjustments to the architecture including adding a new hidden layer, in order to enable the concurrent adversarial optimization for fairness and accuracy. Our framework provides one way to quantify the tradeoff between fairness and accuracy, while also leading to strong empirical performance.

Cite

Text

Adel et al. "One-Network Adversarial Fairness." AAAI Conference on Artificial Intelligence, 2019. doi:10.1609/AAAI.V33I01.33012412

Markdown

[Adel et al. "One-Network Adversarial Fairness." AAAI Conference on Artificial Intelligence, 2019.](https://mlanthology.org/aaai/2019/adel2019aaai-one/) doi:10.1609/AAAI.V33I01.33012412

BibTeX

@inproceedings{adel2019aaai-one,
  title     = {{One-Network Adversarial Fairness}},
  author    = {Adel, Tameem and Valera, Isabel and Ghahramani, Zoubin and Weller, Adrian},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2019},
  pages     = {2412-2420},
  doi       = {10.1609/AAAI.V33I01.33012412},
  url       = {https://mlanthology.org/aaai/2019/adel2019aaai-one/}
}