Balancing Act: Constraining Disparate Impact in Sparse Models
Abstract
Model pruning is a popular approach to enable the deployment of large deep learning models on edge devices with restricted computational or storage capacities. Although sparse models achieve performance comparable to that of their dense counterparts at the level of the entire dataset, they exhibit high accuracy drops for some data sub-groups. Existing methods to mitigate this disparate impact induced by pruning (i) rely on surrogate metrics that address the problem indirectly and have limited interpretability; or (ii) scale poorly with the number of protected sub-groups in terms of computational cost. We propose a constrained optimization approach that _directly addresses the disparate impact of pruning_: our formulation bounds the accuracy change between the dense and sparse models, for each sub-group. This choice of constraints provides an interpretable success criterion to determine if a pruned model achieves acceptable disparity levels. Experimental results demonstrate that our technique scales reliably to problems involving large models and hundreds of protected sub-groups.
Cite
Text
Hashemizadeh et al. "Balancing Act: Constraining Disparate Impact in Sparse Models." International Conference on Learning Representations, 2024.Markdown
[Hashemizadeh et al. "Balancing Act: Constraining Disparate Impact in Sparse Models." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/hashemizadeh2024iclr-balancing/)BibTeX
@inproceedings{hashemizadeh2024iclr-balancing,
title = {{Balancing Act: Constraining Disparate Impact in Sparse Models}},
author = {Hashemizadeh, Meraj and Ramirez, Juan and Sukumaran, Rohan and Farnadi, Golnoosh and Lacoste-Julien, Simon and Gallego-Posada, Jose},
booktitle = {International Conference on Learning Representations},
year = {2024},
url = {https://mlanthology.org/iclr/2024/hashemizadeh2024iclr-balancing/}
}