What Causes a Disparate Impact in a Quantized Model?

Abstract

Post Training Quantization (PTQ) is widely adopted due to its high compression capacity and speed with minimal impact on accuracy. However, we observed that disparate impacts are exacerbated by quantization, especially for minority groups. Our analysis explains that in the course of quantization, the changes in weights and activations cause cascaded impacts in the network, resulting in logits with lower variance, increased loss, and compromised group accuracies. We extend our study to verify the influence of these impacts on group gradient norms and eigenvalues of the Hessian matrix, providing insights into the state of the network from an optimization point of view.

Cite

Text

Bellam and Kim. "What Causes a Disparate Impact in a Quantized Model?." NeurIPS 2024 Workshops: FITML, 2024.

Markdown

[Bellam and Kim. "What Causes a Disparate Impact in a Quantized Model?." NeurIPS 2024 Workshops: FITML, 2024.](https://mlanthology.org/neuripsw/2024/bellam2024neuripsw-causes/)

BibTeX

@inproceedings{bellam2024neuripsw-causes,
  title     = {{What Causes a Disparate Impact in a Quantized Model?}},
  author    = {Bellam, Abhimanyu and Kim, Jung-Eun},
  booktitle = {NeurIPS 2024 Workshops: FITML},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/bellam2024neuripsw-causes/}
}