Bias in Motion: Theoretical Insights into the Dynamics of Bias in SGD Training

Abstract

Machine learning systems often acquire biases by leveraging undesired features in the data, impacting accuracy variably across different sub-populations. This paper explores the evolution of bias in a teacher-student setup modeling different data sub-populations with a Gaussian-mixture model, by providing an analytical description of the stochastic gradient descent dynamics of a linear classifier in this setting. Our analysis reveals how different properties of sub-populations influence bias at different timescales, showing a shifting preference of the classifier during training. We empirically validate our results in more complex scenarios by training deeper networks on real datasets including CIFAR10, MNIST, and CelebA.

Cite

Text

Jain et al. "Bias in Motion: Theoretical Insights into the Dynamics of Bias in SGD Training." NeurIPS 2024 Workshops: M3L, 2024.

Markdown

[Jain et al. "Bias in Motion: Theoretical Insights into the Dynamics of Bias in SGD Training." NeurIPS 2024 Workshops: M3L, 2024.](https://mlanthology.org/neuripsw/2024/jain2024neuripsw-bias/)

BibTeX

@inproceedings{jain2024neuripsw-bias,
  title     = {{Bias in Motion: Theoretical Insights into the Dynamics of Bias in SGD Training}},
  author    = {Jain, Anchit and Nobahari, Rozhin and Baratin, Aristide and Mannelli, Stefano Sarao},
  booktitle = {NeurIPS 2024 Workshops: M3L},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/jain2024neuripsw-bias/}
}