Monitoring Shortcut Learning Using Mutual Information

Abstract

The failure of deep neural networks to generalize to out-of-distribution data is a well-known problem and raises concerns about the deployment of trained networks in safety-critical domains such as healthcare, finance, and autonomous vehicles. We study a particular kind of distribution shift — shortcuts or spurious correlations in the training data. Shortcut learning is often only exposed when models are evaluated on real-world data that does not contain the same spurious correlations, posing a serious dilemma for AI practitioners to properly assess the effectiveness of a trained model for real-world applications. In this work, we propose to use the mutual information (MI) between the learned representation and the input as a metric to find where in training the network latches onto shortcuts. Experiments demonstrate that MI can be used as a domain-agnostic metric for detecting shortcut learning.

Cite

Text

Adnan et al. "Monitoring Shortcut Learning Using Mutual Information." ICML 2022 Workshops: SCIS, 2022.

Markdown

[Adnan et al. "Monitoring Shortcut Learning Using Mutual Information." ICML 2022 Workshops: SCIS, 2022.](https://mlanthology.org/icmlw/2022/adnan2022icmlw-monitoring/)

BibTeX

@inproceedings{adnan2022icmlw-monitoring,
  title     = {{Monitoring Shortcut Learning Using Mutual Information}},
  author    = {Adnan, Mohammed and Ioannou, Yani and Tsai, Kenyon and Galloway, Angus and Tizhoosh, Hamid and Taylor, Graham W.},
  booktitle = {ICML 2022 Workshops: SCIS},
  year      = {2022},
  url       = {https://mlanthology.org/icmlw/2022/adnan2022icmlw-monitoring/}
}