Teaching Invariance Using Privileged Mediation Information

Abstract

The performance of deep neural networks often deteriorates in out-of-distribution settings due to relying on easy-to-learn but unreliable spurious associations known as shortcuts. Recent work attempting to mitigate shortcut learning relies on a priori knowledge of what the shortcut is and requires a strict overlap assumption with respect to the shortcut and the labels. In this paper, we present a causally-motivated teacher-student framework that encourages invariance to all shortcuts by leveraging privileged mediation information. The Teaching Invariance using Privileged Mediation Information (TIPMI) framework distills knowledge from a counterfactually invariant teacher trained using privileged mediation information to a student predictor that uses non-privileged features. We analyze the theoretical properties of our proposed estimator, showing that TIPMI promotes invariance to multiple unknown shortcuts and has better finite-sample efficiency. We empirically verify our theoretical findings by showing that TIPMI outperforms several state-of-the-art methods on two vision datasets and one language dataset.

Cite

Text

Zapzalka and Makar. "Teaching Invariance Using Privileged Mediation Information." NeurIPS 2024 Workshops: CRL, 2024.

Markdown

[Zapzalka and Makar. "Teaching Invariance Using Privileged Mediation Information." NeurIPS 2024 Workshops: CRL, 2024.](https://mlanthology.org/neuripsw/2024/zapzalka2024neuripsw-teaching/)

BibTeX

@inproceedings{zapzalka2024neuripsw-teaching,
  title     = {{Teaching Invariance Using Privileged Mediation Information}},
  author    = {Zapzalka, Dylan and Makar, Maggie},
  booktitle = {NeurIPS 2024 Workshops: CRL},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/zapzalka2024neuripsw-teaching/}
}