Targeted Activation Penalties Help CNNs Ignore Spurious Signals

Abstract

Neural networks (NNs) can learn to rely on spurious signals in the training data, leading to poor generalisation. Recent methods tackle this problem by training NNs with additional ground-truth annotations of such signals. These methods may, however, let spurious signals re-emerge in deep convolutional NNs (CNNs). We propose Targeted Activation Penalty (TAP), a new method tackling the same problem by penalising activations to control the re-emergence of spurious signals in deep CNNs, while also lowering training times and memory usage. In addition, ground-truth annotations can be expensive to obtain. We show that TAP still works well with annotations generated by pre-trained models as effective substitutes of ground-truth annotations. We demonstrate the power of TAP against two state-of-the-art baselines on the MNIST benchmark and on two clinical image datasets, using four different CNN architectures.

Cite

Text

Zhang et al. "Targeted Activation Penalties Help CNNs Ignore Spurious Signals." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I15.29610

Markdown

[Zhang et al. "Targeted Activation Penalties Help CNNs Ignore Spurious Signals." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/zhang2024aaai-targeted/) doi:10.1609/AAAI.V38I15.29610

BibTeX

@inproceedings{zhang2024aaai-targeted,
  title     = {{Targeted Activation Penalties Help CNNs Ignore Spurious Signals}},
  author    = {Zhang, Dekai and Williams, Matt and Toni, Francesca},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {16705-16713},
  doi       = {10.1609/AAAI.V38I15.29610},
  url       = {https://mlanthology.org/aaai/2024/zhang2024aaai-targeted/}
}