Top-GAP: Integrating Size Priors in CNNs for More Interpretability, Robustness, and Bias Mitigation
Abstract
This paper introduces Top-GAP, a novel regularization technique that enhances the explainability and robustness of convolutional neural networks. By constraining the spatial size of the learned feature representation, our method forces the network to focus on the most salient image regions, effectively reducing background influence. Using adversarial attacks and the Effective Receptive Field, we show that Top-GAP directs more attention towards object pixels rather than the background. This leads to enhanced interpretability and robustness. We achieve over 50% robust accuracy on CIFAR-10 with PGD $\epsilon =\nicefrac {8}{255}$ ϵ = 8 255 and 20 iterations while maintaining the original clean accuracy. Furthermore, we see increases of up to 5% accuracy against distribution shifts. Our approach also yields more precise object localization, as evidenced by up to 25% improvement in Intersection over Union (IOU) compared to methods like GradCAM and Recipro-CAM.
Cite
Text
Nieradzik et al. "Top-GAP: Integrating Size Priors in CNNs for More Interpretability, Robustness, and Bias Mitigation." European Conference on Computer Vision Workshops, 2024. doi:10.1007/978-3-031-92648-8_9Markdown
[Nieradzik et al. "Top-GAP: Integrating Size Priors in CNNs for More Interpretability, Robustness, and Bias Mitigation." European Conference on Computer Vision Workshops, 2024.](https://mlanthology.org/eccvw/2024/nieradzik2024eccvw-topgap/) doi:10.1007/978-3-031-92648-8_9BibTeX
@inproceedings{nieradzik2024eccvw-topgap,
title = {{Top-GAP: Integrating Size Priors in CNNs for More Interpretability, Robustness, and Bias Mitigation}},
author = {Nieradzik, Lars and Stephani, Henrike and Keuper, Janis},
booktitle = {European Conference on Computer Vision Workshops},
year = {2024},
pages = {134-151},
doi = {10.1007/978-3-031-92648-8_9},
url = {https://mlanthology.org/eccvw/2024/nieradzik2024eccvw-topgap/}
}