GLANCE: Global to Local Architecture-Neutral Concept-Based Explanations

Abstract

Most of the current explainability techniques focus on capturing the importance of features in input space. However, given the complexity of models and data-generating processes, the resulting explanations are far from being complete, in that they lack an indication of feature interactions and visualization of their effect. In this work, we propose a novel surrogate-model-based explainability framework to explain the decisions of any CNN-based image classifiers by extracting causal relations between the features. These causal relations serve as global explanations from which local explanations of different forms can be obtained. Specifically, we employ a generator to visualize the `effect' of interactions among features in latent space and draw feature importance therefrom as local explanations. We demonstrate and evaluate explanations obtained with our framework on the Morpho-MNIST, the FFHQ, and the AFHQ datasets.

Cite

Text

Kori et al. "GLANCE: Global to Local Architecture-Neutral Concept-Based Explanations." NeurIPS 2023 Workshops: XAIA, 2023.

Markdown

[Kori et al. "GLANCE: Global to Local Architecture-Neutral Concept-Based Explanations." NeurIPS 2023 Workshops: XAIA, 2023.](https://mlanthology.org/neuripsw/2023/kori2023neuripsw-glance/)

BibTeX

@inproceedings{kori2023neuripsw-glance,
  title     = {{GLANCE: Global to Local Architecture-Neutral Concept-Based Explanations}},
  author    = {Kori, Avinash and Glocker, Ben and Toni, Francesca},
  booktitle = {NeurIPS 2023 Workshops: XAIA},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/kori2023neuripsw-glance/}
}