GLANCE: Global to Local Architecture-Neutral Concept-Based Explanations
Abstract
Most of the current explainability techniques focus on capturing the importance of features in input space. However, given the complexity of models and data-generating processes, the resulting explanations are far from being complete, in that they lack an indication of feature interactions and visualization of their effect. In this work, we propose a novel surrogate-model-based explainability framework to explain the decisions of any CNN-based image classifiers by extracting causal relations between the features. These causal relations serve as global explanations from which local explanations of different forms can be obtained. Specifically, we employ a generator to visualize the `effect' of interactions among features in latent space and draw feature importance therefrom as local explanations. We demonstrate and evaluate explanations obtained with our framework on the Morpho-MNIST, the FFHQ, and the AFHQ datasets.
Cite
Text
Kori et al. "GLANCE: Global to Local Architecture-Neutral Concept-Based Explanations." NeurIPS 2023 Workshops: XAIA, 2023.Markdown
[Kori et al. "GLANCE: Global to Local Architecture-Neutral Concept-Based Explanations." NeurIPS 2023 Workshops: XAIA, 2023.](https://mlanthology.org/neuripsw/2023/kori2023neuripsw-glance/)BibTeX
@inproceedings{kori2023neuripsw-glance,
title = {{GLANCE: Global to Local Architecture-Neutral Concept-Based Explanations}},
author = {Kori, Avinash and Glocker, Ben and Toni, Francesca},
booktitle = {NeurIPS 2023 Workshops: XAIA},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/kori2023neuripsw-glance/}
}