DeDUCE: Generating Counterfactual Explanations at Scale
Abstract
When an image classifier outputs a wrong class label, it can be helpful to see what changes in the image would lead to a correct classification. This is the aim of algorithms generating counterfactual explanations. However, there is no easily scalable method to generate such counterfactuals. We develop a new algorithm providing counterfactual explanations for large image classifiers trained with spectral normalisation at low computational cost. We empirically compare this algorithm against baselines from the literature; our novel algorithm consistently finds counterfactuals that are much closer to the original inputs. At the same time, the realism of these counterfactuals is comparable to the baselines.
Cite
Text
Höltgen et al. "DeDUCE: Generating Counterfactual Explanations at Scale." NeurIPS 2021 Workshops: XAI4Debugging, 2021.Markdown
[Höltgen et al. "DeDUCE: Generating Counterfactual Explanations at Scale." NeurIPS 2021 Workshops: XAI4Debugging, 2021.](https://mlanthology.org/neuripsw/2021/holtgen2021neuripsw-deduce/)BibTeX
@inproceedings{holtgen2021neuripsw-deduce,
title = {{DeDUCE: Generating Counterfactual Explanations at Scale}},
author = {Höltgen, Benedikt and Schut, Lisa and Brauner, Jan M. and Gal, Yarin},
booktitle = {NeurIPS 2021 Workshops: XAI4Debugging},
year = {2021},
url = {https://mlanthology.org/neuripsw/2021/holtgen2021neuripsw-deduce/}
}