Two-Stage Holistic and Contrastive Explanation of Image Classification

Abstract

The need to explain the output of a deep neural network classifier is now widely recognized. While previous methods typically explain a single class in the output, we advocate explaining the whole output, which is a probability distribution over multiple classes. A whole-output explanation can help a human user gain an overall understanding of model behaviour instead of only one aspect of it. It can also provide a natural framework where one can examine the evidence used to discriminate between competing classes, and thereby obtain contrastive explanations. In this paper, we propose a contrastive whole-output explanation (CWOX) method for image classification, and evaluate it using quantitative metrics and through human subject studies. The source code of CWOX is available at https://github.com/vaynexie/CWOX.

Cite

Text

Xie et al. "Two-Stage Holistic and Contrastive Explanation of Image Classification." Uncertainty in Artificial Intelligence, 2023.

Markdown

[Xie et al. "Two-Stage Holistic and Contrastive Explanation of Image Classification." Uncertainty in Artificial Intelligence, 2023.](https://mlanthology.org/uai/2023/xie2023uai-twostage/)

BibTeX

@inproceedings{xie2023uai-twostage,
  title     = {{Two-Stage Holistic and Contrastive Explanation of Image Classification}},
  author    = {Xie, Weiyan and Li, Xiao-Hui and Lin, Zhi and Poon, Leonard K. M. and Cao, Caleb Chen and Zhang, Nevin L.},
  booktitle = {Uncertainty in Artificial Intelligence},
  year      = {2023},
  pages     = {2335-2345},
  volume    = {216},
  url       = {https://mlanthology.org/uai/2023/xie2023uai-twostage/}
}