Statistically Significant Concept-Based Explanation of Image Classifiers via Model Knockoffs
Abstract
A concept-based classifier can explain the decision process of a deep learning model by human understandable concepts in image classification problems. However, sometimes concept-based explanations may cause false positives, which misregards unrelated concepts as important for the prediction task. Our goal is to find the statistically significant concept for classification to prevent misinterpretation. In this study, we propose a method using a deep learning model to learn the image concept and then using the knockoff sample to select the important concepts for prediction by controlling the False Discovery Rate (FDR) under a certain value. We evaluate the proposed method in our experiments on both synthetic and real data. Also, it shows that our method can control the FDR properly while selecting highly interpretable concepts to improve the trustworthiness of the model.
Cite
Text
Xu et al. "Statistically Significant Concept-Based Explanation of Image Classifiers via Model Knockoffs." International Joint Conference on Artificial Intelligence, 2023. doi:10.24963/IJCAI.2023/58Markdown
[Xu et al. "Statistically Significant Concept-Based Explanation of Image Classifiers via Model Knockoffs." International Joint Conference on Artificial Intelligence, 2023.](https://mlanthology.org/ijcai/2023/xu2023ijcai-statistically/) doi:10.24963/IJCAI.2023/58BibTeX
@inproceedings{xu2023ijcai-statistically,
title = {{Statistically Significant Concept-Based Explanation of Image Classifiers via Model Knockoffs}},
author = {Xu, Kaiwen and Fukuchi, Kazuto and Akimoto, Youhei and Sakuma, Jun},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2023},
pages = {519-526},
doi = {10.24963/IJCAI.2023/58},
url = {https://mlanthology.org/ijcai/2023/xu2023ijcai-statistically/}
}