Interpreting Interpretations: Organizing Attribution Methods by Criteria
Abstract
Motivated by distinct, though related, criteria, a growing number of attribution methods have been developed to interprete deep learning. While each relies on the interpretability of the concept of "importance" and our ability to visualize patterns, explanations produced by the methods often differ. In this work we expand the foundations of human-understandable concepts with which attributions can be interpreted beyond "importance" and its visualization; we incorporate the logical concepts of necessity and sufficiency, and the concept of proportionality. We define metrics to represent these concepts as quantitative aspects of an attribution. We evaluate our measures on a collection of methods explaining convolutional neural networks (CNN) for image classification. We conclude that some attribution methods are more appropriate for interpretation in terms of necessity while others are in terms of sufficiency, while no method is always the most appropriate in terms of both.
Cite
Text
Wang et al. "Interpreting Interpretations: Organizing Attribution Methods by Criteria." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020. doi:10.1109/CVPRW50498.2020.00013Markdown
[Wang et al. "Interpreting Interpretations: Organizing Attribution Methods by Criteria." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020.](https://mlanthology.org/cvprw/2020/wang2020cvprw-interpreting/) doi:10.1109/CVPRW50498.2020.00013BibTeX
@inproceedings{wang2020cvprw-interpreting,
title = {{Interpreting Interpretations: Organizing Attribution Methods by Criteria}},
author = {Wang, Zifan and Mardziel, Piotr and Datta, Anupam and Fredrikson, Matt},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2020},
pages = {48-55},
doi = {10.1109/CVPRW50498.2020.00013},
url = {https://mlanthology.org/cvprw/2020/wang2020cvprw-interpreting/}
}