HIVE: Evaluating the Human Interpretability of Visual Explanations

Abstract

As AI technology is increasingly applied to high-impact, high-risk domains, there have been a number of new methods aimed at making AI models more human interpretable. Despite the recent growth of interpretability work, there is a lack of systematic evaluation of proposed techniques. In this work, we introduce HIVE (Human Interpretability of Visual Explanations), a novel human evaluation framework that assesses the utility of explanations to human users in AI-assisted decision making scenarios, and enables falsifiable hypothesis testing, cross-method comparison, and human-centered evaluation of visual interpretability methods. To the best of our knowledge, this is the first work of its kind. Using HIVE, we conduct IRB-approved human studies with nearly 1000 participants and evaluate four methods that represent the diversity of computer vision interpretability works: GradCAM, BagNet, ProtoPNet, and ProtoTree. Our results suggest that explanations engender human trust, even for incorrect predictions, yet are not distinct enough for users to distinguish between correct and incorrect predictions. We open-source HIVE to enable future studies and encourage more human-centered approaches to interpretability research.

Cite

Text

Kim et al. "HIVE: Evaluating the Human Interpretability of Visual Explanations." Proceedings of the European Conference on Computer Vision (ECCV), 2022. doi:10.1007/978-3-031-19775-8_17

Markdown

[Kim et al. "HIVE: Evaluating the Human Interpretability of Visual Explanations." Proceedings of the European Conference on Computer Vision (ECCV), 2022.](https://mlanthology.org/eccv/2022/kim2022eccv-hive/) doi:10.1007/978-3-031-19775-8_17

BibTeX

@inproceedings{kim2022eccv-hive,
  title     = {{HIVE: Evaluating the Human Interpretability of Visual Explanations}},
  author    = {Kim, Sunnie S. Y. and Meister, Nicole and Ramaswamy, Vikram V. and Fong, Ruth and Russakovsky, Olga},
  booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
  year      = {2022},
  doi       = {10.1007/978-3-031-19775-8_17},
  url       = {https://mlanthology.org/eccv/2022/kim2022eccv-hive/}
}