ObEy: Quantifiable Object-Based Explainability Without Ground-Truth Annotations

Abstract

Neural networks are at the core of AI systems recently observing accelerated adoption in high-stakes environments. Consequently, understanding their black-box predictive behavior is paramount. Current explainable AI techniques, however, are limited to explaining a single prediction, rather than characterizing the inherent ability of the model to be explained, reducing their usefulness to manual inspection of samples. In this work, we offer a conceptual distinction between explanation methods and explainability. We use this motivation to propose Object-based Explainability (ObEy), a novel model explainability metric that collectively assesses model-produced saliency maps relative to objects in images, inspired by humans’ perception of scenes. To render ObEy independent of the prediction task, we use full-image instance segmentations obtained from a foundation model, making the metric applicable on existing models in any setting. We demonstrate ObEy’s immediate applicability to use cases in model inspection and comparison. As a result, we present new insights into the explainability of adversarially trained models from a quantitative perspective.

Cite

Text

Schulze et al. "ObEy: Quantifiable Object-Based Explainability Without Ground-Truth Annotations." NeurIPS 2023 Workshops: XAIA, 2023.

Markdown

[Schulze et al. "ObEy: Quantifiable Object-Based Explainability Without Ground-Truth Annotations." NeurIPS 2023 Workshops: XAIA, 2023.](https://mlanthology.org/neuripsw/2023/schulze2023neuripsw-obey/)

BibTeX

@inproceedings{schulze2023neuripsw-obey,
  title     = {{ObEy: Quantifiable Object-Based Explainability Without Ground-Truth Annotations}},
  author    = {Schulze, Lennart and Ho, William and Zemel, Richard},
  booktitle = {NeurIPS 2023 Workshops: XAIA},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/schulze2023neuripsw-obey/}
}