Aligning Characteristic Descriptors with Images for Human-Expert-like Explainability
Abstract
In mission-critical domains such as law enforcement and medical diagnosis, the ability to explain and interpret the outputs of deep learning models is crucial for en- suring user trust and supporting informed decision-making. Despite advancements in explainability, existing methods often fall short in providing explanations that mirror the depth and clarity of those given by human experts. Such expert-level explanations are essential for the dependable application of deep learning models in law enforcement and medical contexts. Additionally, we recognize that most explanations in real-world scenarios are communicated primarily through natu- ral language. Addressing these needs, we propose a novel approach that utilizes characteristic descriptors to explain model decisions by identifying their presence in images, thereby generating expert-like explanations. Our method incorporates a concept bottleneck layer within the model architecture, which calculates the similarity between image and descriptor encodings to deliver inherent and faithful explanations. Through experiments in face recognition and chest X-ray diagnosis, we show that our approach offers a significant contrast over existing techniques, which are often limited to the use of saliency maps. Our approach represents a sig- nificant step toward making deep learning systems more accountable, transparent, and trustworthy in the critical domains of face recognition and medical diagnosis.
Cite
Text
Yalavarthi and Ratha. "Aligning Characteristic Descriptors with Images for Human-Expert-like Explainability." NeurIPS 2024 Workshops: InterpretableAI, 2024.Markdown
[Yalavarthi and Ratha. "Aligning Characteristic Descriptors with Images for Human-Expert-like Explainability." NeurIPS 2024 Workshops: InterpretableAI, 2024.](https://mlanthology.org/neuripsw/2024/yalavarthi2024neuripsw-aligning/)BibTeX
@inproceedings{yalavarthi2024neuripsw-aligning,
title = {{Aligning Characteristic Descriptors with Images for Human-Expert-like Explainability}},
author = {Yalavarthi, Bharat Chandra and Ratha, Nalini K.},
booktitle = {NeurIPS 2024 Workshops: InterpretableAI},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/yalavarthi2024neuripsw-aligning/}
}