Sparse Explanations of Neural Networks Using Pruned Layer-Wise Relevance Propagation
Abstract
Explainability is a key component in many applications involving deep neural networks (DNNs). However, current explanation methods for DNNs commonly leave it to the human observer to distinguish relevant explanations from spurious noise. This is not feasible anymore when going from easily human-accessible data such as images to more complex data such as genome sequences. To facilitate the accessibility of DNN outputs from such complex data and to increase explainability, we present a modification of the widely used explanation method layer-wise relevance propagation. Our approach enforces sparsity directly by pruning the relevance propagation for the different layers. Thereby, we achieve sparser relevance attributions for the input features as well as for the intermediate layers. As the relevance propagation is input-specific, we aim to prune the relevance propagation rather than the underlying model architecture. This allows to prune different neurons for different inputs and hence, might be more appropriate to the local nature of explanation methods. To demonstrate the efficacy of our method, we evaluate it on two types of data: images and genome sequences. We show that our modification indeed leads to noise reduction and concentrates relevance on the most important features compared to the baseline.
Cite
Text
Sarmiento et al. "Sparse Explanations of Neural Networks Using Pruned Layer-Wise Relevance Propagation." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2024. doi:10.1007/978-3-031-70359-1_20Markdown
[Sarmiento et al. "Sparse Explanations of Neural Networks Using Pruned Layer-Wise Relevance Propagation." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2024.](https://mlanthology.org/ecmlpkdd/2024/sarmiento2024ecmlpkdd-sparse/) doi:10.1007/978-3-031-70359-1_20BibTeX
@inproceedings{sarmiento2024ecmlpkdd-sparse,
title = {{Sparse Explanations of Neural Networks Using Pruned Layer-Wise Relevance Propagation}},
author = {Sarmiento, Paulo Yanez and Witzke, Simon and Klein, Nadja and Renard, Bernhard Y.},
booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
year = {2024},
pages = {336-351},
doi = {10.1007/978-3-031-70359-1_20},
url = {https://mlanthology.org/ecmlpkdd/2024/sarmiento2024ecmlpkdd-sparse/}
}