Redundancy Undermines the Trustworthiness of Self-Interpretable GNNs

Abstract

This work presents a systematic investigation into the trustworthiness of explanations generated by self-interpretable graph neural networks (GNNs), revealing why models trained with different random seeds yield inconsistent explanations. We identify redundancy—resulting from weak conciseness constraints—as the root cause of both explanation inconsistency and its associated inaccuracy, ultimately hindering user trust and limiting GNN deployment in high-stakes applications. Our analysis demonstrates that redundancy is difficult to eliminate; however, a simple ensemble strategy can mitigate its detrimental effects. We validate our findings through extensive experiments across diverse datasets, model architectures, and self-interpretable GNN frameworks, providing a benchmark to guide future research on addressing redundancy and advancing GNN deployment in critical domains. Our code is available at https://github.com/ICDM-UESTC/TrustworthyExplanation.

Cite

Text

Tai et al. "Redundancy Undermines the Trustworthiness of Self-Interpretable GNNs." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Tai et al. "Redundancy Undermines the Trustworthiness of Self-Interpretable GNNs." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/tai2025icml-redundancy/)

BibTeX

@inproceedings{tai2025icml-redundancy,
  title     = {{Redundancy Undermines the Trustworthiness of Self-Interpretable GNNs}},
  author    = {Tai, Wenxin and Zhong, Ting and Trajcevski, Goce and Zhou, Fan},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {58169-58188},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/tai2025icml-redundancy/}
}