CF-GNNExplainer: Counterfactual Explanations for Graph Neural Networks

Abstract

Given the increasing promise of graph neural networks (GNNs) in real-world applications, several methods have been developed for explaining their predictions. Existing methods for interpreting predictions from GNNs have primarily focused on generating subgraphs that are especially relevant for a particular prediction. However, such methods are not counterfactual (CF) in nature: given a prediction, we want to understand how the prediction can be changed in order to achieve an alternative outcome. In this work, we propose a method for generating CF explanations for GNNs: the minimal perturbation to the input (graph) data such that the prediction changes. Using only edge deletions, we find that our method, CF-GNNExplainer, can generate CF explanations for the majority of instances across three widely used datasets for GNN explanations, while removing less than 3 edges on average, with at least $94%$ accuracy. This indicates that CF-GNNExplainer primarily removes edges that are crucial for the original predictions, resulting in minimal CF explanations.

Cite

Text

Lucic et al. " CF-GNNExplainer: Counterfactual Explanations for Graph Neural Networks ." Artificial Intelligence and Statistics, 2022.

Markdown

[Lucic et al. " CF-GNNExplainer: Counterfactual Explanations for Graph Neural Networks ." Artificial Intelligence and Statistics, 2022.](https://mlanthology.org/aistats/2022/lucic2022aistats-cfgnnexplainer/)

BibTeX

@inproceedings{lucic2022aistats-cfgnnexplainer,
  title     = {{ CF-GNNExplainer: Counterfactual Explanations for Graph Neural Networks }},
  author    = {Lucic, Ana and Ter Hoeve, Maartje A. and Tolomei, Gabriele and De Rijke, Maarten and Silvestri, Fabrizio},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2022},
  pages     = {4499-4511},
  volume    = {151},
  url       = {https://mlanthology.org/aistats/2022/lucic2022aistats-cfgnnexplainer/}
}