CF-GNNExplainer: Counterfactual Explanations for Graph Neural Networks
Abstract
Given the increasing promise of graph neural networks (GNNs) in real-world applications, several methods have been developed for explaining their predictions. Existing methods for interpreting predictions from GNNs have primarily focused on generating subgraphs that are especially relevant for a particular prediction. However, such methods are not counterfactual (CF) in nature: given a prediction, we want to understand how the prediction can be changed in order to achieve an alternative outcome. In this work, we propose a method for generating CF explanations for GNNs: the minimal perturbation to the input (graph) data such that the prediction changes. Using only edge deletions, we find that our method, CF-GNNExplainer, can generate CF explanations for the majority of instances across three widely used datasets for GNN explanations, while removing less than 3 edges on average, with at least $94%$ accuracy. This indicates that CF-GNNExplainer primarily removes edges that are crucial for the original predictions, resulting in minimal CF explanations.
Cite
Text
Lucic et al. " CF-GNNExplainer: Counterfactual Explanations for Graph Neural Networks ." Artificial Intelligence and Statistics, 2022.Markdown
[Lucic et al. " CF-GNNExplainer: Counterfactual Explanations for Graph Neural Networks ." Artificial Intelligence and Statistics, 2022.](https://mlanthology.org/aistats/2022/lucic2022aistats-cfgnnexplainer/)BibTeX
@inproceedings{lucic2022aistats-cfgnnexplainer,
title = {{ CF-GNNExplainer: Counterfactual Explanations for Graph Neural Networks }},
author = {Lucic, Ana and Ter Hoeve, Maartje A. and Tolomei, Gabriele and De Rijke, Maarten and Silvestri, Fabrizio},
booktitle = {Artificial Intelligence and Statistics},
year = {2022},
pages = {4499-4511},
volume = {151},
url = {https://mlanthology.org/aistats/2022/lucic2022aistats-cfgnnexplainer/}
}