CF-OPT: Counterfactual Explanations for Structured Prediction

Abstract

Optimization layers in deep neural networks have enjoyed a growing popularity in structured learning, improving the state of the art on a variety of applications. Yet, these pipelines lack interpretability since they are made of two opaque layers: a highly non-linear prediction model, such as a deep neural network, and an optimization layer, which is typically a complex black-box solver. Our goal is to improve the transparency of such methods by providing counterfactual explanations. We build upon variational autoencoders a principled way of obtaining counterfactuals: working in the latent space leads to a natural notion of plausibility of explanations. We finally introduce a variant of the classic loss for VAE training that improves their performance in our specific structured context. These provide the foundations of CF-OPT, a first-order optimization algorithm that can find counterfactual explanations for a broad class of structured learning architectures. Our numerical results show that both close and plausible explanations can be obtained for problems from the recent literature.

Cite

Text

Vivier-Ardisson et al. "CF-OPT: Counterfactual Explanations for Structured Prediction." International Conference on Machine Learning, 2024.

Markdown

[Vivier-Ardisson et al. "CF-OPT: Counterfactual Explanations for Structured Prediction." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/vivierardisson2024icml-cfopt/)

BibTeX

@inproceedings{vivierardisson2024icml-cfopt,
  title     = {{CF-OPT: Counterfactual Explanations for Structured Prediction}},
  author    = {Vivier-Ardisson, Germain and Forel, Alexandre and Parmentier, Axel and Vidal, Thibaut},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {49558-49579},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/vivierardisson2024icml-cfopt/}
}