AR-Pro: Counterfactual Explanations for Anomaly Repair with Formal Properties
Abstract
Anomaly detection is widely used for identifying critical errors and suspicious behaviors, but current methods lack interpretability.We leverage common properties of existing methods and recent advances in generative models to introduce counterfactual explanations for anomaly detection.Given an input, we generate its counterfactual as a diffusion-based repair that shows what a non-anomalous version $\textit{should have looked like}$.A key advantage of this approach is that it enables a domain-independent formal specification of explainability desiderata, offering a unified framework for generating and evaluating explanations.We demonstrate the effectiveness of our anomaly explainability framework, AR-Pro, on vision (MVTec, VisA) and time-series (SWaT, WADI, HAI) anomaly datasets. The code used for the experiments is accessible at: https://github.com/xjiae/arpro.
Cite
Text
Ji et al. "AR-Pro: Counterfactual Explanations for Anomaly Repair with Formal Properties." Neural Information Processing Systems, 2024. doi:10.52202/079017-0514Markdown
[Ji et al. "AR-Pro: Counterfactual Explanations for Anomaly Repair with Formal Properties." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/ji2024neurips-arpro/) doi:10.52202/079017-0514BibTeX
@inproceedings{ji2024neurips-arpro,
title = {{AR-Pro: Counterfactual Explanations for Anomaly Repair with Formal Properties}},
author = {Ji, Xiayan and Xue, Anton and Wong, Eric and Sokolsky, Oleg and Lee, Insup},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-0514},
url = {https://mlanthology.org/neurips/2024/ji2024neurips-arpro/}
}