"Why Did the Model Fail?": Attributing Model Performance Changes to Distribution Shifts

Abstract

Performance of machine learning models may differ significantly in novel environments compared to during training due to shifts in the underlying data distribution. Attributing performance changes to specific data shifts is critical for identifying sources of model failures and designing stable models. In this work, we design a novel method for attributing performance differences between environments to shifts in the underlying causal mechanisms. We formulate the problem as a cooperative game and derive an importance weighting method for computing the value of a coalition of distributions. The contribution of each distribution to the total performance change is then quantified as its Shapley value. We demonstrate the correctness and utility of our method on two synthetic datasets and two real-world case studies, showing its effectiveness in attributing performance changes to a wide range of distribution shifts.

Cite

Text

Zhang et al. ""Why Did the Model Fail?": Attributing Model Performance Changes to Distribution Shifts." NeurIPS 2022 Workshops: DistShift, 2022.

Markdown

[Zhang et al. ""Why Did the Model Fail?": Attributing Model Performance Changes to Distribution Shifts." NeurIPS 2022 Workshops: DistShift, 2022.](https://mlanthology.org/neuripsw/2022/zhang2022neuripsw-model/)

BibTeX

@inproceedings{zhang2022neuripsw-model,
  title     = {{"Why Did the Model Fail?": Attributing Model Performance Changes to Distribution Shifts}},
  author    = {Zhang, Haoran and Singh, Harvineet and Ghassemi, Marzyeh and Joshi, Shalmali},
  booktitle = {NeurIPS 2022 Workshops: DistShift},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/zhang2022neuripsw-model/}
}