RKHS-SHAP: Shapley Values for Kernel Methods

Abstract

Feature attribution for kernel methods is often heuristic and not individualised for each prediction. To address this, we turn to the concept of Shapley values (SV), a coalition game theoretical framework that has previously been applied to different machine learning model interpretation tasks, such as linear models, tree ensembles and deep networks. By analysing SVs from a functional perspective, we propose RKHS-SHAP, an attribution method for kernel machines that can efficiently compute both Interventional and Observational Shapley values using kernel mean embeddings of distributions. We show theoretically that our method is robust with respect to local perturbations - a key yet often overlooked desideratum for consistent model interpretation. Further, we propose Shapley regulariser, applicable to a general empirical risk minimisation framework, allowing learning while controlling the level of specific feature's contributions to the model. We demonstrate that the Shapley regulariser enables learning which is robust to covariate shift of a given feature and fair learning which controls the SVs of sensitive features.

Cite

Text

Chau et al. "RKHS-SHAP: Shapley Values for Kernel Methods." Neural Information Processing Systems, 2022.

Markdown

[Chau et al. "RKHS-SHAP: Shapley Values for Kernel Methods." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/chau2022neurips-rkhsshap/)

BibTeX

@inproceedings{chau2022neurips-rkhsshap,
  title     = {{RKHS-SHAP: Shapley Values for Kernel Methods}},
  author    = {Chau, Siu Lun and Hu, Robert and González, Javier and Sejdinovic, Dino},
  booktitle = {Neural Information Processing Systems},
  year      = {2022},
  url       = {https://mlanthology.org/neurips/2022/chau2022neurips-rkhsshap/}
}