A Functional Information Perspective on Model Interpretation
Abstract
Contemporary predictive models are hard to interpret as their deep nets exploit numerous complex relations between input elements. This work suggests a theoretical framework for model interpretability by measuring the contribution of relevant features to the functional entropy of the network with respect to the input. We rely on the log-Sobolev inequality that bounds the functional entropy by the functional Fisher information with respect to the covariance of the data. This provides a principled way to measure the amount of information contribution of a subset of features to the decision function. Through extensive experiments, we show that our method surpasses existing interpretability sampling-based methods on various data signals such as image, text, and audio.
Cite
Text
Gat et al. "A Functional Information Perspective on Model Interpretation." International Conference on Machine Learning, 2022.Markdown
[Gat et al. "A Functional Information Perspective on Model Interpretation." International Conference on Machine Learning, 2022.](https://mlanthology.org/icml/2022/gat2022icml-functional/)BibTeX
@inproceedings{gat2022icml-functional,
title = {{A Functional Information Perspective on Model Interpretation}},
author = {Gat, Itai and Calderon, Nitay and Reichart, Roi and Hazan, Tamir},
booktitle = {International Conference on Machine Learning},
year = {2022},
pages = {7266-7278},
volume = {162},
url = {https://mlanthology.org/icml/2022/gat2022icml-functional/}
}