Regional Tree Regularization for Interpretability in Deep Neural Networks

Abstract

The lack of interpretability remains a barrier to adopting deep neural networks across many safety-critical domains. Tree regularization was recently proposed to encourage a deep neural network's decisions to resemble those of a globally compact, axis-aligned decision tree. However, it is often unreasonable to expect a single tree to predict well across all possible inputs. In practice, doing so could lead to neither interpretable nor performant optima. To address this issue, we propose regional tree regularization – a method that encourages a deep model to be well-approximated by several separate decision trees specific to predefined regions of the input space. Across many datasets, including two healthcare applications, we show our approach delivers simpler explanations than other regularization schemes without compromising accuracy. Specifically, our regional regularizer finds many more “desirable” optima compared to global analogues.

Cite

Text

Wu et al. "Regional Tree Regularization for Interpretability in Deep Neural Networks." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I04.6112

Markdown

[Wu et al. "Regional Tree Regularization for Interpretability in Deep Neural Networks." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/wu2020aaai-regional/) doi:10.1609/AAAI.V34I04.6112

BibTeX

@inproceedings{wu2020aaai-regional,
  title     = {{Regional Tree Regularization for Interpretability in Deep Neural Networks}},
  author    = {Wu, Mike and Parbhoo, Sonali and Hughes, Michael C. and Kindle, Ryan and Celi, Leo A. and Zazzi, Maurizio and Roth, Volker and Doshi-Velez, Finale},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2020},
  pages     = {6413-6421},
  doi       = {10.1609/AAAI.V34I04.6112},
  url       = {https://mlanthology.org/aaai/2020/wu2020aaai-regional/}
}