Information Bottleneck for Multi-Task LSTMs

Abstract

Neural networks, which have had a profound effect on how researchers study complex phenomena, do so through a complex, nonlinear mathematical structure which can be difficult to interpret or understand. This is especially true for recurrent models, as their dynamic structure can be difficult to measure and analyze. However, interpretability is a key factor in understanding certain problems such as text and language analysis. In this paper, we present a novel introspection method for LSTMs trained to solve complex language problems, such as sentiment analysis. Inspired by Information Bottleneck theory, our method uses a state-of-the-art information theoretic framework to visualize shared information around labels, features, and between layers. We apply our approach on simulated data, and real sentiment analysis datasets, providing novel, information-theoretic insights into internal model dynamics.

Cite

Text

Baker et al. "Information Bottleneck for Multi-Task LSTMs." NeurIPS 2022 Workshops: InfoCog, 2022.

Markdown

[Baker et al. "Information Bottleneck for Multi-Task LSTMs." NeurIPS 2022 Workshops: InfoCog, 2022.](https://mlanthology.org/neuripsw/2022/baker2022neuripsw-information/)

BibTeX

@inproceedings{baker2022neuripsw-information,
  title     = {{Information Bottleneck for Multi-Task LSTMs}},
  author    = {Baker, Bradley Thomas and Lewis, Noah and Saha, Debratta and Rahaman, Md Abdur and Plis, Sergey and Calhoun, Vince},
  booktitle = {NeurIPS 2022 Workshops: InfoCog},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/baker2022neuripsw-information/}
}