Learning Dynamics of Attention: Human Prior for Interpretable Machine Reasoning
Abstract
Without relevant human priors, neural networks may learn uninterpretable features. We propose Dynamics of Attention for Focus Transition (DAFT) as a human prior for machine reasoning. DAFT is a novel method that regularizes attention-based reasoning by modelling it as a continuous dynamical system using neural ordinary differential equations. As a proof of concept, we augment a state-of-the-art visual reasoning model with DAFT. Our experiments reveal that applying DAFT yields similar performance to the original model while using fewer reasoning steps, showing that it implicitly learns to skip unnecessary steps. We also propose a new metric, Total Length of Transition (TLT), which represents the effective reasoning step size by quantifying how much a given model's focus drifts while reasoning about a question. We show that adding DAFT results in lower TLT, demonstrating that our method indeed obeys the human prior towards shorter reasoning paths in addition to producing more interpretable attention maps.
Cite
Text
Kim and Lee. "Learning Dynamics of Attention: Human Prior for Interpretable Machine Reasoning." Neural Information Processing Systems, 2019.Markdown
[Kim and Lee. "Learning Dynamics of Attention: Human Prior for Interpretable Machine Reasoning." Neural Information Processing Systems, 2019.](https://mlanthology.org/neurips/2019/kim2019neurips-learning/)BibTeX
@inproceedings{kim2019neurips-learning,
title = {{Learning Dynamics of Attention: Human Prior for Interpretable Machine Reasoning}},
author = {Kim, Wonjae and Lee, Yoonho},
booktitle = {Neural Information Processing Systems},
year = {2019},
pages = {6021-6032},
url = {https://mlanthology.org/neurips/2019/kim2019neurips-learning/}
}