Usable Information and Evolution of Optimal Representations During Training
Abstract
We introduce a notion of usable information contained in the representation learned by a deep network, and use it to study how optimal representations for the task emerge during training, and how they adapt to different tasks. We use this to characterize the transient dynamics of deep neural networks on perceptual decision-making tasks inspired by neuroscience literature. In particular, we show that both the random initialization and the implicit regularization from Stochastic Gradient Descent play an important role in learning minimal sufficient representations for the task. If the network is not randomly initialized, we show that the training may not recover an optimal representation, increasing the chance of overfitting.
Cite
Text
Kleinman et al. "Usable Information and Evolution of Optimal Representations During Training." NeurIPS 2020 Workshops: SVRHM, 2020.Markdown
[Kleinman et al. "Usable Information and Evolution of Optimal Representations During Training." NeurIPS 2020 Workshops: SVRHM, 2020.](https://mlanthology.org/neuripsw/2020/kleinman2020neuripsw-usable/)BibTeX
@inproceedings{kleinman2020neuripsw-usable,
title = {{Usable Information and Evolution of Optimal Representations During Training}},
author = {Kleinman, Michael and Idnani, Daksh and Achille, Alessandro and Kao, Jonathan},
booktitle = {NeurIPS 2020 Workshops: SVRHM},
year = {2020},
url = {https://mlanthology.org/neuripsw/2020/kleinman2020neuripsw-usable/}
}