Generalisation and the Geometry of Class Separability
Abstract
Recent results in deep learning show that considering only the capacity of machines does not adequately explain the generalisation performance we can observe. We propose that by considering the geometry of the data we can better explain generalisation achieved in deep learning. In particular we show that in classification the separability of the data can explain how good generalisation can be achieved in high dimensions. Further we show that layers within a CNNs sequentially increase the linear separability of data, and that the information these layers retain or discard can help explain why these models generalise.
Cite
Text
Belcher et al. "Generalisation and the Geometry of Class Separability." NeurIPS 2020 Workshops: DL-IG, 2020.Markdown
[Belcher et al. "Generalisation and the Geometry of Class Separability." NeurIPS 2020 Workshops: DL-IG, 2020.](https://mlanthology.org/neuripsw/2020/belcher2020neuripsw-generalisation/)BibTeX
@inproceedings{belcher2020neuripsw-generalisation,
title = {{Generalisation and the Geometry of Class Separability}},
author = {Belcher, Dominic and Prugel-Bennett, Adam and Dasmahapatra, Srinandan},
booktitle = {NeurIPS 2020 Workshops: DL-IG},
year = {2020},
url = {https://mlanthology.org/neuripsw/2020/belcher2020neuripsw-generalisation/}
}