Geometric Interpretation of a CNN's Last Layer
Abstract
Training Convolutional Neural Networks (CNNs) remains a non-trivial task that in many cases relies on the skills and experience of the person conducting the training. Choosing hyperparameters, knowing when the training should be interrupted, or even when to stop trying training strategies are some difficult decisions that have to be made. These decisions are difficult partly because we still know little about the internal behaviour of CNNs, especially during training. In this work we conduct a methodical experimentation on MNIST public database of handwritten digits to better understand the evolution of the last layer from a geometric perspective: namely the classification vectors and the image embedding vectors. Within this context we present the problem of the variability across equal set-up trainings due to the random component of the initialisation method. We propose a novel approach that guides the initialisation of the parameters in the classification layer. This method reduces 12% the variability across repetitions and leads to accuracies 18% higher on average.
Cite
Text
de la Calle et al. "Geometric Interpretation of a CNN's Last Layer." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019.Markdown
[de la Calle et al. "Geometric Interpretation of a CNN's Last Layer." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019.](https://mlanthology.org/cvprw/2019/delacalle2019cvprw-geometric/)BibTeX
@inproceedings{delacalle2019cvprw-geometric,
title = {{Geometric Interpretation of a CNN's Last Layer}},
author = {de la Calle, Alejandro and Aller, Aitor and Tovar, Javier and Almazán, Emilio J.},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2019},
pages = {79-82},
url = {https://mlanthology.org/cvprw/2019/delacalle2019cvprw-geometric/}
}