Links Between Perceptrons, MLPs and SVMs
Abstract
We propose to study links between three important classification algorithms:Perceptrons, Multi-Layer Perceptrons (MLPs) and Support Vector Machines(SVMs). We first study ways to control the capacity of Perceptrons (mainlyregularization parameters and early stopping), using the margin ideaintroduced with SVMs. After showing that under simple conditions a Perceptronis equivalent to an SVM, we show it can be computationally expensive in timeto train an SVM (and thus a Perceptron) with stochastic gradient descent, mainly because of the margin maximization term in the cost function. We thenshow that if we remove this margin maximization term, the learning rate or theuse of early stopping can still control the margin. These ideas are extendedafterward to the case of MLPs. Moreover, under some assumptions it alsoappears that MLPs are a kind of mixture of SVMs, maximizing the margin in thehidden layer space. Finally, we present a very simple MLP based on theprevious findings, which yields better performances in generalization andspeed than the other models.
Cite
Text
Collobert and Bengio. "Links Between Perceptrons, MLPs and SVMs." International Conference on Machine Learning, 2004. doi:10.1145/1015330.1015415Markdown
[Collobert and Bengio. "Links Between Perceptrons, MLPs and SVMs." International Conference on Machine Learning, 2004.](https://mlanthology.org/icml/2004/collobert2004icml-links/) doi:10.1145/1015330.1015415BibTeX
@inproceedings{collobert2004icml-links,
title = {{Links Between Perceptrons, MLPs and SVMs}},
author = {Collobert, Ronan and Bengio, Samy},
booktitle = {International Conference on Machine Learning},
year = {2004},
doi = {10.1145/1015330.1015415},
url = {https://mlanthology.org/icml/2004/collobert2004icml-links/}
}