A Fast Natural Newton Method
Abstract
Nowadays, for many tasks such as object recognition or language modeling, data is plentiful. As such, an important challenge has become to find learning algorithms which can make use of all the available data. In this setting, called ``large-scale learning'' by Bottou and Bousquet (2008), learning and optimization become different and powerful optimization algorithms are suboptimal learning algorithms. While most efforts are focused on adapting optimization algorithms for learning by efficiently using the information contained in the Hessian, Le Roux et al. (2008) exploited the special structure of the learning problem to achieve faster convergence. In this paper, we investigate a natural way of combining these two directions to yield fast and robust learning algorithms.
Cite
Text
Le Roux and Fitzgibbon. "A Fast Natural Newton Method." International Conference on Machine Learning, 2010.Markdown
[Le Roux and Fitzgibbon. "A Fast Natural Newton Method." International Conference on Machine Learning, 2010.](https://mlanthology.org/icml/2010/roux2010icml-fast/)BibTeX
@inproceedings{roux2010icml-fast,
title = {{A Fast Natural Newton Method}},
author = {Le Roux, Nicolas and Fitzgibbon, Andrew W.},
booktitle = {International Conference on Machine Learning},
year = {2010},
pages = {623-630},
url = {https://mlanthology.org/icml/2010/roux2010icml-fast/}
}