Curriculum Learning

Abstract

Humans and animals learn much better when the examples are not randomly presented but organized in a meaningful order which illustrates gradually more concepts, and more complex ones. Here, we formalize such training strategies in the context of machine learning, and call them "curriculum learning". In the context of recent research studying the difficulty of training in the presence of non-convex training criteria (for deep deterministic and stochastic neural networks), we explore curriculum learning in various set-ups. The experiments show that significant improvements in generalization can be achieved by using a particular curriculum, i.e., the selection and order of training examples. We hypothesize that curriculum learning has both an effect on the speed of convergence of the training process to a minimum and, in the case of non-convex criteria, on the quality of the local minima obtained: curriculum learning can be seen as a particular form of continuation method (a general strategy for global optimization of non-convex functions).

Cite

Text

Bengio et al. "Curriculum Learning." International Conference on Machine Learning, 2009. doi:10.1145/1553374.1553380

Markdown

[Bengio et al. "Curriculum Learning." International Conference on Machine Learning, 2009.](https://mlanthology.org/icml/2009/bengio2009icml-curriculum/) doi:10.1145/1553374.1553380

BibTeX

@inproceedings{bengio2009icml-curriculum,
  title     = {{Curriculum Learning}},
  author    = {Bengio, Yoshua and Louradour, Jérôme and Collobert, Ronan and Weston, Jason},
  booktitle = {International Conference on Machine Learning},
  year      = {2009},
  pages     = {41-48},
  doi       = {10.1145/1553374.1553380},
  url       = {https://mlanthology.org/icml/2009/bengio2009icml-curriculum/}
}