Adaptive Momentum Coefficient for Neural Network Optimization

Abstract

We propose a novel and efficient momentum-based first-order algorithm for optimizing neural networks which uses an adaptive coefficient for the momentum term. Our algorithm, called Adaptive Momentum Coefficient (AMoC), utilizes the inner product of the gradient and the previous update to the parameters, to effectively control the amount of weight put on the momentum term based on the change of direction in the optimization path. The algorithm is easy to implement and its computational overhead over momentum methods is negligible. Extensive empirical results on both convex and neural network objectives show that AMoC performs well in practise and compares favourably with other first and second-order optimization algorithms. We also provide a convergence analysis and a convergence rate for AMoC, showing theoretical guarantees similar to those provided by other efficient first-order methods.

Cite

Text

Rashidi et al. "Adaptive Momentum Coefficient for Neural Network Optimization." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2020. doi:10.1007/978-3-030-67661-2_3

Markdown

[Rashidi et al. "Adaptive Momentum Coefficient for Neural Network Optimization." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2020.](https://mlanthology.org/ecmlpkdd/2020/rashidi2020ecmlpkdd-adaptive/) doi:10.1007/978-3-030-67661-2_3

BibTeX

@inproceedings{rashidi2020ecmlpkdd-adaptive,
  title     = {{Adaptive Momentum Coefficient for Neural Network Optimization}},
  author    = {Rashidi, Zana and Kasra Ahmadi, K. A. and An, Aijun and Wang, Xiaogang},
  booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
  year      = {2020},
  pages     = {35-51},
  doi       = {10.1007/978-3-030-67661-2_3},
  url       = {https://mlanthology.org/ecmlpkdd/2020/rashidi2020ecmlpkdd-adaptive/}
}