Stochastic DCA with Variance Reduction and Applications in Machine Learning

Abstract

We design stochastic Difference-of-Convex-functions Algorithms (DCA) for solving a class of structured Difference-of-Convex-functions (DC) problems. As the standard DCA requires the full information of (sub)gradients which could be expensive in large-scale settings, stochastic approaches rely upon stochastic information instead. However, stochastic estimations generate additional variance terms making stochastic algorithms unstable. Therefore, we integrate some novel variance reduction techniques including SVRG and SAGA into our design. The almost sure convergence to critical points of the proposed algorithms is established and the algorithms' complexities are analyzed. To study the efficiency of our algorithms, we apply them to three important problems in machine learning: nonnegative principal component analysis, group variable selection in multiclass logistic regression, and sparse linear regression. Numerical experiments have shown the merits of our proposed algorithms in comparison with other state-of-the-art stochastic methods for solving nonconvex large-sum problems.

Cite

Text

Le Thi et al. "Stochastic DCA with Variance Reduction and Applications in Machine Learning." Journal of Machine Learning Research, 2022.

Markdown

[Le Thi et al. "Stochastic DCA with Variance Reduction and Applications in Machine Learning." Journal of Machine Learning Research, 2022.](https://mlanthology.org/jmlr/2022/thi2022jmlr-stochastic/)

BibTeX

@article{thi2022jmlr-stochastic,
  title     = {{Stochastic DCA with Variance Reduction and Applications in Machine Learning}},
  author    = {Le Thi, Hoai An and Luu, Hoang Phuc Hau and Le, Hoai Minh and Dinh, Tao Pham},
  journal   = {Journal of Machine Learning Research},
  year      = {2022},
  pages     = {1-44},
  volume    = {23},
  url       = {https://mlanthology.org/jmlr/2022/thi2022jmlr-stochastic/}
}