Dual Coordinate Descent Methods for Logistic Regression and Maximum Entropy Models

Abstract

Most optimization methods for logistic regression or maximum entropy solve the primal problem. They range from iterative scaling, coordinate descent, quasi-Newton, and truncated Newton. Less efforts have been made to solve the dual problem. In contrast, for linear support vector machines ( SVM ), methods have been shown to be very effective for solving the dual problem. In this paper, we apply coordinate descent methods to solve the dual form of logistic regression and maximum entropy. Interestingly, many details are different from the situation in linear SVM . We carefully study the theoretical convergence as well as numerical issues. The proposed method is shown to be faster than most state of the art methods for training logistic regression and maximum entropy.

Cite

Text

Yu et al. "Dual Coordinate Descent Methods for Logistic Regression and Maximum Entropy Models." Machine Learning, 2011. doi:10.1007/S10994-010-5221-8

Markdown

[Yu et al. "Dual Coordinate Descent Methods for Logistic Regression and Maximum Entropy Models." Machine Learning, 2011.](https://mlanthology.org/mlj/2011/yu2011mlj-dual/) doi:10.1007/S10994-010-5221-8

BibTeX

@article{yu2011mlj-dual,
  title     = {{Dual Coordinate Descent Methods for Logistic Regression and Maximum Entropy Models}},
  author    = {Yu, Hsiang-Fu and Huang, Fang-Lan and Lin, Chih-Jen},
  journal   = {Machine Learning},
  year      = {2011},
  pages     = {41-75},
  doi       = {10.1007/S10994-010-5221-8},
  volume    = {85},
  url       = {https://mlanthology.org/mlj/2011/yu2011mlj-dual/}
}