Dual Coordinate Descent Methods for Logistic Regression and Maximum Entropy Models
Abstract
Most optimization methods for logistic regression or maximum entropy solve the primal problem. They range from iterative scaling, coordinate descent, quasi-Newton, and truncated Newton. Less efforts have been made to solve the dual problem. In contrast, for linear support vector machines ( SVM ), methods have been shown to be very effective for solving the dual problem. In this paper, we apply coordinate descent methods to solve the dual form of logistic regression and maximum entropy. Interestingly, many details are different from the situation in linear SVM . We carefully study the theoretical convergence as well as numerical issues. The proposed method is shown to be faster than most state of the art methods for training logistic regression and maximum entropy.
Cite
Text
Yu et al. "Dual Coordinate Descent Methods for Logistic Regression and Maximum Entropy Models." Machine Learning, 2011. doi:10.1007/S10994-010-5221-8Markdown
[Yu et al. "Dual Coordinate Descent Methods for Logistic Regression and Maximum Entropy Models." Machine Learning, 2011.](https://mlanthology.org/mlj/2011/yu2011mlj-dual/) doi:10.1007/S10994-010-5221-8BibTeX
@article{yu2011mlj-dual,
title = {{Dual Coordinate Descent Methods for Logistic Regression and Maximum Entropy Models}},
author = {Yu, Hsiang-Fu and Huang, Fang-Lan and Lin, Chih-Jen},
journal = {Machine Learning},
year = {2011},
pages = {41-75},
doi = {10.1007/S10994-010-5221-8},
volume = {85},
url = {https://mlanthology.org/mlj/2011/yu2011mlj-dual/}
}