Proximal Quasi-Newton for Computationally Intensive L1-Regularized M-Estimators
Abstract
We consider the class of optimization problems arising from computationally intensive L1-regularized M-estimators, where the function or gradient values are very expensive to compute. A particular instance of interest is the L1-regularized MLE for learning Conditional Random Fields (CRFs), which are a popular class of statistical models for varied structured prediction problems such as sequence labeling, alignment, and classification with label taxonomy. L1-regularized MLEs for CRFs are particularly expensive to optimize since computing the gradient values requires an expensive inference step. In this work, we propose the use of a carefully constructed proximal quasi-Newton algorithm for such computationally intensive M-estimation problems, where we employ an aggressive active set selection technique. In a key contribution of the paper, we show that our proximal quasi-Newton algorithm is provably super-linearly convergent, even in the absence of strong convexity, by leveraging a restricted variant of strong convexity. In our experiments, the proposed algorithm converges considerably faster than current state-of-the-art on the problems of sequence labeling and hierarchical classification.
Cite
Text
Zhong et al. "Proximal Quasi-Newton for Computationally Intensive L1-Regularized M-Estimators." Neural Information Processing Systems, 2014.Markdown
[Zhong et al. "Proximal Quasi-Newton for Computationally Intensive L1-Regularized M-Estimators." Neural Information Processing Systems, 2014.](https://mlanthology.org/neurips/2014/zhong2014neurips-proximal/)BibTeX
@inproceedings{zhong2014neurips-proximal,
title = {{Proximal Quasi-Newton for Computationally Intensive L1-Regularized M-Estimators}},
author = {Zhong, Kai and Yen, Ian En-Hsu and Dhillon, Inderjit S and Ravikumar, Pradeep K},
booktitle = {Neural Information Processing Systems},
year = {2014},
pages = {2375-2383},
url = {https://mlanthology.org/neurips/2014/zhong2014neurips-proximal/}
}