An Accelerated Gradient Method for Trace Norm Minimization
Abstract
We consider the minimization of a smooth loss function regularized by the trace norm of the matrix variable. Such formulation finds applications in many machine learning tasks including multi-task learning, matrix classification, and matrix completion. The standard semidefinite programming formulation for this problem is computationally expensive. In addition, due to the non-smoothness nature of the trace norm, the optimal first-order black-box method for solving such class of problems converges as O(1/sqrt(k)), where k is the iteration counter. In this paper, we exploit the special structure of the trace norm, based on which we propose an extended gradient algorithm that converges as O(1/k). We further propose an accelerated gradient algorithm, which achieves the optimal convergence rate of O(1/k^2) for smooth problems. Experiments on multi-task learning problems demonstrate the efficiency of the proposed algorithms.
Cite
Text
Ji and Ye. "An Accelerated Gradient Method for Trace Norm Minimization." International Conference on Machine Learning, 2009. doi:10.1145/1553374.1553434Markdown
[Ji and Ye. "An Accelerated Gradient Method for Trace Norm Minimization." International Conference on Machine Learning, 2009.](https://mlanthology.org/icml/2009/ji2009icml-accelerated/) doi:10.1145/1553374.1553434BibTeX
@inproceedings{ji2009icml-accelerated,
title = {{An Accelerated Gradient Method for Trace Norm Minimization}},
author = {Ji, Shuiwang and Ye, Jieping},
booktitle = {International Conference on Machine Learning},
year = {2009},
pages = {457-464},
doi = {10.1145/1553374.1553434},
url = {https://mlanthology.org/icml/2009/ji2009icml-accelerated/}
}