Communication-Efficient Distributed Multi-Task Learning with Matrix Sparsity Regularization
Abstract
This work focuses on distributed optimization for multi-task learning with matrix sparsity regularization. We propose a fast communication-efficient distributed optimization method for solving the problem. With the proposed method, training data of different tasks can be geo-distributed over different local machines, and the tasks can be learned jointly through the matrix sparsity regularization without a need to centralize the data. We theoretically prove that our proposed method enjoys a fast convergence rate for different types of loss functions in the distributed environment. To further reduce the communication cost during the distributed optimization procedure, we propose a data screening approach to safely filter inactive features or variables. Finally, we conduct extensive experiments on both synthetic and real-world datasets to demonstrate the effectiveness of our proposed method.
Cite
Text
Zhou et al. "Communication-Efficient Distributed Multi-Task Learning with Matrix Sparsity Regularization." Machine Learning, 2020. doi:10.1007/S10994-019-05847-6Markdown
[Zhou et al. "Communication-Efficient Distributed Multi-Task Learning with Matrix Sparsity Regularization." Machine Learning, 2020.](https://mlanthology.org/mlj/2020/zhou2020mlj-communicationefficient/) doi:10.1007/S10994-019-05847-6BibTeX
@article{zhou2020mlj-communicationefficient,
title = {{Communication-Efficient Distributed Multi-Task Learning with Matrix Sparsity Regularization}},
author = {Zhou, Qiang and Chen, Yu and Pan, Sinno Jialin},
journal = {Machine Learning},
year = {2020},
pages = {569-601},
doi = {10.1007/S10994-019-05847-6},
volume = {109},
url = {https://mlanthology.org/mlj/2020/zhou2020mlj-communicationefficient/}
}