Effective Distributed Learning with Random Features: Improved Bounds and Algorithms

Abstract

In this paper, we study the statistical properties of distributed kernel ridge regression together with random features (DKRR-RF), and obtain optimal generalization bounds under the basic setting, which can substantially relax the restriction on the number of local machines in the existing state-of-art bounds. Specifically, we first show that the simple combination of divide-and-conquer technique and random features can achieve the same statistical accuracy as the exact KRR in expectation requiring only $\mathcal{O}(|\mathcal{D}|)$ memory and $\mathcal{O}(|\mathcal{D}|^{1.5})$ time. Then, beyond the generalization bounds in expectation that demonstrate the average information for multiple trails, we derive generalization bounds in probability to capture the learning performance for a single trail. Finally, we propose an effective communication strategy to further improve the performance of DKRR-RF, and validate the theoretical bounds via numerical experiments.

Cite

Text

Liu et al. "Effective Distributed Learning with Random Features: Improved Bounds and Algorithms." International Conference on Learning Representations, 2021.

Markdown

[Liu et al. "Effective Distributed Learning with Random Features: Improved Bounds and Algorithms." International Conference on Learning Representations, 2021.](https://mlanthology.org/iclr/2021/liu2021iclr-effective/)

BibTeX

@inproceedings{liu2021iclr-effective,
  title     = {{Effective Distributed Learning with Random Features: Improved Bounds and Algorithms}},
  author    = {Liu, Yong and Liu, Jiankun and Wang, Shuqiang},
  booktitle = {International Conference on Learning Representations},
  year      = {2021},
  url       = {https://mlanthology.org/iclr/2021/liu2021iclr-effective/}
}