Distributed Personalized Empirical Risk Minimization

Abstract

This paper advocates a new paradigm Personalized Empirical Risk Minimization (PERM) to facilitate learning from heterogeneous data sources without imposing stringent constraints on computational resources shared by participating devices. In PERM, we aim at learning a distinct model for each client by personalizing the aggregation of local empirical losses by effectively estimating the statistical discrepancy among data distributions, which entails optimal statistical accuracy for all local distributions and overcomes the data heterogeneity issue. To learn personalized models at scale, we propose a distributed algorithm that replaces the standard model averaging with model shuffling to simultaneously optimize PERM objectives for all devices. This also allows to learn distinct model architectures (e.g., neural networks with different number of parameters) for different clients, thus confining to underlying memory and compute resources of individual clients. We rigorously analyze the convergence of proposed algorithm and conduct experiments that corroborates the effectiveness of proposed paradigm.

Cite

Text

Deng et al. "Distributed Personalized Empirical Risk Minimization." Neural Information Processing Systems, 2023.

Markdown

[Deng et al. "Distributed Personalized Empirical Risk Minimization." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/deng2023neurips-distributed/)

BibTeX

@inproceedings{deng2023neurips-distributed,
  title     = {{Distributed Personalized Empirical Risk Minimization}},
  author    = {Deng, Yuyang and Kamani, Mohammad Mahdi and Mahdavinia, Pouria and Mahdavi, Mehrdad},
  booktitle = {Neural Information Processing Systems},
  year      = {2023},
  url       = {https://mlanthology.org/neurips/2023/deng2023neurips-distributed/}
}