Provably Near-Optimal Federated Ensemble Distillation with Negligible Overhead

Abstract

Federated ensemble distillation addresses client heterogeneity by generating pseudo-labels for an unlabeled server dataset based on client predictions and training the server model using the pseudo-labeled dataset. The unlabeled server dataset can either be pre-existing or generated through a data-free approach. The effectiveness of this approach critically depends on the method of assigning weights to client predictions when creating pseudo-labels, especially in highly heterogeneous settings. Inspired by theoretical results from GANs, we propose a provably near-optimal weighting method that leverages client discriminators trained with a server-distributed generator and local datasets. Our experiments on various image classification tasks demonstrate that the proposed method significantly outperforms baselines. Furthermore, we show that the additional communication cost, client-side privacy leakage, and client-side computational overhead introduced by our method are negligible, both in scenarios with and without a pre-existing server dataset.

Cite

Text

Jang et al. "Provably Near-Optimal Federated Ensemble Distillation with Negligible Overhead." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Jang et al. "Provably Near-Optimal Federated Ensemble Distillation with Negligible Overhead." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/jang2025icml-provably/)

BibTeX

@inproceedings{jang2025icml-provably,
  title     = {{Provably Near-Optimal Federated Ensemble Distillation with Negligible Overhead}},
  author    = {Jang, Won-Jun and Park, Hyeon-Seo and Lee, Si-Hyeon},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {26896-26924},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/jang2025icml-provably/}
}