Distributed Conformal Prediction via Message Passing
Abstract
Post-hoc calibration of pre-trained models is critical for ensuring reliable inference, especially in safety-critical domains such as healthcare. Conformal Prediction (CP) offers a robust post-hoc calibration framework, providing distribution-free statistical coverage guarantees for prediction sets by leveraging held-out datasets. In this work, we address a decentralized setting where each device has limited calibration data and can communicate only with its neighbors over an arbitrary graph topology. We propose two message-passing-based approaches for achieving reliable inference via CP: quantile-based distributed conformal prediction (Q-DCP) and histogram-based distributed conformal prediction (H-DCP). Q-DCP employs distributed quantile regression enhanced with tailored smoothing and regularization terms to accelerate convergence, while H-DCP uses a consensus-based histogram estimation approach. Through extensive experiments, we investigate the trade-offs between hyperparameter tuning requirements, communication overhead, coverage guarantees, and prediction set sizes across different network topologies. The code of our work is released on: https://github.com/HaifengWen/Distributed-Conformal-Prediction.
Cite
Text
Wen et al. "Distributed Conformal Prediction via Message Passing." Proceedings of the 42nd International Conference on Machine Learning, 2025.Markdown
[Wen et al. "Distributed Conformal Prediction via Message Passing." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/wen2025icml-distributed/)BibTeX
@inproceedings{wen2025icml-distributed,
title = {{Distributed Conformal Prediction via Message Passing}},
author = {Wen, Haifeng and Xing, Hong and Simeone, Osvaldo},
booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
year = {2025},
pages = {66539-66557},
volume = {267},
url = {https://mlanthology.org/icml/2025/wen2025icml-distributed/}
}