Reliable Uncertainty Quantification in Machine Learning via Conformal Prediction
Abstract
Deploying machine learning (ML) models in high-stakes domains such as healthcare and autonomous systems requires reliable uncertainty quantification (UQ) to ensure safe and accurate decision-making. Conformal prediction (CP) offers a robust, distribution-agnostic framework for UQ, providing valid prediction sets that guarantee a specified coverage probability. However, existing CP methods are often limited by assumptions that are violated in real-world scenarios, such as non-i.i.d. data, and by a lack of integration with modern machine learning workflows, particularly in large generative models. This research aims to address these limitations by advancing CP techniques to operate effectively in non-i.i.d. settings, improving predictive efficiency without sacrificing theoretical guarantees, and integrating CP directly into model training processes. These developments will enhance the practical applicability of CP for a wide range of ML tasks, enabling more reliable and interpretable models in high-stakes applications.
Cite
Text
Shi. "Reliable Uncertainty Quantification in Machine Learning via Conformal Prediction." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I28.35227Markdown
[Shi. "Reliable Uncertainty Quantification in Machine Learning via Conformal Prediction." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/shi2025aaai-reliable/) doi:10.1609/AAAI.V39I28.35227BibTeX
@inproceedings{shi2025aaai-reliable,
title = {{Reliable Uncertainty Quantification in Machine Learning via Conformal Prediction}},
author = {Shi, Yuanjie},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2025},
pages = {29299-29300},
doi = {10.1609/AAAI.V39I28.35227},
url = {https://mlanthology.org/aaai/2025/shi2025aaai-reliable/}
}