Deep Errors-in-Variables Using a Diffusion Model
Abstract
Errors-in-Variables is the statistical concept used to explicitly model input variable errors caused, for example, by noise. While it has long been known in statistics that not accounting for such errors can produce a substantial bias, the vast majority of deep learning models have thus far neglected Errors-in-Variables approaches. Reasons for this include a significant increase of the numerical burden and the challenge in assigning an appropriate prior in a Bayesian treatment. To date, the attempts made to use Errors-in-Variables for neural networks do not scale to deep networks or are too simplistic to enhance the prediction performance. This work shows for the first time how Bayesian deep Errors-in-Variables models can increase the prediction performance. We present a scalable variational inference scheme for Bayesian Errors-in-Variables and demonstrate a significant increase in prediction performance for the case of image classification. Concretely, we use a diffusion model as input posterior to obtain a distribution over the denoised image data. We also observe that training the diffusion model on an unnoisy surrogate dataset can suffice to achieve an improved prediction performance on noisy data.
Cite
Text
Faller et al. "Deep Errors-in-Variables Using a Diffusion Model." Machine Learning, 2025. doi:10.1007/S10994-025-06744-XMarkdown
[Faller et al. "Deep Errors-in-Variables Using a Diffusion Model." Machine Learning, 2025.](https://mlanthology.org/mlj/2025/faller2025mlj-deep/) doi:10.1007/S10994-025-06744-XBibTeX
@article{faller2025mlj-deep,
title = {{Deep Errors-in-Variables Using a Diffusion Model}},
author = {Faller, Josua and Martin, Jörg and Elster, Clemens},
journal = {Machine Learning},
year = {2025},
pages = {107},
doi = {10.1007/S10994-025-06744-X},
volume = {114},
url = {https://mlanthology.org/mlj/2025/faller2025mlj-deep/}
}