Input Uncertainty Propagation Through Trained Neural Networks
Abstract
When physical sensors are involved, such as image sensors, the uncertainty over the input data is often a major component of the output uncertainty of machine learning models. In this work, we address the problem of input uncertainty propagation through trained neural networks. We do not rely on a Gaussian distribution assumption of the output or of any intermediate layer. We propagate instead a Gaussian Mixture Model (GMM) that offers much more flexibility, using the Split&Merge algorithm. This paper’s main contribution is the computation of a Wasserstein criterion to control the Gaussian splitting procedure for which theoretical guarantees of convergence on the output distribution estimates are derived. The methodology is tested against a wide range of datasets and networks. It shows robustness, and genericity and offers highly accurate output probability density function estimation while maintaining a reasonable computational cost compared with the standard Monte Carlo (MC) approach.
Cite
Text
Monchot et al. "Input Uncertainty Propagation Through Trained Neural Networks." International Conference on Machine Learning, 2023.Markdown
[Monchot et al. "Input Uncertainty Propagation Through Trained Neural Networks." International Conference on Machine Learning, 2023.](https://mlanthology.org/icml/2023/monchot2023icml-input/)BibTeX
@inproceedings{monchot2023icml-input,
title = {{Input Uncertainty Propagation Through Trained Neural Networks}},
author = {Monchot, Paul and Coquelin, Loic and Petit, Sébastien Julien and Marmin, Sébastien and Le Pennec, Erwan and Fischer, Nicolas},
booktitle = {International Conference on Machine Learning},
year = {2023},
pages = {25140-25173},
volume = {202},
url = {https://mlanthology.org/icml/2023/monchot2023icml-input/}
}