Set-Based Training for Neural Network Verification
Abstract
Neural networks are vulnerable to adversarial attacks, i.e., small input perturbations can significantly affect the outputs of a neural network. Therefore, to ensure safety of neural networks in safety-critical environments, the robustness of a neural network must be formally verified against input perturbations, e.g., from noisy sensors. To improve the robustness of neural networks and thus simplify the formal verification, we present a novel set-based training procedure in which we compute the set of possible outputs given the set of possible inputs and compute for the first time a gradient set, i.e., each possible output has a different gradient. Therefore, we can directly reduce the size of the output enclosure by choosing gradients toward its center. Small output enclosures increase the robustness of a neural network and, at the same time, simplify its formal verification. The latter benefit is due to the fact that a larger size of propagated sets increases the conservatism of most verification methods. Our extensive evaluation demonstrates that set-based training produces robust neural networks with competitive performance, which can be verified using fast (polynomial-time) verification algorithms due to the reduced output set.
Cite
Text
Koller et al. "Set-Based Training for Neural Network Verification." Transactions on Machine Learning Research, 2025.Markdown
[Koller et al. "Set-Based Training for Neural Network Verification." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/koller2025tmlr-setbased/)BibTeX
@article{koller2025tmlr-setbased,
title = {{Set-Based Training for Neural Network Verification}},
author = {Koller, Lukas and Ladner, Tobias and Althoff, Matthias},
journal = {Transactions on Machine Learning Research},
year = {2025},
url = {https://mlanthology.org/tmlr/2025/koller2025tmlr-setbased/}
}