Set-Based Training for Neural Network Verification

Abstract

Neural networks are vulnerable to adversarial attacks, i.e., small input perturbations can significantly affect the outputs of a neural network. Therefore, to ensure safety of neural networks in safety-critical environments, the robustness of a neural network must be formally verified against input perturbations, e.g., from noisy sensors. To improve the robustness of neural networks and thus simplify the formal verification, we present a novel set-based training procedure in which we compute the set of possible outputs given the set of possible inputs and compute for the first time a gradient set, i.e., each possible output has a different gradient. Therefore, we can directly reduce the size of the output enclosure by choosing gradients toward its center. Small output enclosures increase the robustness of a neural network and, at the same time, simplify its formal verification. The latter benefit is due to the fact that a larger size of propagated sets increases the conservatism of most verification methods. Our extensive evaluation demonstrates that set-based training produces robust neural networks with competitive performance, which can be verified using fast (polynomial-time) verification algorithms due to the reduced output set.

Cite

Text

Koller et al. "Set-Based Training for Neural Network Verification." Transactions on Machine Learning Research, 2025.

Markdown

[Koller et al. "Set-Based Training for Neural Network Verification." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/koller2025tmlr-setbased/)

BibTeX

@article{koller2025tmlr-setbased,
  title     = {{Set-Based Training for Neural Network Verification}},
  author    = {Koller, Lukas and Ladner, Tobias and Althoff, Matthias},
  journal   = {Transactions on Machine Learning Research},
  year      = {2025},
  url       = {https://mlanthology.org/tmlr/2025/koller2025tmlr-setbased/}
}