Robust-by-Design Classification via Unitary-Gradient Neural Networks
Abstract
The use of neural networks in safety-critical systems requires safe and robust models, due to the existence of adversarial attacks. Knowing the minimal adversarial perturbation of any input x, or, equivalently, knowing the distance of x from the classification boundary, allows evaluating the classification robustness, providing certifiable predictions. Unfortunately, state-of-the-art techniques for computing such a distance are computationally expensive and hence not suited for online applications. This work proposes a novel family of classifiers, namely Signed Distance Classifiers (SDCs), that, from a theoretical perspective, directly output the exact distance of x from the classification boundary, rather than a probability score (e.g., SoftMax). SDCs represent a family of robust-by-design classifiers. To practically address the theoretical requirements of an SDC, a novel network architecture named Unitary-Gradient Neural Network is presented. Experimental results show that the proposed architecture approximates a signed distance classifier, hence allowing an online certifiable classification of x at the cost of a single inference.
Cite
Text
Brau et al. "Robust-by-Design Classification via Unitary-Gradient Neural Networks." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I12.26721Markdown
[Brau et al. "Robust-by-Design Classification via Unitary-Gradient Neural Networks." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/brau2023aaai-robust/) doi:10.1609/AAAI.V37I12.26721BibTeX
@inproceedings{brau2023aaai-robust,
title = {{Robust-by-Design Classification via Unitary-Gradient Neural Networks}},
author = {Brau, Fabio and Rossolini, Giulio and Biondi, Alessandro and Buttazzo, Giorgio C.},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {14729-14737},
doi = {10.1609/AAAI.V37I12.26721},
url = {https://mlanthology.org/aaai/2023/brau2023aaai-robust/}
}