Certified Adversarial Robustness Within Multiple Perturbation Bounds
Abstract
Randomized smoothing (RS) is a well known certified defense against adversarial attacks, which creates a smoothed classifier by predicting the most likely class under random noise perturbations of inputs during inference. While initial work focused on robustness to ℓ2 norm perturbations using noise sampled from a Gaussian distribution, subsequent works have shown that different noise distributions can result in robustness to other ℓp norm bounds as well. In general, a specific noise distribution is optimal for defending against a given ℓp norm based attack. In this work, we aim to improve the certified adversarial robustness against multiple perturbation bounds simultaneously. Towards this, we firstly present a novel certification scheme, that effectively combines the certificates obtained using different noise distributions to obtain optimal results against multiple perturbation bounds. We further propose a novel training noise distribution along with a regularized training scheme to improve the certification within both ℓ1 and ℓ2 perturbation norms simultaneously. Contrary to prior works, we compare the certified robustness of different training algorithms across the same natural (clean) accuracy, rather than across fixed noise levels used for training and certification. We also empirically invalidate the argument that training and certifying the classifier with the same amount of noise gives the best results. The proposed approach achieves improvements on the ACR (Average Certified Radius) metric across both ℓ1 and ℓ2 perturbation bounds. Code available at https://github.com/valiisc/NU-Certified-Robustness
Cite
Text
Nandi et al. "Certified Adversarial Robustness Within Multiple Perturbation Bounds." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2023. doi:10.1109/CVPRW59228.2023.00225Markdown
[Nandi et al. "Certified Adversarial Robustness Within Multiple Perturbation Bounds." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2023.](https://mlanthology.org/cvprw/2023/nandi2023cvprw-certified/) doi:10.1109/CVPRW59228.2023.00225BibTeX
@inproceedings{nandi2023cvprw-certified,
title = {{Certified Adversarial Robustness Within Multiple Perturbation Bounds}},
author = {Nandi, Soumalya and Addepalli, Sravanti and Rangwani, Harsh and Babu, R. Venkatesh},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2023},
pages = {2298-2305},
doi = {10.1109/CVPRW59228.2023.00225},
url = {https://mlanthology.org/cvprw/2023/nandi2023cvprw-certified/}
}