Provable Robustness Against a Union of L_0 Adversarial Attacks

Abstract

Sparse or L0 adversarial attacks arbitrarily perturb an unknown subset of the features. L0 robustness analysis is particularly well-suited for heterogeneous (tabular) data where features have different types or scales. State-of-the-art L0 certified defenses are based on randomized smoothing and apply to evasion attacks only. This paper proposes feature partition aggregation (FPA) -- a certified defense against the union of L0 evasion, backdoor, and poisoning attacks. FPA generates its stronger robustness guarantees via an ensemble whose submodels are trained on disjoint feature sets. Compared to state-of-the-art L0 defenses, FPA is up to 3,000x faster and provides larger median robustness guarantees (e.g., median certificates of 13 pixels over 10 for CIFAR10, 12 pixels over 10 for MNIST, 4 features over 1 for Weather, and 3 features over 1 for Ames), meaning FPA provides the additional dimensions of robustness essentially for free.

Cite

Text

Hammoudeh and Lowd. "Provable Robustness Against a Union of L_0 Adversarial Attacks." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I19.30106

Markdown

[Hammoudeh and Lowd. "Provable Robustness Against a Union of L_0 Adversarial Attacks." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/hammoudeh2024aaai-provable/) doi:10.1609/AAAI.V38I19.30106

BibTeX

@inproceedings{hammoudeh2024aaai-provable,
  title     = {{Provable Robustness Against a Union of L_0 Adversarial Attacks}},
  author    = {Hammoudeh, Zayd and Lowd, Daniel},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {21134-21142},
  doi       = {10.1609/AAAI.V38I19.30106},
  url       = {https://mlanthology.org/aaai/2024/hammoudeh2024aaai-provable/}
}