Adversarial Robustness in Parameter-Space Classifiers

Abstract

Implicit Neural Representations (INRs) have been recently garnering increasing interest in various research fields, mainly due to their ability to represent large, complex data in a compact and continuous manner. Past work further showed that numerous popular downstream tasks can be performed directly in the INR parameter-space. Doing so can substantially reduce the computational resources required to process the represented data in their native domain. A major difficulty in using modern machine-learning approaches, is their high susceptibility to adversarial attacks, which have been shown to greatly limit the reliability and applicability of such methods in a wide range of settings. In this work, we show that parameter-space models trained for classification are inherently robust to adversarial attacks -- without the need of any robust training. To support our claims, we develop a novel suite of adversarial attacks targeting parameter-space classifiers, and furthermore analyze practical considerations of attacking parameter-space classifiers.

Cite

Text

Shor et al. "Adversarial Robustness in Parameter-Space Classifiers." ICLR 2025 Workshops: WSL, 2025. doi:10.48550/arxiv.2502.20314

Markdown

[Shor et al. "Adversarial Robustness in Parameter-Space Classifiers." ICLR 2025 Workshops: WSL, 2025.](https://mlanthology.org/iclrw/2025/shor2025iclrw-adversarial/) doi:10.48550/arxiv.2502.20314

BibTeX

@inproceedings{shor2025iclrw-adversarial,
  title     = {{Adversarial Robustness in Parameter-Space Classifiers}},
  author    = {Shor, Tamir and Fetaya, Ethan and Baskin, Chaim and Bronstein, Alex M.},
  booktitle = {ICLR 2025 Workshops: WSL},
  year      = {2025},
  doi       = {10.48550/arxiv.2502.20314},
  url       = {https://mlanthology.org/iclrw/2025/shor2025iclrw-adversarial/}
}