Attacking Few-Shot Classifiers with Adversarial Support Poisoning

Abstract

This paper examines the robustness of deployed few-shot meta-learning systems when they are fed an imperceptibly perturbed few-shot dataset, showing that the resulting predictions on test inputs can become worse than chance. This is achieved by developing a novel attack, Adversarial Support Poisoning or ASP, which crafts a poisoned set of examples. When even a small subset of malicious data points is inserted into the support set of a meta-learner, accuracy is significantly reduced. We evaluate the new attack on a variety of few-shot classification algorithms and scenarios, and propose a form of adversarial training that significantly improves robustness against both poisoning and evasion attacks.

Cite

Text

Oldewage et al. "Attacking Few-Shot Classifiers with Adversarial Support Poisoning." ICML 2021 Workshops: AML, 2021.

Markdown

[Oldewage et al. "Attacking Few-Shot Classifiers with Adversarial Support Poisoning." ICML 2021 Workshops: AML, 2021.](https://mlanthology.org/icmlw/2021/oldewage2021icmlw-attacking/)

BibTeX

@inproceedings{oldewage2021icmlw-attacking,
  title     = {{Attacking Few-Shot Classifiers with Adversarial Support Poisoning}},
  author    = {Oldewage, Elre Talea and Bronskill, John F and Turner, Richard E},
  booktitle = {ICML 2021 Workshops: AML},
  year      = {2021},
  url       = {https://mlanthology.org/icmlw/2021/oldewage2021icmlw-attacking/}
}