Foundation Model-Oriented Robustness: Robust Image Model Evaluation with Pretrained Models

Abstract

Machine learning has demonstrated remarkable performance over finite datasets, yet whether the scores over the fixed benchmarks can sufficiently indicate the model’s performance in the real world is still in discussion. In reality, an ideal robust model will probably behave similarly to the oracle (e.g., the human users), thus a good evaluation protocol is probably to evaluate the models’ behaviors in comparison to the oracle. In this paper, we introduce a new robustness measurement that directly measures the image classification model’s performance compared with a surrogate oracle (i.e., a zoo of foundation models). Besides, we design a simple method that can accomplish the evaluation beyond the scope of the benchmarks. Our method extends the image datasets with new samples that are sufficiently perturbed to be distinct from the ones in the original sets, but are still bounded within the same image-label structure the original test image represents, constrained by a zoo of foundation models pretrained with a large amount of samples. As a result, our new method will offer us a new way to evaluate the models’ robustness performance, free of limitations of fixed benchmarks or constrained perturbations, although scoped by the power of the oracle. In addition to the evaluation results, we also leverage our generated data to understand the behaviors of the model and our new evaluation strategies.

Cite

Text

Zhang et al. "Foundation Model-Oriented Robustness: Robust Image Model Evaluation with Pretrained Models." International Conference on Learning Representations, 2024.

Markdown

[Zhang et al. "Foundation Model-Oriented Robustness: Robust Image Model Evaluation with Pretrained Models." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/zhang2024iclr-foundation/)

BibTeX

@inproceedings{zhang2024iclr-foundation,
  title     = {{Foundation Model-Oriented Robustness: Robust Image Model Evaluation with Pretrained Models}},
  author    = {Zhang, Peiyan and Liu, Haoyang and Li, Chaozhuo and Xie, Xing and Kim, Sunghun and Wang, Haohan},
  booktitle = {International Conference on Learning Representations},
  year      = {2024},
  url       = {https://mlanthology.org/iclr/2024/zhang2024iclr-foundation/}
}