On the Robustness of GUI Grounding Models Against Image Attacks

Abstract

Graphical User Interface (GUI) grounding models are crucial for enabling intelligent agents to understand and interact with complex visual interfaces. However, these models face significant robustness challenges in real-world scenarios due to natural noise and adversarial perturbations, and their robustness remains underexplored. In this study, we systematically evaluate the robustness of state-of-the-art GUI grounding models, such as UGround, under three conditions: natural noise, untargeted adversarial attacks, and targeted adversarial attacks. Our experiments, which were conducted across a wide range of GUI environments, including mobile, desktop, and web interfaces, have clearly demonstrated that GUI grounding models exhibit a high degree of sensitivity to adversarial perturbations and low-resolution conditions. These findings provide valuable insights into the vulnerabilities of GUI grounding models and establish a strong benchmark for future research aimed at enhancing their robustness in practical applications. Our code is available at https://github.com/ZZZhr-1/Robust_GUI_Grounding.

Cite

Text

Zhao et al. "On the Robustness of GUI Grounding Models Against Image Attacks." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2025.

Markdown

[Zhao et al. "On the Robustness of GUI Grounding Models Against Image Attacks." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2025.](https://mlanthology.org/cvprw/2025/zhao2025cvprw-robustness/)

BibTeX

@inproceedings{zhao2025cvprw-robustness,
  title     = {{On the Robustness of GUI Grounding Models Against Image Attacks}},
  author    = {Zhao, Haoren and Chen, Tianyi and Wang, Zhen},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2025},
  pages     = {1618-1623},
  url       = {https://mlanthology.org/cvprw/2025/zhao2025cvprw-robustness/}
}