Adversarial Self-Attention for Language Understanding

Abstract

Deep neural models (e.g. Transformer) naturally learn spurious features, which create a ``shortcut'' between the labels and inputs, thus impairing the generalization and robustness. This paper advances self-attention mechanism to its robust variant for Transformer-based pre-trained language models (e.g. BERT). We propose Adversarial Self-Attention mechanism (ASA), which adversarially biases the attentions to effectively suppress the model reliance on features (e.g. specific keywords) and encourage its exploration of broader semantics. We conduct comprehensive evaluation across a wide range of tasks for both pre-training and fine-tuning stages. For pre-training, ASA unfolds remarkable performance gain compared to naive training for longer steps. For fine-tuning, ASA-empowered models outweigh naive models by a large margin considering both generalization and robustness.

Cite

Text

Wu et al. "Adversarial Self-Attention for Language Understanding." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I11.26608

Markdown

[Wu et al. "Adversarial Self-Attention for Language Understanding." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/wu2023aaai-adversarial-a/) doi:10.1609/AAAI.V37I11.26608

BibTeX

@inproceedings{wu2023aaai-adversarial-a,
  title     = {{Adversarial Self-Attention for Language Understanding}},
  author    = {Wu, Hongqiu and Ding, Ruixue and Zhao, Hai and Xie, Pengjun and Huang, Fei and Zhang, Min},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2023},
  pages     = {13727-13735},
  doi       = {10.1609/AAAI.V37I11.26608},
  url       = {https://mlanthology.org/aaai/2023/wu2023aaai-adversarial-a/}
}