BVT-IMA: Binary Vision Transformer with Information-Modified Attention

Abstract

As a compression method that can significantly reduce the cost of calculations and memories, model binarization has been extensively studied in convolutional neural networks. However, the recently popular vision transformer models pose new challenges to such a technique, in which the binarized models suffer from serious performance drops. In this paper, an attention shifting is observed in the binary multi-head self-attention module, which can influence the information fusion between tokens and thus hurts the model performance. From the perspective of information theory, we find a correlation between attention scores and the information quantity, further indicating that a reason for such a phenomenon may be the loss of the information quantity induced by constant moduli of binarized tokens. Finally, we reveal the information quantity hidden in the attention maps of binary vision transformers and propose a simple approach to modify the attention values with look-up information tables so that improve the model performance. Extensive experiments on CIFAR-100/TinyImageNet/ImageNet-1k demonstrate the effectiveness of the proposed information-modified attention on binary vision transformers.

Cite

Text

Wang et al. "BVT-IMA: Binary Vision Transformer with Information-Modified Attention." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I14.29505

Markdown

[Wang et al. "BVT-IMA: Binary Vision Transformer with Information-Modified Attention." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/wang2024aaai-bvt/) doi:10.1609/AAAI.V38I14.29505

BibTeX

@inproceedings{wang2024aaai-bvt,
  title     = {{BVT-IMA: Binary Vision Transformer with Information-Modified Attention}},
  author    = {Wang, Zhenyu and Luo, Hao and Xie, Xuemei and Wang, Fan and Shi, Guangming},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {15761-15769},
  doi       = {10.1609/AAAI.V38I14.29505},
  url       = {https://mlanthology.org/aaai/2024/wang2024aaai-bvt/}
}