The Inhibitor: ReLU and Addition-Based Attention for Efficient Transformers (Student Abstract)
Abstract
To enhance the computational efficiency of quantized Transformers, we replace the dot-product and Softmax-based attention with an alternative mechanism involving addition and ReLU activation only. This side-steps the expansion to double precision often required by matrix multiplication and avoids costly Softmax evaluations but maintains much of the core functionality of conventional dot-product attention. It can enable more efficient execution and support larger quantized Transformer models on resource-constrained hardware or alternative arithmetic systems like homomorphic encryption. Training experiments on four common benchmark tasks show test set prediction scores comparable to those of conventional Transformers with dot-product attention. Our scaling experiments also suggest significant computational savings, both in plaintext and under encryption. In particular, we believe that the ReLU and addition-based attention mechanism introduced in this paper may enable privacy-preserving AI applications operating under homomorphic encryption by avoiding the costly multiplication of encrypted variables.
Cite
Text
Brännvall. "The Inhibitor: ReLU and Addition-Based Attention for Efficient Transformers (Student Abstract)." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I21.30422Markdown
[Brännvall. "The Inhibitor: ReLU and Addition-Based Attention for Efficient Transformers (Student Abstract)." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/brannvall2024aaai-inhibitor/) doi:10.1609/AAAI.V38I21.30422BibTeX
@inproceedings{brannvall2024aaai-inhibitor,
title = {{The Inhibitor: ReLU and Addition-Based Attention for Efficient Transformers (Student Abstract)}},
author = {Brännvall, Rickard},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2024},
pages = {23445-23446},
doi = {10.1609/AAAI.V38I21.30422},
url = {https://mlanthology.org/aaai/2024/brannvall2024aaai-inhibitor/}
}