Self-Attention in Colors: Another Take on Encoding Graph Structure in Transformers

Abstract

We introduce a novel self-attention mechanism, which we call CSA (Chromatic Self-Attention), which extends the notion of attention scores to attention _filters_, independently modulating the feature channels. We showcase CSA in a fully-attentional graph Transformer CGT (Chromatic Graph Transformer) which integrates both graph structural information and edge features, completely bypassing the need for local message-passing components. Our method flexibly encodes graph structure through node-node interactions, by enriching the original edge features with a relative positional encoding scheme. We propose a new scheme based on random walks that encodes both structural and positional information, and show how to incorporate higher-order topological information, such as rings in molecular graphs. Our approach achieves state-of-the-art results on the ZINC benchmark dataset, while providing a flexible framework for encoding graph structure and incorporating higher-order topology.

Cite

Text

Menegaux et al. "Self-Attention in Colors: Another Take on Encoding Graph Structure in Transformers." Transactions on Machine Learning Research, 2023.

Markdown

[Menegaux et al. "Self-Attention in Colors: Another Take on Encoding Graph Structure in Transformers." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/menegaux2023tmlr-selfattention/)

BibTeX

@article{menegaux2023tmlr-selfattention,
  title     = {{Self-Attention in Colors: Another Take on Encoding Graph Structure in Transformers}},
  author    = {Menegaux, Romain and Jehanno, Emmanuel and Selosse, Margot and Mairal, Julien},
  journal   = {Transactions on Machine Learning Research},
  year      = {2023},
  url       = {https://mlanthology.org/tmlr/2023/menegaux2023tmlr-selfattention/}
}