LGI-GT: Graph Transformers with Local and Global Operators Interleaving
Abstract
Since Transformers can alleviate some critical and fundamental problems of graph neural networks (GNNs), such as over-smoothing, over-squashing and limited expressiveness, they have been successfully applied to graph representation learning and achieved impressive results. However, although there are many works dedicated to make graph Transformers (GTs) aware of the structure and edge information by specifically tailored attention forms or graph-related positional and structural encodings, few works address the problem of how to construct high-performing GTs with modules of GNNs and Transformers. In this paper, we propose a novel graph Transformer with local and global operators interleaving (LGI-GT), in which we further design a new method propagating embeddings of the [CLS] token for global information representation. Additionally, we propose an effective message passing module called edge enhanced local attention (EELA), which makes LGI-GT a full-attention GT. Extensive experiments demonstrate that LGI-GT performs consistently better than previous state-of-the-art GNNs and GTs, while ablation studies show the effectiveness of the proposed LGI scheme and EELA. The source code of LGI-GT is available at https://github.com/shuoyinn/LGI-GT.
Cite
Text
Yin and Zhong. "LGI-GT: Graph Transformers with Local and Global Operators Interleaving." International Joint Conference on Artificial Intelligence, 2023. doi:10.24963/IJCAI.2023/501Markdown
[Yin and Zhong. "LGI-GT: Graph Transformers with Local and Global Operators Interleaving." International Joint Conference on Artificial Intelligence, 2023.](https://mlanthology.org/ijcai/2023/yin2023ijcai-lgi/) doi:10.24963/IJCAI.2023/501BibTeX
@inproceedings{yin2023ijcai-lgi,
title = {{LGI-GT: Graph Transformers with Local and Global Operators Interleaving}},
author = {Yin, Shuo and Zhong, Guoqiang},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2023},
pages = {4504-4512},
doi = {10.24963/IJCAI.2023/501},
url = {https://mlanthology.org/ijcai/2023/yin2023ijcai-lgi/}
}