Internet publication
Graph inductive biases in transformers without message passing
- Abstract:
-
Transformers for graph data are increasingly widely studied and successful in numerous learning tasks. Graph inductive biases are crucial for Graph Transformers, and previous works incorporate them using message-passing modules and/or positional encodings. However, Graph Transformers that use message-passing inherit known issues of message-passing, and differ significantly from Transformers used in other domains, thus making transfer of research advances more difficult. On the other hand, Graph Transformers without message-passing often perform poorly on smaller datasets, where inductive biases are more crucial. To bridge this gap, we propose the Graph Inductive bias Transformer (GRIT) -- a new Graph Transformer that incorporates graph inductive biases without using message passing. GRIT is based on several architectural changes that are each theoretically and empirically justified, including: learned relative positional encodings initialized with random walk probabilities, a flexible attention mechanism that updates node and node-pair representations, and injection of degree information in each layer. We prove that GRIT is expressive -- it can express shortest path distances and various graph propagation matrices. GRIT achieves state-of-the-art empirical performance across a variety of graph datasets, thus showing the power that Graph Transformers without message-passing can deliver.
- Publication status:
- Published
- Peer review status:
- Not peer reviewed
Actions
Access Document
- Files:
-
-
(Preview, Pre-print, pdf, 852.5KB, Terms of use)
-
- Publisher copy:
- 10.48550/arXiv.2305.17589
Authors
- Host title:
- arXiv
- Publication date:
- 2023-05-27
- DOI:
- Language:
-
English
- Keywords:
- Pubs id:
-
1775416
- Local pid:
-
pubs:1775416
- Deposit date:
-
2024-05-16
Terms of use
- Copyright holder:
- Ma et al.
- Copyright date:
- 2023
- Rights statement:
- © 2023 The Author(s). This article is licensed under an arXiv.org perpetual, non-exclusive license 1.0.
- Notes:
- This is the pre-print version of the article. The final version is available online from PMLR at https://proceedings.mlr.press/v202/ma23c.html
- Licence:
- Other
If you are the owner of this record, you can report an update to it here: Report update to this record