Graph Inductive Biases in Transformers without Message Passing

05/27/2023
by   Liheng Ma, et al.
0

Transformers for graph data are increasingly widely studied and successful in numerous learning tasks. Graph inductive biases are crucial for Graph Transformers, and previous works incorporate them using message-passing modules and/or positional encodings. However, Graph Transformers that use message-passing inherit known issues of message-passing, and differ significantly from Transformers used in other domains, thus making transfer of research advances more difficult. On the other hand, Graph Transformers without message-passing often perform poorly on smaller datasets, where inductive biases are more crucial. To bridge this gap, we propose the Graph Inductive bias Transformer (GRIT) – a new Graph Transformer that incorporates graph inductive biases without using message passing. GRIT is based on several architectural changes that are each theoretically and empirically justified, including: learned relative positional encodings initialized with random walk probabilities, a flexible attention mechanism that updates node and node-pair representations, and injection of degree information in each layer. We prove that GRIT is expressive – it can express shortest path distances and various graph propagation matrices. GRIT achieves state-of-the-art empirical performance across a variety of graph datasets, thus showing the power that Graph Transformers without message-passing can deliver.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/23/2022

Equiformer: Equivariant Graph Attention Transformer for 3D Atomistic Graphs

3D-related inductive biases like translational invariance and rotational...
research
05/12/2023

AGFormer: Efficient Graph Representation with Anchor-Graph Transformer

To alleviate the local receptive issue of GCN, Transformers have been ex...
research
04/01/2023

Abstractors: Transformer Modules for Symbolic Message Passing and Relational Reasoning

A framework is proposed that casts relational learning in terms of trans...
research
06/02/2022

Shortest Path Networks for Graph Property Prediction

Most graph neural network models rely on a particular message passing pa...
research
02/13/2022

Flowformer: Linearizing Transformers with Conservation Flows

Transformers based on the attention mechanism have achieved impressive s...
research
06/07/2023

Enabling tabular deep learning when d ≫ n with an auxiliary knowledge graph

Machine learning models exhibit strong performance on datasets with abun...
research
12/15/2022

Essentials of Parallel Graph Analytics

We identify the graph data structure, frontiers, operators, an iterative...

Please sign up or login with your details

Forgot password? Click here to reset