Representing Long-Range Context for Graph Neural Networks with Global Attention

01/21/2022
by   Zhanghao Wu, et al.
2

Graph neural networks are powerful architectures for structured datasets. However, current methods struggle to represent long-range dependencies. Scaling the depth or width of GNNs is insufficient to broaden receptive fields as larger GNNs encounter optimization instabilities such as vanishing gradients and representation oversmoothing, while pooling-based approaches have yet to become as universally useful as in computer vision. In this work, we propose the use of Transformer-based self-attention to learn long-range pairwise relationships, with a novel "readout" mechanism to obtain a global graph embedding. Inspired by recent computer vision results that find position-invariant attention performant in learning long-range relationships, our method, which we call GraphTrans, applies a permutation-invariant Transformer module after a standard GNN module. This simple architecture leads to state-of-the-art results on several graph classification tasks, outperforming methods that explicitly encode graph structure. Our results suggest that purely-learning-based approaches without graph structure may be suitable for learning high-level, long-range relationships on graphs. Code for GraphTrans is available at https://github.com/ucbrise/graphtrans.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/19/2020

Improving the Long-Range Performance of Gated Graph Neural Networks

Many popular variants of graph neural networks (GNNs) that are capable o...
research
12/27/2022

A Generalization of ViT/MLP-Mixer to Graphs

Graph Neural Networks (GNNs) have shown great potential in the field of ...
research
02/17/2023

Search to Capture Long-range Dependency with Stacking GNNs for Graph Classification

In recent years, Graph Neural Networks (GNNs) have been popular in the g...
research
03/10/2021

Graph Neural Networks Inspired by Classical Iterative Algorithms

Despite the recent success of graph neural networks (GNN), common archit...
research
11/26/2022

PatchGT: Transformer over Non-trainable Clusters for Learning Graph Representations

Recently the Transformer structure has shown good performances in graph ...
research
07/23/2019

Compact Global Descriptor for Neural Networks

Long-range dependencies modeling, widely used in capturing spatiotempora...
research
07/13/2022

Eliminating Gradient Conflict in Reference-based Line-art Colorization

Reference-based line-art colorization is a challenging task in computer ...

Please sign up or login with your details

Forgot password? Click here to reset