Deformable Graph Transformer

06/29/2022
by   Jinyoung Park, et al.
6

Transformer-based models have been widely used and achieved state-of-the-art performance in various domains such as natural language processing and computer vision. Recent works show that Transformers can also be generalized to graph-structured data. However, the success is limited to small-scale graphs due to technical challenges such as the quadratic complexity in regards to the number of nodes and non-local aggregation that often leads to inferior generalization performance to conventional graph neural networks. In this paper, to address these issues, we propose Deformable Graph Transformer (DGT) that performs sparse attention with dynamically sampled key and value pairs. Specifically, our framework first constructs multiple node sequences with various criteria to consider both structural and semantic proximity. Then, the sparse attention is applied to the node sequences for learning node representations with a reduced computational cost. We also design simple and effective positional encodings to capture structural similarity and distance between nodes. Experiments demonstrate that our novel graph Transformer consistently outperforms existing Transformer-based models and shows competitive performance compared to state-of-the-art models on 8 graph benchmark datasets including large-scale graphs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/10/2022

NAGphormer: Neighborhood Aggregation Graph Transformer for Node Classification in Large Graphs

Graph Transformers have demonstrated superiority on various graph learni...
research
10/25/2021

Gophormer: Ego-Graph Transformer for Node Classification

Transformers have achieved remarkable performance in a myriad of fields ...
research
10/24/2022

DAGformer: Directed Acyclic Graph Transformer

In many fields, such as natural language processing and computer vision,...
research
02/18/2022

Unleashing the Power of Transformer for Graphs

Despite recent successes in natural language processing and computer vis...
research
03/25/2022

Gransformer: Transformer-based Graph Generation

Transformers have become widely used in modern models for various tasks ...
research
07/05/2019

Graph Representation Learning via Hard and Channel-Wise Attention Networks

Attention operators have been widely applied in various fields, includin...
research
10/08/2022

Hierarchical Graph Transformer with Adaptive Node Sampling

The Transformer architecture has achieved remarkable success in a number...

Please sign up or login with your details

Forgot password? Click here to reset