Relational Attention: Generalizing Transformers for Graph-Structured Tasks

10/11/2022
by   Cameron Diao, et al.
0

Transformers flexibly operate over sets of real-valued vectors representing task-specific entities and their attributes, where each vector might encode one word-piece token and its position in a sequence, or some piece of information that carries no position at all. But as set processors, transformers are at a disadvantage in reasoning over more general graph-structured data where nodes represent entities and edges represent relations between entities. To address this shortcoming, we generalize transformer attention to consider and update edge vectors in each transformer layer. We evaluate this relational transformer on a diverse array of graph-structured tasks, including the large and challenging CLRS Algorithmic Reasoning Benchmark. There, it dramatically outperforms state-of-the-art graph neural networks expressly designed to reason over graph-structured data. Our analysis demonstrates that these gains are attributable to relational attention's inherent ability to leverage the greater expressivity of graphs over sets.

READ FULL TEXT

page 1

page 18

research
12/01/2021

Systematic Generalization with Edge Transformers

Recent research suggests that systematic generalization in natural langu...
research
12/13/2022

Bridging Graph Position Encodings for Transformers with Weighted Graph-Walking Automata

A current goal in the graph neural network literature is to enable trans...
research
06/10/2022

NAGphormer: Neighborhood Aggregation Graph Transformer for Node Classification in Large Graphs

Graph Transformers have demonstrated superiority on various graph learni...
research
06/23/2022

Equiformer: Equivariant Graph Attention Transformer for 3D Atomistic Graphs

3D-related inductive biases like translational invariance and rotational...
research
05/10/2021

R2D2: Relational Text Decoding with Transformers

We propose a novel framework for modeling the interaction between graphi...
research
01/17/2020

Siamese Graph Neural Networks for Data Integration

Data integration has been studied extensively for decades and approached...
research
09/14/2022

Graph Perceiver IO: A General Architecture for Graph Structured Data

Multimodal machine learning has been widely studied for the development ...

Please sign up or login with your details

Forgot password? Click here to reset