Infusing Lattice Symmetry Priors in Attention Mechanisms for Sample-Efficient Abstract Geometric Reasoning

06/05/2023
by   Mattia Atzeni, et al.
0

The Abstraction and Reasoning Corpus (ARC) (Chollet, 2019) and its most recent language-complete instantiation (LARC) has been postulated as an important step towards general AI. Yet, even state-of-the-art machine learning models struggle to achieve meaningful performance on these problems, falling behind non-learning based approaches. We argue that solving these tasks requires extreme generalization that can only be achieved by proper accounting for core knowledge priors. As a step towards this goal, we focus on geometry priors and introduce LatFormer, a model that incorporates lattice symmetry priors in attention masks. We show that, for any transformation of the hypercubic lattice, there exists a binary attention mask that implements that group action. Hence, our study motivates a modification to the standard attention mechanism, where attention weights are scaled using soft masks generated by a convolutional network. Experiments on synthetic geometric reasoning show that LatFormer requires 2 orders of magnitude fewer data than standard attention and transformers. Moreover, our results on ARC and LARC tasks that incorporate geometric priors provide preliminary evidence that these complex datasets do not lie out of the reach of deep learning models.

READ FULL TEXT

page 2

page 4

page 6

page 12

page 13

research
11/08/2021

Lattice gauge symmetry in neural networks

We review a novel neural network architecture called lattice gauge equiv...
research
05/28/2019

Learning Dynamics of Attention: Human Prior for Interpretable Machine Reasoning

Without relevant human priors, neural networks may learn uninterpretable...
research
03/27/2022

A General Survey on Attention Mechanisms in Deep Learning

Attention is an important mechanism that can be employed for a variety o...
research
11/14/2019

Attention on Abstract Visual Reasoning

Attention mechanisms have been boosting the performance of deep learning...
research
05/25/2022

Guiding Visual Question Answering with Attention Priors

The current success of modern visual reasoning systems is arguably attri...
research
02/28/2019

Link Prediction with Mutual Attention for Text-Attributed Networks

In this extended abstract, we present an algorithm that learns a similar...

Please sign up or login with your details

Forgot password? Click here to reset