Do Transformers Need Deep Long-Range Memory

07/07/2020
by   Jack W Rae, et al.
0

Deep attention models have advanced the modelling of sequential data across many domains. For language modelling in particular, the Transformer-XL – a Transformer augmented with a long-range memory of past activations – has been shown to be state-of-the-art across a variety of well-studied benchmarks. The Transformer-XL incorporates a long-range memory at every layer of the network, which renders its state to be thousands of times larger than RNN predecessors. However it is unclear whether this is necessary. We perform a set of interventions to show that comparable performance can be obtained with 6X fewer long range memories and better performance can be obtained by limiting the range of attention in lower layers of the network.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/13/2019

Compressive Transformers for Long-Range Sequence Modelling

We present the Compressive Transformer, an attentive sequence model whic...
research
02/18/2023

Neural Attention Memory

We propose a novel perspective of the attention mechanism by reinventing...
research
03/10/2023

Mode-locking Theory for Long-Range Interaction in Artificial Neural Networks

Visual long-range interaction refers to modeling dependencies between di...
research
10/14/2021

Causal Transformers Perform Below Chance on Recursive Nested Constructions, Unlike Humans

Recursive processing is considered a hallmark of human linguistic abilit...
research
03/17/2023

HDformer: A Higher Dimensional Transformer for Diabetes Detection Utilizing Long Range Vascular Signals

Diabetes mellitus is a worldwide concern, and early detection can help t...
research
05/24/2023

Focus Your Attention (with Adaptive IIR Filters)

We present a new layer in which dynamic (i.e.,input-dependent) Infinite ...
research
08/05/2021

FMMformer: Efficient and Flexible Transformer via Decomposed Near-field and Far-field Attention

We propose FMMformers, a class of efficient and flexible transformers in...

Please sign up or login with your details

Forgot password? Click here to reset