Memory in humans and deep language models: Linking hypotheses for model augmentation

10/04/2022
by   Omri Raccah, et al.
0

The computational complexity of the self-attention mechanism in Transformer models significantly limits their ability to generalize over long temporal durations. Memory-augmentation, or the explicit storing of past information in external memory for subsequent predictions, has become a constructive avenue for mitigating this limitation. We argue that memory-augmented Transformers can benefit substantially from considering insights from the memory literature in humans. We detail an approach to integrating evidence from the human memory system through the specification of cross-domain linking hypotheses. We then provide an empirical demonstration to evaluate the use of surprisal as a linking hypothesis, and further identify the limitations of this approach to inform future research.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/23/2022

FastRPB: a Scalable Relative Positional Encoding for Long Sequence Tasks

Transformers achieve remarkable performance in various domains, includin...
research
02/04/2022

Temporal Attention for Language Models

Pretrained language models based on the transformer architecture have sh...
research
06/02/2023

The Information Pathways Hypothesis: Transformers are Dynamic Self-Ensembles

Transformers use the dense self-attention mechanism which gives a lot of...
research
01/23/2023

AttMEMO : Accelerating Transformers with Memoization on Big Memory Systems

Transformer models gain popularity because of their superior inference a...
research
05/22/2023

Adaptive Chameleon or Stubborn Sloth: Unraveling the Behavior of Large Language Models in Knowledge Clashes

By providing external information to large language models (LLMs), tool ...
research
04/22/2023

Read-once machines and the thermodynamic complexity of Maxwell's demons

The thermodynamical costs imposed by computational resource limitations ...

Please sign up or login with your details

Forgot password? Click here to reset