Neural Attention Memory

02/18/2023
by   Hyoungwook Nam, et al.
0

We propose a novel perspective of the attention mechanism by reinventing it as a memory architecture for neural networks, namely Neural Attention Memory (NAM). NAM is a memory structure that is both readable and writable via differentiable linear algebra operations. We explore three use cases of NAM: memory-augmented neural network (MANN), few-shot learning, and efficient long-range attention. First, we design two NAM-based MANNs of Long Short-term Memory (LSAM) and NAM Turing Machine (NAM-TM) that show better computational powers in algorithmic zero-shot generalization tasks compared to other baselines such as differentiable neural computer (DNC). Next, we apply NAM to the N-way K-shot learning task and show that it is more effective at reducing false positives compared to the baseline cosine classifier. Finally, we implement an efficient Transformer with NAM and evaluate it with long-range arena tasks to show that NAM can be an efficient and effective alternative for scaled dot-product attention.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/07/2020

Do Transformers Need Deep Long-Range Memory

Deep attention models have advanced the modelling of sequential data acr...
research
11/13/2022

Enhancing Few-shot Image Classification with Cosine Transformer

This paper addresses the few-shot image classification problem. One nota...
research
03/09/2015

Compositional Distributional Semantics with Long Short Term Memory

We are proposing an extension of the recursive neural network that makes...
research
05/25/2019

Neural Stored-program Memory

Neural networks powered with external memory simulate computer behaviors...
research
11/05/2019

Memory Augmented Recursive Neural Networks

Recursive neural networks have shown an impressive performance for model...
research
12/19/2016

Few-Shot Object Recognition from Machine-Labeled Web Images

With the tremendous advances of Convolutional Neural Networks (ConvNets)...
research
06/13/2019

Multigrid Neural Memory

We introduce a novel architecture that integrates a large addressable me...

Please sign up or login with your details

Forgot password? Click here to reset