Differentiable Random Access Memory using Lattices

07/07/2021
by   Adam P. Goucher, et al.
0

We introduce a differentiable random access memory module with O(1) performance regardless of size, scaling to billions of entries. The design stores entries on points of a chosen lattice to calculate nearest neighbours of arbitrary points efficiently by exploiting symmetries. Augmenting a standard neural network architecture with a single memory layer based on this, we can scale the parameter count up to memory limits with negligible computational overhead, giving better accuracy at similar cost. On large language modelling tasks, these enhanced models with larger capacity significantly outperform the unmodified transformer baseline. We found continued scaling with memory size up to the limits tested.

READ FULL TEXT
research
07/10/2019

Large Memory Layers with Product Keys

This paper introduces a structured memory which can be easily integrated...
research
06/13/2019

Multigrid Neural Memory

We introduce a novel architecture that integrates a large addressable me...
research
10/27/2016

Scaling Memory-Augmented Neural Networks with Sparse Reads and Writes

Neural networks augmented with external memory have the ability to learn...
research
11/19/2015

Neural Random-Access Machines

In this paper, we propose and investigate a new neural network architect...
research
12/31/2022

Pseudo-Inverted Bottleneck Convolution for DARTS Search Space

Differentiable Architecture Search (DARTS) has attracted considerable at...
research
01/27/2021

CNN with large memory layers

This work is centred around the recently proposed product key memory str...
research
05/31/2021

Memory-Efficient Differentiable Transformer Architecture Search

Differentiable architecture search (DARTS) is successfully applied in ma...

Please sign up or login with your details

Forgot password? Click here to reset