Learning What to Remember: Long-term Episodic Memory Networks for Learning from Streaming Data

12/11/2018
by   Hyunwoo Jung, et al.
0

Current generation of memory-augmented neural networks has limited scalability as they cannot efficiently process data that are too large to fit in the external memory storage. One example of this is lifelong learning scenario where the model receives unlimited length of data stream as an input which contains vast majority of uninformative entries. We tackle this problem by proposing a memory network fit for long-term lifelong learning scenario, which we refer to as Long-term Episodic Memory Networks (LEMN), that features a RNN-based retention agent that learns to replace less important memory entries based on the retention probability generated on each entry that is learned to identify data instances of generic importance relative to other memory entries, as well as its historical importance. Such learning of retention agent allows our long-term episodic memory network to retain memory entries of generic importance for a given task. We validate our model on a path-finding task as well as synthetic and real question answering tasks, on which our model achieves significant improvements over the memory augmented networks with rule-based memory scheduling as well as an RL-based baseline that does not consider relative or historical importance of the memory.

READ FULL TEXT
research
03/14/2019

Episodic Memory Reader: Learning What to Remember for Question Answering from Streaming Data

We consider a novel question answering (QA) task where the machine needs...
research
10/26/2017

Rotational Unit of Memory

The concepts of unitary evolution matrices and associative memory have b...
research
10/15/2014

Memory Networks

We describe a new class of learning models called memory networks. Memor...
research
07/17/2017

Visual Question Answering with Memory-Augmented Networks

This paper exploits a memory-augmented neural network to predict accurat...
research
06/06/2020

Do RNN and LSTM have Long Memory?

The LSTM network was proposed to overcome the difficulty in learning lon...
research
04/26/2023

Unleashing Infinite-Length Input Capacity for Large-scale Language Models with Self-Controlled Memory System

Large-scale Language Models (LLMs) are constrained by their inability to...
research
10/06/2020

Learning to Ignore: Long Document Coreference with Bounded Memory Neural Networks

Long document coreference resolution remains a challenging task due to t...

Please sign up or login with your details

Forgot password? Click here to reset