Learning Sequence Representations by Non-local Recurrent Neural Memory

07/20/2022
by   Wenjie Pei, et al.
0

The key challenge of sequence representation learning is to capture the long-range temporal dependencies. Typical methods for supervised sequence representation learning are built upon recurrent neural networks to capture temporal dependencies. One potential limitation of these methods is that they only model one-order information interactions explicitly between adjacent time steps in a sequence, hence the high-order interactions between nonadjacent time steps are not fully exploited. It greatly limits the capability of modeling the long-range temporal dependencies since the temporal features learned by one-order interactions cannot be maintained for a long term due to temporal information dilution and gradient vanishing. To tackle this limitation, we propose the Non-local Recurrent Neural Memory (NRNM) for supervised sequence representation learning, which performs non-local operations by means of self-attention mechanism to learn full-order interactions within a sliding temporal memory block and models global interactions between memory blocks in a gated recurrent manner. Consequently, our model is able to capture long-range dependencies. Besides, the latent high-level features contained in high-order interactions can be distilled by our model. We validate the effectiveness and generalization of our NRNM on three types of sequence applications across different modalities, including sequence classification, step-wise sequential prediction and sequence similarity learning. Our model compares favorably against other state-of-the-art methods specifically designed for each of these sequence applications.

READ FULL TEXT

page 6

page 12

page 16

page 17

page 20

research
08/26/2019

Non-local Recurrent Neural Memory for Supervised Sequence Modeling

Typical methods for supervised sequence modeling are built upon the recu...
research
10/13/2021

Non-local Recurrent Regularization Networks for Multi-view Stereo

In deep multi-view stereo networks, cost regularization is crucial to ac...
research
04/08/2019

A Statistical Investigation of Long Memory in Language and Music

Representation and learning of long-range dependencies is a central chal...
research
09/14/2019

Temporal FiLM: Capturing Long-Range Sequence Dependencies with Feature-Wise Modulations

Learning representations that accurately capture long-range dependencies...
research
02/15/2017

Generative Temporal Models with Memory

We consider the general problem of modeling temporal data with long-rang...
research
06/23/2016

Algorithmic Composition of Melodies with Deep Recurrent Neural Networks

A big challenge in algorithmic composition is to devise a model that is ...
research
05/25/2023

Online learning of long-range dependencies

Online learning holds the promise of enabling efficient long-term credit...

Please sign up or login with your details

Forgot password? Click here to reset