Attention-based Memory Selection Recurrent Network for Language Modeling

11/26/2016
by   Da-Rong Liu, et al.
0

Recurrent neural networks (RNNs) have achieved great success in language modeling. However, since the RNNs have fixed size of memory, their memory cannot store all the information about the words it have seen before in the sentence, and thus the useful long-term information may be ignored when predicting the next words. In this paper, we propose Attention-based Memory Selection Recurrent Network (AMSRN), in which the model can review the information stored in the memory at each previous time step and select the relevant information to help generate the outputs. In AMSRN, the attention mechanism finds the time steps storing the relevant information in the memory, and memory selection determines which dimensions of the memory are involved in computing the attention weights and from which the information is extracted.In the experiments, AMSRN outperformed long short-term memory (LSTM) based language models on both English and Chinese corpora. Moreover, we investigate using entropy as a regularizer for attention weights and visualize how the attention mechanism helps language modeling.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/30/2018

Long Short-Term Attention

In order to learn effective features from temporal sequences, the long s...
research
01/06/2016

Recurrent Memory Networks for Language Modeling

Recurrent Neural Networks (RNN) have obtained excellent result in many n...
research
02/15/2017

Frustratingly Short Attention Spans in Neural Language Modeling

Neural language models predict the next token using a latent representat...
research
05/31/2018

Attention-Based LSTM for Psychological Stress Detection from Spoken Language Using Distant Supervision

We propose a Long Short-Term Memory (LSTM) with attention mechanism to c...
research
10/10/2018

Persistence pays off: Paying Attention to What the LSTM Gating Mechanism Persists

Language Models (LMs) are important components in several Natural Langua...
research
09/22/2022

A Case Report On The "A.I. Locked-In Problem": social concerns with modern NLP

Modern NLP models are becoming better conversational agents than their p...
research
06/30/2020

Learning to Combine Top-Down and Bottom-Up Signals in Recurrent Neural Networks with Attention over Modules

Robust perception relies on both bottom-up and top-down signals. Bottom-...

Please sign up or login with your details

Forgot password? Click here to reset