Sequential Recurrent Neural Networks for Language Modeling

03/23/2017
by   Youssef Oualil, et al.
0

Feedforward Neural Network (FNN)-based language models estimate the probability of the next word based on the history of the last N words, whereas Recurrent Neural Networks (RNN) perform the same task based only on the last word and some context information that cycles in the network. This paper presents a novel approach, which bridges the gap between these two categories of networks. In particular, we propose an architecture which takes advantage of the explicit, sequential enumeration of the word history in FNN structure while enhancing each word representation at the projection layer through recurrent context information that evolves in the network. The context integration is performed using an additional word-dependent weight matrix that is also learned during the training. Extensive experiments conducted on the Penn Treebank (PTB) and the Large Text Compression Benchmark (LTCB) corpus showed a significant reduction of the perplexity when compared to state-of-the-art feedforward as well as recurrent neural network architectures.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/23/2017

A Neural Network Approach for Mixing Language Models

The performance of Neural Network (NN)-based language models is steadily...
research
03/17/2015

genCNN: A Convolutional Architecture for Word Sequence Prediction

We propose a novel convolutional architecture, named genCNN, for word se...
research
08/24/2017

A Study on Neural Network Language Modeling

An exhaustive study on neural network language modeling (NNLM) is perfor...
research
08/18/2017

Future Word Contexts in Neural Network Language Models

Recently, bidirectional recurrent network language models (bi-RNNLMs) ha...
research
02/13/2019

Which Neural Network Architecture matches Human Behavior in Artificial Grammar Learning?

In recent years artificial neural networks achieved performance close to...
research
01/27/2019

Variational Smoothing in Recurrent Neural Network Language Models

We present a new theoretical perspective of data noising in recurrent ne...
research
12/23/2016

Language Modeling with Gated Convolutional Networks

The pre-dominant approach to language modeling to date is based on recur...

Please sign up or login with your details

Forgot password? Click here to reset