Improving Language Modeling using Densely Connected Recurrent Neural Networks

07/19/2017
by   Fréderic Godin, et al.
0

In this paper, we introduce the novel concept of densely connected layers into recurrent neural networks. We evaluate our proposed architecture on the Penn Treebank language modeling task. We show that we can obtain similar perplexity scores with six times fewer parameters compared to a standard stacked 2-layer LSTM model trained with dropout (Zaremba et al. 2014). In contrast with the current usage of skip connections, we show that densely connecting only a few stacked layers with skip connections already yields significant perplexity reductions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/25/2017

Dual Rectified Linear Units (DReLUs): A Replacement for Tanh Activation Functions in Quasi-Recurrent Neural Networks

In this paper, we introduce a novel type of Rectified Linear Unit (ReLU)...
research
11/05/2016

Quasi-Recurrent Neural Networks

Recurrent neural networks are a powerful tool for modeling sequential da...
research
01/18/2018

Sparsely Connected Convolutional Networks

Residual learning with skip connections permits training ultra-deep neur...
research
01/03/2017

Shortcut Sequence Tagging

Deep stacked RNNs are usually hard to train. Adding shortcut connections...
research
10/11/2016

An Empirical Exploration of Skip Connections for Sequential Tagging

In this paper, we empirically explore the effects of various kinds of sk...
research
12/23/2016

Language Modeling with Gated Convolutional Networks

The pre-dominant approach to language modeling to date is based on recur...
research
02/03/2018

Densely Connected Bidirectional LSTM with Applications to Sentence Classification

Deep neural networks have recently been shown to achieve highly competit...

Please sign up or login with your details

Forgot password? Click here to reset