Noisin: Unbiased Regularization for Recurrent Neural Networks

05/03/2018
by   Adji B. Dieng, et al.
0

Recurrent neural networks (RNNs) are powerful models of sequential data. They have been successfully used in domains such as text and speech. However, RNNs are susceptible to overfitting; regularization is important. In this paper we develop Noisin, a new method for regularizing RNNs. Noisin injects random noise into the hidden states of the RNN and then maximizes the corresponding marginal likelihood of the data. We show how Noisin applies to any RNN and we study many different types of noise. Noisin is unbiased--it preserves the underlying RNN on average. We characterize how Noisin regularizes its RNN both theoretically and empirically. On language modeling benchmarks, Noisin improves over dropout by as much as 12.2 also compared the state-of-the-art language model of Yang et al. 2017, both with and without Noisin. On the Penn Treebank, the method with Noisin more quickly reaches state-of-the-art performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/03/2017

Revisiting Activation Regularization for Language RNNs

Recurrent neural networks (RNNs) serve as a fundamental building block f...
research
01/23/2013

Regularization and nonlinearities for neural language models: when are they needed?

Neural language models (LMs) based on recurrent neural networks (RNN) ar...
research
11/11/2019

RNN-Test: Adversarial Testing Framework for Recurrent Neural Network Systems

While huge efforts have been investigated in the adversarial testing of ...
research
05/03/2021

Robust Learning of Recurrent Neural Networks in Presence of Exogenous Noise

Recurrent Neural networks (RNN) have shown promising potential for learn...
research
05/20/2018

Multi-Step Prediction of Dynamic Systems with Recurrent Neural Networks

Recurrent Neural Networks (RNNs) can encode rich dynamics which makes th...
research
09/25/2017

Predictive-State Decoders: Encoding the Future into Recurrent Networks

Recurrent neural networks (RNNs) are a vital modeling technique that rel...
research
09/06/2019

RNN Architecture Learning with Sparse Regularization

Neural models for NLP typically use large numbers of parameters to reach...

Please sign up or login with your details

Forgot password? Click here to reset