Training Very Deep Networks

07/22/2015
by   Rupesh Kumar Srivastava, et al.
0

Theoretical and empirical evidence indicates that the depth of neural networks is crucial for their success. However, training becomes more difficult as depth increases, and training of very deep networks remains an open problem. Here we introduce a new architecture designed to overcome this. Our so-called highway networks allow unimpeded information flow across many layers on information highways. They are inspired by Long Short-Term Memory recurrent networks and use adaptive gating units to regulate the information flow. Even with hundreds of layers, highway networks can be trained directly through simple gradient descent. This enables the study of extremely deep and efficient architectures.

READ FULL TEXT
research
05/03/2015

Highway Networks

There is plenty of theoretical and empirical evidence that depth of neur...
research
03/30/2016

Deep Networks with Stochastic Depth

Very deep convolutional networks with hundreds of layers have led to sig...
research
05/23/2018

Highway State Gating for Recurrent Highway Networks: improving information flow through time

Recurrent Neural Networks (RNNs) play a major role in the field of seque...
research
08/19/2017

A Data and Model-Parallel, Distributed and Scalable Framework for Training of Deep Networks in Apache Spark

Training deep networks is expensive and time-consuming with the training...
research
11/21/2015

Adding Gradient Noise Improves Learning for Very Deep Networks

Deep feedforward and recurrent networks have achieved impressive results...
research
11/16/2017

A Forward-Backward Approach for Visualizing Information Flow in Deep Networks

We introduce a new, systematic framework for visualizing information flo...
research
08/11/2016

Faster Training of Very Deep Networks Via p-Norm Gates

A major contributing factor to the recent advances in deep neural networ...

Please sign up or login with your details

Forgot password? Click here to reset