Robust Implicit Backpropagation

08/07/2018
by   Francois Fagan, et al.
0

Arguably the biggest challenge in applying neural networks is tuning the hyperparameters, in particular the learning rate. The sensitivity to the learning rate is due to the reliance on backpropagation to train the network. In this paper we present the first application of Implicit Stochastic Gradient Descent (ISGD) to train neural networks, a method known in convex optimization to be unconditionally stable and robust to the learning rate. Our key contribution is a novel layer-wise approximation of ISGD which makes its updates tractable for neural networks. Experiments show that our method is more robust to high learning rates and generally outperforms standard backpropagation on a variety of tasks.

READ FULL TEXT
research
10/02/2018

Learning with Random Learning Rates

Hyperparameter tuning is a bothersome step in the training of deep learn...
research
06/25/2020

Learning compositional functions via multiplicative weight updates

Compositionality is a basic structural feature of both biological and ar...
research
01/29/2022

A Stochastic Bundle Method for Interpolating Networks

We propose a novel method for training deep neural networks that are cap...
research
01/06/2020

Self learning robot using real-time neural networks

With the advancements in high volume, low precision computational techno...
research
05/22/2017

Training Deep Networks without Learning Rates Through Coin Betting

Deep learning methods achieve state-of-the-art performance in many appli...
research
09/11/2019

An Implicit Form of Krasulina's k-PCA Update without the Orthonormality Constraint

We shed new insights on the two commonly used updates for the online k-P...
research
09/11/2014

Selection of Most Appropriate Backpropagation Training Algorithm in Data Pattern Recognition

There are several training algorithms for backpropagation method in neur...

Please sign up or login with your details

Forgot password? Click here to reset