L4: Practical loss-based stepsize adaptation for deep learning

02/14/2018
by   Rolinek Michal, et al.
0

We propose a stepsize adaptation scheme for stochastic gradient descent. It operates directly with the loss function and rescales the gradient in order to make fixed predicted progress on the loss. We demonstrate its capabilities by strongly improving the performance of Adam and Momentum optimizers. The enhanced optimizers with default hyperparameters consistently outperform their constant stepsize counterparts, even the best ones, without a measurable increase in computational cost. The performance is validated on multiple architectures including ResNets and the Differential Neural Computer. A prototype implementation as a TensorFlow optimizer is released.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/01/2023

QLAB: Quadratic Loss Approximation-Based Optimal Learning Rate for Deep Learning

We propose a learning rate adaptation scheme, called QLAB, for descent o...
research
08/21/2022

Critical Bach Size Minimizes Stochastic First-Order Oracle Complexity of Deep Learning Optimizer using Hyperparameters Close to One

Practical results have shown that deep learning optimizers using small c...
research
03/28/2019

PAL: A fast DNN optimization method based on curvature information

We present a novel optimizer for deep neural networks that combines the ...
research
12/15/2022

Neuroevolution Surpasses Stochastic Gradient Descent for Physics-Informed Neural Networks

The potential of learned models for fundamental scientific research and ...
research
11/09/2022

Resource frugal optimizer for quantum machine learning

Quantum-enhanced data science, also known as quantum machine learning (Q...
research
09/14/2020

Deforming the Loss Surface to Affect the Behaviour of the Optimizer

In deep learning, it is usually assumed that the optimization process is...
research
06/24/2016

Stochastic Multiple Choice Learning for Training Diverse Deep Ensembles

Many practical perception systems exist within larger processes that inc...

Please sign up or login with your details

Forgot password? Click here to reset