Gradient Correction beyond Gradient Descent

03/16/2022
by   Zefan Li, et al.
0

The great success neural networks have achieved is inseparable from the application of gradient-descent (GD) algorithms. Based on GD, many variant algorithms have emerged to improve the GD optimization process. The gradient for back-propagation is apparently the most crucial aspect for the training of a neural network. The quality of the calculated gradient can be affected by multiple aspects, e.g., noisy data, calculation error, algorithm limitation, and so on. To reveal gradient information beyond gradient descent, we introduce a framework (GCGD) to perform gradient correction. GCGD consists of two plug-in modules: 1) inspired by the idea of gradient prediction, we propose a GC-W module for weight gradient correction; 2) based on Neural ODE, we propose a GC-ODE module for hidden states gradient correction. Experiment results show that our gradient correction framework can effectively improve the gradient quality to reduce training epochs by ∼ 20% and also improve the network performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/22/2022

Langevin algorithms for Markovian Neural Networks and Deep Stochastic control

Stochastic Gradient Descent Langevin Dynamics (SGLD) algorithms, which a...
research
09/27/2020

Faster Biological Gradient Descent Learning

Back-propagation is a popular machine learning algorithm that uses gradi...
research
03/11/2019

Gradient Descent based Optimization Algorithms for Deep Learning Models Training

In this paper, we aim at providing an introduction to the gradient desce...
research
07/04/2022

Automating the Design and Development of Gradient Descent Trained Expert System Networks

Prior work introduced a gradient descent trained expert system that conc...
research
07/07/2021

Fast and Accurate Optimization of Metasurfaces with Gradient Descent and the Woodbury Matrix Identity

A fast metasurface optimization strategy for finite-size metasurfaces mo...
research
10/27/2017

Automated Design using Neural Networks and Gradient Descent

We propose a novel method that makes use of deep neural networks and gra...
research
05/14/2019

Task-Driven Data Verification via Gradient Descent

We introduce a novel algorithm for the detection of possible sample corr...

Please sign up or login with your details

Forgot password? Click here to reset