Gradient Descent based Optimization Algorithms for Deep Learning Models Training

03/11/2019
by   Jiawei Zhang, et al.
0

In this paper, we aim at providing an introduction to the gradient descent based optimization algorithms for learning deep neural network models. Deep learning models involving multiple nonlinear projection layers are very challenging to train. Nowadays, most of the deep learning model training still relies on the back propagation algorithm actually. In back propagation, the model variables will be updated iteratively until convergence with gradient descent based optimization algorithms. Besides the conventional vanilla gradient descent algorithm, many gradient descent variants have also been proposed in recent years to improve the learning performance, including Momentum, Adagrad, Adam, Gadam, etc., which will all be introduced in this paper respectively.

READ FULL TEXT
research
04/19/2019

Derivative-Free Global Optimization Algorithms: Bayesian Method and Lipschitzian Approaches

In this paper, we will provide an introduction to the derivative-free op...
research
12/23/2022

Exploring the Optimized Value of Each Hyperparameter in Various Gradient Descent Algorithms

In the recent years, various gradient descent algorithms including the m...
research
06/15/2023

Stochastic Re-weighted Gradient Descent via Distributionally Robust Optimization

We develop a re-weighted gradient descent technique for boosting the per...
research
04/19/2019

Derivative-Free Global Optimization Algorithms: Population based Methods and Random Search Approaches

In this paper, we will provide an introduction to the derivative-free op...
research
06/02/2020

Acceleration of Descent-based Optimization Algorithms via Carathéodory's Theorem

We propose a new technique to accelerate algorithms based on Gradient De...
research
09/09/2023

A Gentle Introduction to Gradient-Based Optimization and Variational Inequalities for Machine Learning

The rapid progress in machine learning in recent years has been based on...
research
03/16/2022

Gradient Correction beyond Gradient Descent

The great success neural networks have achieved is inseparable from the ...

Please sign up or login with your details

Forgot password? Click here to reset