On the Analysis of Trajectories of Gradient Descent in the Optimization of Deep Neural Networks

07/21/2018
by   Adepu Ravi Sankar, et al.
0

Theoretical analysis of the error landscape of deep neural networks has garnered significant interest in recent years. In this work, we theoretically study the importance of noise in the trajectories of gradient descent towards optimal solutions in multi-layer neural networks. We show that adding noise (in different ways) to a neural network while training increases the rank of the product of weight matrices of a multi-layer linear neural network. We thus study how adding noise can assist reaching a global optimum when the product matrix is full-rank (under certain conditions). We establish theoretical foundations between the noise induced into the neural network - either to the gradient, to the architecture, or to the input/output to a neural network - and the rank of product of weight matrices. We corroborate our theoretical findings with empirical results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/01/2017

Theoretical Properties for Neural Networks with Weight Matrices of Low Displacement Rank

Recently low displacement rank (LDR) matrices, or so-called structured m...
research
05/27/2019

Fast Convergence of Natural Gradient Descent for Overparameterized Neural Networks

Natural gradient descent has proven effective at mitigating the effects ...
research
11/05/2018

PILAE: A Non-gradient Descent Learning Scheme for Deep Feedforward Neural Networks

In this work, a non-gradient descent learning scheme is proposed for dee...
research
02/09/2017

Energy Saving Additive Neural Network

In recent years, machine learning techniques based on neural networks fo...
research
12/14/2021

Training Multi-Layer Over-Parametrized Neural Network in Subquadratic Time

We consider the problem of training a multi-layer over-parametrized neur...
research
06/13/2022

Rank Diminishing in Deep Neural Networks

The rank of neural networks measures information flowing across layers. ...
research
06/07/2017

Are Saddles Good Enough for Deep Learning?

Recent years have seen a growing interest in understanding deep neural n...

Please sign up or login with your details

Forgot password? Click here to reset