Analytic expressions for the output evolution of a deep neural network

12/18/2019
by   Anastasia Borovykh, et al.
0

We present a novel methodology based on a Taylor expansion of the network output for obtaining analytical expressions for the expected value of the network weights and output under stochastic training. Using these analytical expressions the effects of the hyperparameters and the noise variance of the optimization algorithm on the performance of the deep neural network are studied. In the early phases of training with a small noise coefficient, the output is equivalent to a linear model. In this case the network can generalize better due to the noise preventing the output from fully converging on the train data, however the noise does not result in any explicit regularization. In the later training stages, when higher order approximations are required, the impact of the noise becomes more significant, i.e. in a model which is non-linear in the weights noise can regularize the output function resulting in better generalization as witnessed by its influence on the weight Hessian, a commonly used metric for generalization capabilities.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/24/2022

Noise Injection as a Probe of Deep Learning Dynamics

We propose a new method to probe the learning mechanism of Deep Neural N...
research
04/16/2011

Adding noise to the input of a model trained with a regularized objective

Regularization is a well studied problem in the context of neural networ...
research
09/27/2018

Introducing Noise in Decentralized Training of Neural Networks

It has been shown that injecting noise into the neural network weights d...
research
12/19/2019

FQ-Conv: Fully Quantized Convolution for Efficient and Accurate Inference

Deep neural networks (DNNs) can be made hardware-efficient by reducing t...
research
07/20/2021

Edge of chaos as a guiding principle for modern neural network training

The success of deep neural networks in real-world problems has prompted ...
research
03/13/2017

Comparison of echo state network output layer classification methods on noisy data

Echo state networks are a recently developed type of recurrent neural ne...
research
08/04/2023

A stochastic optimization approach to train non-linear neural networks with a higher-order variation regularization

While highly expressive parametric models including deep neural networks...

Please sign up or login with your details

Forgot password? Click here to reset