Neuron Campaign for Initialization Guided by Information Bottleneck Theory

08/14/2021
by   Haitao Mao, et al.
0

Initialization plays a critical role in the training of deep neural networks (DNN). Existing initialization strategies mainly focus on stabilizing the training process to mitigate gradient vanish/explosion problems. However, these initialization methods are lacking in consideration about how to enhance generalization ability. The Information Bottleneck (IB) theory is a well-known understanding framework to provide an explanation about the generalization of DNN. Guided by the insights provided by IB theory, we design two criteria for better initializing DNN. And we further design a neuron campaign initialization algorithm to efficiently select a good initialization for a neural network on a given dataset. The experiments on MNIST dataset show that our method can lead to a better generalization performance with faster convergence.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/13/2018

Understanding training and generalization in deep learning by Fourier analysis

Background: It is still an open research area to theoretically understan...
research
01/07/2019

Generalization in Deep Networks: The Role of Distance from Initialization

Why does training deep neural networks using stochastic gradient descent...
research
04/20/2020

Revisiting Initialization of Neural Networks

Good initialization of weights is crucial for effective training of deep...
research
04/16/2020

Deep Neural Network (DNN) for Water/Fat Separation: Supervised Training, Unsupervised Training, and No Training

Purpose: To use a deep neural network (DNN) for solving the optimization...
research
03/27/2019

A Sober Look at Neural Network Initializations

Initializing the weights and the biases is a key part of the training pr...
research
03/22/2022

On the (Non-)Robustness of Two-Layer Neural Networks in Different Learning Regimes

Neural networks are known to be highly sensitive to adversarial examples...
research
12/11/2019

Is Feature Diversity Necessary in Neural Network Initialization?

Standard practice in training neural networks involves initializing the ...

Please sign up or login with your details

Forgot password? Click here to reset