Removing the Feature Correlation Effect of Multiplicative Noise

09/19/2018
by   Zijun Zhang, et al.
0

Multiplicative noise, including dropout, is widely used to regularize deep neural networks (DNNs), and is shown to be effective in a wide range of architectures and tasks. From an information perspective, we consider injecting multiplicative noise into a DNN as training the network to solve the task with noisy information pathways, which leads to the observation that multiplicative noise tends to increase the correlation between features, so as to increase the signal-to-noise ratio of information pathways. However, high feature correlation is undesirable, as it increases redundancy in representations. In this work, we propose non-correlating multiplicative noise (NCMN), which exploits batch normalization to remove the correlation effect in a simple yet effective way. We show that NCMN significantly improves the performance of standard multiplicative noise on image classification tasks, providing a better alternative to dropout for batch-normalized networks. Additionally, we present a unified view of NCMN and shake-shake regularization, which explains the performance gain of the latter.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/10/2015

A Scale Mixture Perspective of Multiplicative Noise in Neural Networks

Corrupting the input and hidden layers of deep neural networks (DNNs) wi...
research
12/20/2022

Walking Noise: Understanding Implications of Noisy Computations on Classification Tasks

Machine learning methods like neural networks are extremely successful a...
research
10/16/2018

Supplementary Material for "Estimation of a Multiplicative Correlation Structure in the Large Dimensional Case"

Supplementary Material for "Estimation of a Multiplicative Correlation S...
research
11/01/2018

Critical initialisation for deep signal propagation in noisy rectifier neural networks

Stochastic regularisation is an important weapon in the arsenal of a dee...
research
05/22/2023

Parameter estimation from an Ornstein-Uhlenbeck process with measurement noise

This article aims to investigate the impact of noise on parameter fittin...
research
10/09/2018

Unifying the Dropout Family Through Structured Shrinkage Priors

Dropout regularization of deep neural networks has been a mysterious yet...
research
11/04/2016

Information Dropout: Learning Optimal Representations Through Noisy Computation

The cross-entropy loss commonly used in deep learning is closely related...

Please sign up or login with your details

Forgot password? Click here to reset