Volumization as a Natural Generalization of Weight Decay

03/25/2020
by   Liu Ziyin, et al.
2

We propose a novel regularization method, called volumization, for neural networks. Inspired by physics, we define a physical volume for the weight parameters in neural networks, and we show that this method is an effective way of regularizing neural networks. Intuitively, this method interpolates between an L_2 and L_∞ regularization. Therefore, weight decay and weight clipping become special cases of the proposed algorithm. We prove, on a toy example, that the essence of this method is a regularization technique to control bias-variance tradeoff. The method is shown to do well in the categories where the standard weight decay method is shown to work well, including improving the generalization of networks and preventing memorization. Moreover, we show that the volumization might lead to a simple method for training a neural network whose weight is binary or ternary.

READ FULL TEXT

page 6

page 12

page 13

page 16

page 17

research
10/29/2018

Three Mechanisms of Weight Decay Regularization

Weight decay is one of the standard tricks in the neural network toolbox...
research
11/23/2020

Stable Weight Decay Regularization

Weight decay is a popular regularization technique for training of deep ...
research
07/21/2019

Adaptive Weight Decay for Deep Neural Networks

Regularization in the optimization of deep neural networks is often crit...
research
04/12/2022

NARX Identification using Derivative-Based Regularized Neural Networks

This work presents a novel regularization method for the identification ...
research
03/02/2020

Fiedler Regularization: Learning Neural Networks with Graph Sparsity

We introduce a novel regularization approach for deep learning that inco...
research
06/29/2023

Weight Compander: A Simple Weight Reparameterization for Regularization

Regularization is a set of techniques that are used to improve the gener...
research
06/06/2020

MMA Regularization: Decorrelating Weights of Neural Networks by Maximizing the Minimal Angles

The strong correlation between neurons or filters can significantly weak...

Please sign up or login with your details

Forgot password? Click here to reset