Data-driven Regularization via Racecar Training for Generalizing Neural Networks

06/30/2020
by   You Xie, et al.
0

We propose a novel training approach for improving the generalization in neural networks. We show that in contrast to regular constraints for orthogonality, our approach represents a data-dependent orthogonality constraint, and is closely related to singular value decompositions of the weight matrices. We also show how our formulation is easy to realize in practical network architectures via a reverse pass, which aims for reconstructing the full sequence of internal states of the network. Despite being a surprisingly simple change, we demonstrate that this forward-backward training approach, which we refer to as racecar training, leads to significantly more generic features being extracted from a given data set. Networks trained with our approach show more balanced mutual information between input and output throughout all layers, yield improved explainability and, exhibit improved performance for a variety of tasks and task transfers.

READ FULL TEXT

page 1

page 7

page 8

page 9

page 14

page 17

page 21

page 23

research
06/08/2022

Boundary between noise and information applied to filtering neural network weight matrices

Deep neural networks have been successfully applied to a broad range of ...
research
11/24/2021

Softmax Gradient Tampering: Decoupling the Backward Pass for Improved Fitting

We introduce Softmax Gradient Tampering, a technique for modifying the g...
research
03/23/2020

Sample-Specific Output Constraints for Neural Networks

Neural networks reach state-of-the-art performance in a variety of learn...
research
06/17/2021

Frustratingly Easy Transferability Estimation

Transferability estimation has been an essential tool in selecting a pre...
research
06/29/2020

Neural Mutual Information Estimation for Channel Coding: State-of-the-Art Estimators, Analysis, and Performance Comparison

Deep learning based physical layer design, i.e., using dense neural netw...
research
12/28/2018

On Computation and Generalization of GANs with Spectrum Control

Generative Adversarial Networks (GANs), though powerful, is hard to trai...
research
01/04/2022

Sparse Super-Regular Networks

It has been argued by Thom and Palm that sparsely-connected neural netwo...

Please sign up or login with your details

Forgot password? Click here to reset