REPAIR: REnormalizing Permuted Activations for Interpolation Repair

11/15/2022
by   Keller Jordan, et al.
0

In this paper we look into the conjecture of Entezari et al.(2021) which states that if the permutation invariance of neural networks is taken into account, then there is likely no loss barrier to the linear interpolation between SGD solutions. First, we observe that neuron alignment methods alone are insufficient to establish low-barrier linear connectivity between SGD solutions due to a phenomenon we call variance collapse: interpolated deep networks suffer a collapse in the variance of their activations, causing poor performance. Next, we propose REPAIR (REnormalizing Permuted Activations for Interpolation Repair) which mitigates variance collapse by rescaling the preactivations of such interpolated networks. We explore the interaction between our method and the choice of normalization layer, network width, and depth, and demonstrate that using REPAIR on top of neuron alignment methods leads to 60 architecture families and tasks. In particular, we report a 74 reduction for ResNet50 on ImageNet and 90 CIFAR10.

READ FULL TEXT
research
10/12/2021

The Role of Permutation Invariance in Linear Mode Connectivity of Neural Networks

In this paper, we conjecture that if the permutation invariance of neura...
research
09/15/2022

Random initialisations performing above chance and how to find them

Neural networks trained with stochastic gradient descent (SGD) starting ...
research
09/05/2020

Optimizing Mode Connectivity via Neuron Alignment

The loss landscapes of deep neural networks are not well understood due ...
research
11/18/2020

NeVer 2.0: Learning, Verification and Repair of Deep Neural Networks

In this work, we present an early prototype of NeVer 2.0, a new system f...
research
07/15/2020

Convexifying Sparse Interpolation with Infinitely Wide Neural Networks: An Atomic Norm Approach

This work examines the problem of exact data interpolation via sparse (n...
research
05/29/2023

A Rainbow in Deep Network Black Boxes

We introduce rainbow networks as a probabilistic model of trained deep n...
research
11/21/2018

Regularizing by the Variance of the Activations' Sample-Variances

Normalization techniques play an important role in supporting efficient ...

Please sign up or login with your details

Forgot password? Click here to reset