Understanding Weight Similarity of Neural Networks via Chain Normalization Rule and Hypothesis-Training-Testing

08/08/2022
by   Guangcong Wang, et al.
1

We present a weight similarity measure method that can quantify the weight similarity of non-convex neural networks. To understand the weight similarity of different trained models, we propose to extract the feature representation from the weights of neural networks. We first normalize the weights of neural networks by introducing a chain normalization rule, which is used for weight representation learning and weight similarity measure. We extend the traditional hypothesis-testing method to a hypothesis-training-testing statistical inference method to validate the hypothesis on the weight similarity of neural networks. With the chain normalization rule and the new statistical inference, we study the weight similarity measure on Multi-Layer Perceptron (MLP), Convolutional Neural Network (CNN), and Recurrent Neural Network (RNN), and find that the weights of an identical neural network optimized with the Stochastic Gradient Descent (SGD) algorithm converge to a similar local solution in a metric space. The weight similarity measure provides more insight into the local solutions of neural networks. Experiments on several datasets consistently validate the hypothesis of weight similarity measure.

READ FULL TEXT

page 7

page 8

research
10/29/2018

On the Convergence Rate of Training Recurrent Neural Networks

Despite the huge success of deep learning, our understanding to how the ...
research
08/02/2019

Network with Sub-Networks

We introduce network with sub-network, a neural network which their weig...
research
09/03/2020

It's Hard for Neural Networks To Learn the Game of Life

Efforts to improve the learning abilities of neural networks have focuse...
research
01/08/2021

Towards Accelerating Training of Batch Normalization: A Manifold Perspective

Batch normalization (BN) has become a crucial component across diverse d...
research
06/03/2022

Canonical convolutional neural networks

We introduce canonical weight normalization for convolutional neural net...
research
01/26/2023

Permutation-based Hypothesis Testing for Neural Networks

Neural networks are powerful predictive models, but they provide little ...
research
10/29/2022

Strong Lottery Ticket Hypothesis with ε–perturbation

The strong Lottery Ticket Hypothesis (LTH) claims the existence of a sub...

Please sign up or login with your details

Forgot password? Click here to reset