Neural (Tangent Kernel) Collapse

05/25/2023
by   Mariia Seleznova, et al.
0

This work bridges two important concepts: the Neural Tangent Kernel (NTK), which captures the evolution of deep neural networks (DNNs) during training, and the Neural Collapse (NC) phenomenon, which refers to the emergence of symmetry and structure in the last-layer features of well-trained classification DNNs. We adopt the natural assumption that the empirical NTK develops a block structure aligned with the class labels, i.e., samples within the same class have stronger correlations than samples from different classes. Under this assumption, we derive the dynamics of DNNs trained with mean squared (MSE) loss and break them into interpretable phases. Moreover, we identify an invariant that captures the essence of the dynamics, and use it to prove the emergence of NC in DNNs with block-structured NTK. We provide large-scale numerical experiments on three common DNN architectures and three benchmark datasets to support our theory.

READ FULL TEXT
research
08/10/2018

Hierarchical Block Sparse Neural Networks

Sparse deep neural networks(DNNs) are efficient in both memory and compu...
research
06/13/2020

Understanding Learning Dynamics of Binary Neural Networks via Information Bottleneck

Compact neural networks are essential for affordable and power efficient...
research
06/21/2019

Theory of the Frequency Principle for General Deep Neural Networks

Along with fruitful applications of Deep Neural Networks (DNNs) to reali...
research
04/14/2021

Do Neural Network Weights account for Classes Centers?

The exploitation of Deep Neural Networks (DNNs) as descriptors in featur...
research
06/01/2020

One Versus all for deep Neural Network Incertitude (OVNNI) quantification

Deep neural networks (DNNs) are powerful learning models yet their resul...
research
06/01/2020

Emergence of Separable Manifolds in Deep Language Representations

Deep neural networks (DNNs) have shown much empirical success in solving...
research
02/16/2022

The learning phases in NN: From Fitting the Majority to Fitting a Few

The learning dynamics of deep neural networks are subject to controversy...

Please sign up or login with your details

Forgot password? Click here to reset