Imbalance Trouble: Revisiting Neural-Collapse Geometry

08/10/2022
by   Christos Thrampoulidis, et al.
0

Neural Collapse refers to the remarkable structural properties characterizing the geometry of class embeddings and classifier weights, found by deep nets when trained beyond zero training error. However, this characterization only holds for balanced data. Here we thus ask whether it can be made invariant to class imbalances. Towards this end, we adopt the unconstrained-features model (UFM), a recent theoretical model for studying neural collapse, and introduce Simplex-Encoded-Labels Interpolation (SELI) as an invariant characterization of the neural collapse phenomenon. Specifically, we prove for the UFM with cross-entropy loss and vanishing regularization that, irrespective of class imbalances, the embeddings and classifiers always interpolate a simplex-encoded label matrix and that their individual geometries are determined by the SVD factors of this same label matrix. We then present extensive experiments on synthetic and real datasets that confirm convergence to the SELI geometry. However, we caution that convergence worsens with increasing imbalances. We theoretically support this finding by showing that unlike the balanced case, when minorities are present, ridge-regularization plays a critical role in tweaking the geometry. This defines new questions and motivates further investigations into the impact of class imbalances on the rates at which first-order methods converge to their asymptotically preferred solutions.

READ FULL TEXT
research
06/13/2023

Supervised-Contrastive Loss Learns Orthogonal Frames and Batching Matters

Supervised contrastive loss (SCL) is a competitive and often superior al...
research
03/14/2023

On the Implicit Geometry of Cross-Entropy Parameterizations for Label-Imbalanced Data

Various logit-adjusted parameterizations of the cross-entropy (CE) loss ...
research
09/18/2023

Neural Collapse for Unconstrained Feature Model under Cross-entropy Loss with Imbalanced Data

Recent years have witnessed the huge success of deep neural networks (DN...
research
11/25/2019

Rethinking Softmax with Cross-Entropy: Neural Network Classifier as Mutual Information Estimator

Mutual information is widely applied to learn latent representations of ...
research
06/05/2023

Deep Learning From Crowdsourced Labels: Coupled Cross-entropy Minimization, Identifiability, and Regularization

Using noisy crowdsourced labels from multiple annotators, a deep learnin...
research
10/22/2019

J Regularization Improves Imbalanced Multiclass Segmentation

We propose a new loss formulation to further advance the multiclass segm...
research
03/18/2022

Unbiased Subclass Regularization for Semi-Supervised Semantic Segmentation

Semi-supervised semantic segmentation learns from small amounts of label...

Please sign up or login with your details

Forgot password? Click here to reset