On the Implicit Geometry of Cross-Entropy Parameterizations for Label-Imbalanced Data

03/14/2023
by   Tina Behnia, et al.
0

Various logit-adjusted parameterizations of the cross-entropy (CE) loss have been proposed as alternatives to weighted CE for training large models on label-imbalanced data far beyond the zero train error regime. The driving force behind those designs has been the theory of implicit bias, which for linear(ized) models, explains why they successfully induce bias on the optimization path towards solutions that favor minorities. Aiming to extend this theory to non-linear models, we investigate the implicit geometry of classifiers and embeddings that are learned by different CE parameterizations. Our main result characterizes the global minimizers of a non-convex cost-sensitive SVM classifier for the unconstrained features model, which serves as an abstraction of deep nets. We derive closed-form formulas for the angles and norms of classifiers and embeddings as a function of the number of classes, the imbalance and the minority ratios, and the loss hyperparameters. Using these, we show that logit-adjusted parameterizations can be appropriately tuned to learn symmetric geometries irrespective of the imbalance ratio. We complement our analysis with experiments and an empirical study of convergence accuracy in deep-nets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/10/2022

Imbalance Trouble: Revisiting Neural-Collapse Geometry

Neural Collapse refers to the remarkable structural properties character...
research
06/25/2022

On how to avoid exacerbating spurious correlations when models are overparameterized

Overparameterized models fail to generalize well in the presence of data...
research
09/18/2023

Neural Collapse for Unconstrained Feature Model under Cross-entropy Loss with Imbalanced Data

Recent years have witnessed the huge success of deep neural networks (DN...
research
03/02/2021

Label-Imbalanced and Group-Sensitive Classification under Overparameterization

Label-imbalanced and group-sensitive classification seeks to appropriate...
research
06/03/2021

Neural Collapse Under MSE Loss: Proximity to and Dynamics on the Central Path

Recent work [Papyan, Han, and Donoho, 2020] discovered a phenomenon call...
research
01/04/2022

AutoBalance: Optimized Loss Functions for Imbalanced Data

Imbalanced datasets are commonplace in modern machine learning problems....
research
06/13/2023

Supervised-Contrastive Loss Learns Orthogonal Frames and Batching Matters

Supervised contrastive loss (SCL) is a competitive and often superior al...

Please sign up or login with your details

Forgot password? Click here to reset