Inducing Neural Collapse to a Fixed Hierarchy-Aware Frame for Reducing Mistake Severity

03/10/2023
by   Tong Liang, et al.
0

There is a recently discovered and intriguing phenomenon called Neural Collapse: at the terminal phase of training a deep neural network for classification, the within-class penultimate feature means and the associated classifier vectors of all flat classes collapse to the vertices of a simplex Equiangular Tight Frame (ETF). Recent work has tried to exploit this phenomenon by fixing the related classifier weights to a pre-computed ETF to induce neural collapse and maximize the separation of the learned features when training with imbalanced data. In this work, we propose to fix the linear classifier of a deep neural network to a Hierarchy-Aware Frame (HAFrame), instead of an ETF, and use a cosine similarity-based auxiliary loss to learn hierarchy-aware penultimate features that collapse to the HAFrame. We demonstrate that our approach reduces the mistake severity of the model's predictions while maintaining its top-1 accuracy on several datasets of varying scales with hierarchies of heights ranging from 3 to 12. We will release our code on GitHub in the near future.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/17/2022

Do We Really Need a Learnable Classifier at the End of Deep Neural Network?

Modern deep neural networks for classification usually jointly learn a b...
research
07/26/2022

Learning Hierarchy Aware Features for Reducing Mistake Severity

Label hierarchies are often available apriori as part of biological taxo...
research
02/20/2021

Inducing a hierarchy for multi-class classification problems

In applications where categorical labels follow a natural hierarchy, cla...
research
08/20/2018

Class2Str: End to End Latent Hierarchy Learning

Deep neural networks for image classification typically consists of a co...
research
01/03/2023

Understanding Imbalanced Semantic Segmentation Through Neural Collapse

A recent study has shown a phenomenon called neural collapse in that the...
research
04/01/2021

No Cost Likelihood Manipulation at Test Time for Making Better Mistakes in Deep Networks

There has been increasing interest in building deep hierarchy-aware clas...
research
10/19/2022

Hierarchical classification at multiple operating points

Many classification problems consider classes that form a hierarchy. Cla...

Please sign up or login with your details

Forgot password? Click here to reset