The Tree Loss: Improving Generalization with Many Classes

04/16/2022
by   Yujie Wang, et al.
0

Multi-class classification problems often have many semantically similar classes. For example, 90 of ImageNet's 1000 classes are for different breeds of dog. We should expect that these semantically similar classes will have similar parameter vectors, but the standard cross entropy loss does not enforce this constraint. We introduce the tree loss as a drop-in replacement for the cross entropy loss. The tree loss re-parameterizes the parameter matrix in order to guarantee that semantically similar classes will have similar parameter vectors. Using simple properties of stochastic gradient descent, we show that the tree loss's generalization error is asymptotically better than the cross entropy loss's. We then validate these theoretical results on synthetic data, image data (CIFAR100, ImageNet), and text data (Twitter).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/15/2020

Neural Collapse with Cross-Entropy Loss

We consider the variational problem of cross-entropy loss with n feature...
research
06/14/2022

Loss Functions for Classification using Structured Entropy

Cross-entropy loss is the standard metric used to train classification m...
research
02/08/2023

Cut your Losses with Squentropy

Nearly all practical neural models for classification are trained using ...
research
12/07/2021

Understanding Square Loss in Training Overparametrized Neural Network Classifiers

Deep learning has achieved many breakthroughs in modern classification t...
research
09/26/2019

Adaptive Class Weight based Dual Focal Loss for Improved Semantic Segmentation

In this paper, we propose a Dual Focal Loss (DFL) function, as a replace...
research
09/01/2017

Hierarchical loss for classification

Failing to distinguish between a sheepdog and a skyscraper should be wor...
research
10/18/2022

Multi-Source Transformer Architectures for Audiovisual Scene Classification

In this technical report, the systems we submitted for subtask 1B of the...

Please sign up or login with your details

Forgot password? Click here to reset