Redesigning the classification layer by randomizing the class representation vectors

11/16/2020
by   Gabi Shalev, et al.
0

Neural image classification models typically consist of two components. The first is an image encoder, which is responsible for encoding a given raw image into a representative vector. The second is the classification component, which is often implemented by projecting the representative vector onto target class vectors. The target class vectors, along with the rest of the model parameters, are estimated so as to minimize the loss function. In this paper, we analyze how simple design choices for the classification layer affect the learning dynamics. We show that the standard cross-entropy training implicitly captures visual similarities between different classes, which might deteriorate accuracy or even prevents some models from converging. We propose to draw the class vectors randomly and set them as fixed during training, thus invalidating the visual similarities encoded in these vectors. We analyze the effects of keeping the class vectors fixed and show that it can increase the inter-class separability, intra-class compactness, and the overall model accuracy, while maintaining the robustness to image corruptions and the generalization of the learned concepts.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/12/2020

CC-Loss: Channel Correlation Loss For Image Classification

The loss function is a key component in deep learning models. A commonly...
research
03/06/2020

SimLoss: Class Similarities in Cross Entropy

One common loss function in neural network classification tasks is Categ...
research
11/28/2019

A Discriminative Learned CNN Embedding For Remote Senseing Image Scene Classification

In this work, a discriminatively learned CNN embedding is proposed for r...
research
11/28/2019

A Discriminative Learned CNN Embedding for Remote Sensing Image Scene Classification

In this work, a discriminatively learned CNN embedding is proposed for r...
research
09/22/2022

CAMRI Loss: Improving Recall of a Specific Class without Sacrificing Accuracy

In real-world applications of multi-class classification models, misclas...
research
06/16/2019

Mixture separability loss in a deep convolutional network for image classification

In machine learning, the cost function is crucial because it measures ho...
research
03/11/2023

Generalizing and Decoupling Neural Collapse via Hyperspherical Uniformity Gap

The neural collapse (NC) phenomenon describes an underlying geometric sy...

Please sign up or login with your details

Forgot password? Click here to reset