The Principle of Logit Separation

05/29/2017
by   Gil Keren, et al.
0

We consider neural network training, in applications in which there are many possible classes, but at test-time, the task is to identify only whether the given example belongs to a specific class, which can be different in different applications of the classifier. For instance, this is the case in an image search engine. We consider the Single Logit Classification (SLC) task: training the network so that at test-time, it would be possible to accurately identify if the example belongs to a given class, based only on the output logit for this class. We propose a natural principle, the Principle of Logit Separation, as a guideline for choosing and designing losses suitable for the SLC. We show that the cross-entropy loss function is not aligned with the Principle of Logit Separation. In contrast, there are known loss functions, as well as novel batch loss functions that we propose, which are aligned with this principle. In total, we study seven loss functions. Our experiments show that indeed in almost all cases, losses that are aligned with Principle of Logit Separation obtain a 20 losses that are not aligned with it. We therefore conclude that the Principle of Logit Separation sheds light on an important property of the most common loss functions used by neural network classifiers. Tensorflow code for optimizing the new batch losses is publicly available in https://github.com/cruvadom/Logit_Separation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/20/2022

Test-Time Adaptation via Conjugate Pseudo-labels

Test-time adaptation (TTA) refers to adapting neural networks to distrib...
research
09/13/2022

On the Optimal Combination of Cross-Entropy and Soft Dice Losses for Lesion Segmentation with Out-of-Distribution Robustness

We study the impact of different loss functions on lesion segmentation f...
research
03/06/2018

The Contextual Loss for Image Transformation with Non-Aligned Data

Feed-forward CNNs trained for image transformation problems rely on loss...
research
01/25/2019

Towards a Deeper Understanding of Adversarial Losses

Recent work has proposed various adversarial losses for training generat...
research
08/06/2022

Towards Robust Deep Learning using Entropic Losses

Current deep learning solutions are well known for not informing whether...
research
02/17/2022

General Cyclical Training of Neural Networks

This paper describes the principle of "General Cyclical Training" in mac...
research
11/13/2020

Neural Network Training Techniques Regularize Optimization Trajectory: An Empirical Study

Modern deep neural network (DNN) trainings utilize various training tech...

Please sign up or login with your details

Forgot password? Click here to reset