Squared Earth Mover's Distance-based Loss for Training Deep Neural Networks

11/17/2016
by   Le Hou, et al.
0

In the context of single-label classification, despite the huge success of deep learning, the commonly used cross-entropy loss function ignores the intricate inter-class relationships that often exist in real-life tasks such as age classification. In this work, we propose to leverage these relationships between classes by training deep nets with the exact squared Earth Mover's Distance (also known as Wasserstein distance) for single-label classification. The squared EMD loss uses the predicted probabilities of all classes and penalizes the miss-predictions according to a ground distance matrix that quantifies the dissimilarities between classes. We demonstrate that on datasets with strong inter-class relationships such as an ordering between classes, our exact squared EMD losses yield new state-of-the-art results. Furthermore, we propose a method to automatically learn this matrix using the CNN's own features during training. We show that our method can learn a ground distance matrix efficiently with no inter-class relationship priors and yield the same performance gain. Finally, we show that our method can be generalized to applications that lack strong inter-class relationships and still maintain state-of-the-art performance. Therefore, with limited computational overhead, one can always deploy the proposed loss function on any dataset over the conventional cross-entropy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/16/2019

On the Separability of Classes with the Cross-Entropy Loss Function

In this paper, we focus on the separability of classes with the cross-en...
research
11/22/2016

Relaxed Earth Mover's Distances for Chain- and Tree-connected Spaces and their use as a Loss Function in Deep Learning

The Earth Mover's Distance (EMD) computes the optimal cost of transformi...
research
05/01/2019

Introducing Graph Smoothness Loss for Training Deep Learning Architectures

We introduce a novel loss function for training deep learning architectu...
research
09/04/2020

Imbalanced Image Classification with Complement Cross Entropy

Recently, deep learning models have achieved great success in computer v...
research
11/03/2019

Conservative Wasserstein Training for Pose Estimation

This paper targets the task with discrete and periodic class labels (e.g...
research
03/11/2023

Generalizing and Decoupling Neural Collapse via Hyperspherical Uniformity Gap

The neural collapse (NC) phenomenon describes an underlying geometric sy...
research
08/11/2020

Reinforced Wasserstein Training for Severity-Aware Semantic Segmentation in Autonomous Driving

Semantic segmentation is important for many real-world systems, e.g., au...

Please sign up or login with your details

Forgot password? Click here to reset