On the Learning Property of Logistic and Softmax Losses for Deep Neural Networks

03/04/2020
by   Xiangrui Li, et al.
37

Deep convolutional neural networks (CNNs) trained with logistic and softmax losses have made significant advancement in visual recognition tasks in computer vision. When training data exhibit class imbalances, the class-wise reweighted version of logistic and softmax losses are often used to boost performance of the unweighted version. In this paper, motivated to explain the reweighting mechanism, we explicate the learning property of those two loss functions by analyzing the necessary condition (e.g., gradient equals to zero) after training CNNs to converge to a local minimum. The analysis immediately provides us explanations for understanding (1) quantitative effects of the class-wise reweighting mechanism: deterministic effectiveness for binary classification using logistic loss yet indeterministic for multi-class classification using softmax loss; (2) disadvantage of logistic loss for single-label multi-class classification via one-vs.-all approach, which is due to the averaging effect on predicted probabilities for the negative class (e.g., non-target classes) in the learning process. With the disadvantage and advantage of logistic loss disentangled, we thereafter propose a novel reweighted logistic loss for multi-class classification. Our simple yet effective formulation improves ordinary logistic loss by focusing on learning hard non-target classes (target vs. non-target class in one-vs.-all) and turned out to be competitive with softmax loss. We evaluate our method on several benchmark datasets to demonstrate its effectiveness.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/20/2018

Single-Label Multi-Class Image Classification by Deep Logistic Regression

The objective learning formulation is essential for the success of convo...
research
08/06/2019

Explaining Convolutional Neural Networks using Softmax Gradient Layer-wise Relevance Propagation

Convolutional Neural Networks (CNN) have become state-of-the-art in the ...
research
08/03/2021

SphereFace2: Binary Classification is All You Need for Deep Face Recognition

State-of-the-art deep face recognition methods are mostly trained with a...
research
09/17/2019

Relaxed Softmax for learning from Positive and Unlabeled data

In recent years, the softmax model and its fast approximations have beco...
research
12/01/2015

Loss Functions for Top-k Error: Analysis and Insights

In order to push the performance on realistic computer vision tasks, the...
research
05/01/2021

Non-asymptotic Excess Risk Bounds for Classification with Deep Convolutional Neural Networks

In this paper, we consider the problem of binary classification with a c...
research
08/08/2020

HASeparator: Hyperplane-Assisted Softmax

Efficient feature learning with Convolutional Neural Networks (CNNs) con...

Please sign up or login with your details

Forgot password? Click here to reset