A Heaviside Function Approximation for Neural Network Binary Classification

09/02/2020
by   Nathan Tsoi, et al.
23

Neural network binary classifiers are often evaluated on metrics like accuracy and F_1-Score, which are based on confusion matrix values (True Positives, False Positives, False Negatives, and True Negatives). However, these classifiers are commonly trained with a different loss, e.g. log loss. While it is preferable to perform training on the same loss as the evaluation metric, this is difficult in the case of confusion matrix based metrics because set membership is a step function without a derivative useful for backpropagation. To address this challenge, we propose an approximation of the step function that adheres to the properties necessary for effective training of binary networks using confusion matrix based metrics. This approach allows for end-to-end training of binary deep neural classifiers via batch gradient descent. We demonstrate the flexibility of this approach in several applications with varying levels of class imbalance. We also demonstrate how the approximation allows balancing between precision and recall in the appropriate ratio for the task at hand.

READ FULL TEXT

page 1

page 2

page 3

page 4

page 5

page 6

page 7

page 8

research
04/06/2022

Emphasis on the Minimization of False Negatives or False Positives in Binary Classification

The minimization of specific cases in binary classification, such as fal...
research
10/21/2022

Extending F_1 metric, probabilistic approach

This article explores the extension of well-known F_1 score used for ass...
research
09/06/2018

Yes, IoU loss is submodular - as a function of the mispredictions

This note is a response to [7] in which it is claimed that [13, Proposit...
research
04/30/2023

The MCC approaches the geometric mean of precision and recall as true negatives approach infinity

The performance of a binary classifier is described by a confusion matri...
research
02/21/2023

A Log-linear Gradient Descent Algorithm for Unbalanced Binary Classification using the All Pairs Squared Hinge Loss

Receiver Operating Characteristic (ROC) curves are plots of true positiv...
research
02/21/2021

Constrained Optimization for Training Deep Neural Networks Under Class Imbalance

Deep neural networks (DNNs) are notorious for making more mistakes for t...
research
06/22/2020

DeepTopPush: Simple and Scalable Method for Accuracy at the Top

Accuracy at the top is a special class of binary classification problems...

Please sign up or login with your details

Forgot password? Click here to reset