Reformulating van Rijsbergen's F_β metric for weighted binary cross-entropy

10/29/2022
by   Satesh Ramdhani, et al.
0

The separation of performance metrics from gradient based loss functions may not always give optimal results and may miss vital aggregate information. This paper investigates incorporating a performance metric alongside differentiable loss functions to inform training outcomes. The goal is to guide model performance and interpretation by assuming statistical distributions on this performance metric for dynamic weighting. The focus is on van Rijsbergens F_β metric – a popular choice for gauging classification performance. Through distributional assumptions on the F_β, an intermediary link can be established to the standard binary cross-entropy via dynamic penalty weights. First, the F_β metric is reformulated to facilitate assuming statistical distributions with accompanying proofs for the cumulative density function. These probabilities are used within a knee curve algorithm to find an optimal β or β_opt. This β_opt is used as a weight or penalty in the proposed weighted binary cross-entropy. Experimentation on publicly available data with imbalanced classes mostly yields better and interpretable results as compared to the baseline. For example, for the IMDB text data with known labeling errors, a 14 can accelerate training and provide better interpretation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/14/2022

Loss Functions for Classification using Structured Entropy

Cross-entropy loss is the standard metric used to train classification m...
research
04/03/2021

A surrogate loss function for optimization of F_β score in binary classification with imbalanced data

The F_β score is a commonly used measure of classification performance, ...
research
07/18/2019

On the relation between Loss Functions and T-Norms

Deep learning has been shown to achieve impressive results in several do...
research
01/03/2020

The Real-World-Weight Cross-Entropy Loss Function: Modeling the Costs of Mislabeling

In this paper, we propose a new metric to measure goodness-of-fit for cl...
research
01/03/2023

Effective and Efficient Training for Sequential Recommendation Using Cumulative Cross-Entropy Loss

Increasing research interests focus on sequential recommender systems, a...
research
09/12/2021

Mixing between the Cross Entropy and the Expectation Loss Terms

The cross entropy loss is widely used due to its effectiveness and solid...

Please sign up or login with your details

Forgot password? Click here to reset