PACMAN: PAC-style bounds accounting for the Mismatch between Accuracy and Negative log-loss

12/10/2021
by   Matías Vera, et al.
0

The ultimate performance of machine learning algorithms for classification tasks is usually measured in terms of the empirical error probability (or accuracy) based on a testing dataset. Whereas, these algorithms are optimized through the minimization of a typically different–more convenient–loss function based on a training set. For classification tasks, this loss function is often the negative log-loss that leads to the well-known cross-entropy risk which is typically better behaved (from a numerical perspective) than the error probability. Conventional studies on the generalization error do not usually take into account the underlying mismatch between losses at training and testing phases. In this work, we introduce an analysis based on point-wise PAC approach over the generalization gap considering the mismatch of testing based on the accuracy metric and training on the negative log-loss. We label this analysis PACMAN. Building on the fact that the mentioned mismatch can be written as a likelihood ratio, concentration inequalities can be used to provide some insights for the generalization problem in terms of some point-wise PAC bounds depending on some meaningful information-theoretic quantities. An analysis of the obtained bounds and a comparison with available results in the literature are also provided.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/27/2016

PAC-Bayesian Theory Meets Bayesian Inference

We exhibit a strong link between frequentist PAC-Bayesian risk bounds an...
research
05/04/2021

Information Complexity and Generalization Bounds

We present a unifying picture of PAC-Bayesian and mutual information-bas...
research
06/16/2020

PAC-Bayesian Generalization Bounds for MultiLayer Perceptrons

We study PAC-Bayesian generalization bounds for Multilayer Perceptrons (...
research
02/10/2021

Learning under Distribution Mismatch and Model Misspecification

We study learning algorithms when there is a mismatch between the distri...
research
05/01/2019

On Expected Accuracy

We empirically investigate the (negative) expected accuracy as an altern...
research
05/28/2019

Understanding the Behaviour of the Empirical Cross-Entropy Beyond the Training Distribution

Machine learning theory has mostly focused on generalization to samples ...
research
12/31/2022

Inequality for the variance of an asymmetric loss

We assume that the forecast error follows a probability distribution whi...

Please sign up or login with your details

Forgot password? Click here to reset