A New Family of Generalization Bounds Using Samplewise Evaluated CMI

10/12/2022
by   Fredrik Hellström, et al.
0

We present a new family of information-theoretic generalization bounds, in which the training loss and the population loss are compared through a jointly convex function. This function is upper-bounded in terms of the disintegrated, samplewise, evaluated conditional mutual information (CMI), an information measure that depends on the losses incurred by the selected hypothesis, rather than on the hypothesis itself, as is common in probably approximately correct (PAC)-Bayesian results. We demonstrate the generality of this framework by recovering and extending previously known information-theoretic bounds. Furthermore, using the evaluated CMI, we derive a samplewise, average version of Seeger's PAC-Bayesian bound, where the convex function is the binary KL divergence. In some scenarios, this novel bound results in a tighter characterization of the population loss of deep neural networks than previous bounds. Finally, we derive high-probability versions of some of these average bounds. We demonstrate the unifying nature of the evaluated CMI bounds by using them to recover average and high-probability generalization bounds for multiclass classification with finite Natarajan dimension.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/16/2020

Generalization Bounds via Information Density and Conditional Information Density

We present a general approach, based on an exponential inequality, to de...
research
05/18/2023

A unified framework for information-theoretic generalization bounds

This paper presents a general methodology for deriving information-theor...
research
03/02/2022

Chained Generalisation Bounds

This work discusses how to derive upper bounds for the expected generali...
research
06/21/2023

More PAC-Bayes bounds: From bounded losses, to losses with general tail behaviors, to anytime-validity

In this paper, we present new high-probability PAC-Bayes bounds for diff...
research
04/20/2020

Generalization Error Bounds via mth Central Moments of the Information Density

We present a general approach to deriving bounds on the generalization e...
research
10/22/2020

Nonvacuous Loss Bounds with Fast Rates for Neural Networks via Conditional Information Measures

We present a framework to derive bounds on the test loss of randomized l...
research
05/01/2023

Exactly Tight Information-Theoretic Generalization Error Bound for the Quadratic Gaussian Problem

We provide a new information-theoretic generalization error bound that i...

Please sign up or login with your details

Forgot password? Click here to reset