Nonvacuous Loss Bounds with Fast Rates for Neural Networks via Conditional Information Measures

10/22/2020
by   Fredrik Hellström, et al.
0

We present a framework to derive bounds on the test loss of randomized learning algorithms for the case of bounded loss functions. This framework leads to bounds that depend on the conditional information density between the the output hypothesis and the choice of the training set, given a larger set of data samples from which the training set is formed. Furthermore, the bounds pertain to the average test loss as well as to its tail probability, both for the PAC-Bayesian and the single-draw settings. If the conditional information density is bounded uniformly in the size n of the training set, our bounds decay as 1/n, which is referred to as a fast rate. This is in contrast with the tail bounds involving conditional information measures available in the literature, which have a less benign 1/√(n) dependence. We demonstrate the usefulness of our tail bounds by showing that they lead to estimates of the test loss achievable with several neural network architectures trained on MNIST and Fashion-MNIST that match the state-of-the-art bounds available in the literature.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/16/2020

Generalization Bounds via Information Density and Conditional Information Density

We present a general approach, based on an exponential inequality, to de...
research
04/20/2020

Generalization Error Bounds via mth Central Moments of the Information Density

We present a general approach to deriving bounds on the generalization e...
research
06/21/2023

More PAC-Bayes bounds: From bounded losses, to losses with general tail behaviors, to anytime-validity

In this paper, we present new high-probability PAC-Bayes bounds for diff...
research
10/12/2022

A New Family of Generalization Bounds Using Samplewise Evaluated CMI

We present a new family of information-theoretic generalization bounds, ...
research
05/01/2016

Fast Rates for General Unbounded Loss Functions: from ERM to Generalized Bayes

We present new excess risk bounds for general unbounded loss functions i...
research
03/29/2021

Risk Bounds for Learning via Hilbert Coresets

We develop a formalism for constructing stochastic upper bounds on the e...
research
06/30/2011

Explicit Learning Curves for Transduction and Application to Clustering and Compression Algorithms

Inductive learning is based on inferring a general rule from a finite da...

Please sign up or login with your details

Forgot password? Click here to reset