Tighter Expected Generalization Error Bounds via Convexity of Information Measures

02/24/2022
by   Gholamali Aminian, et al.
8

Generalization error bounds are essential to understanding machine learning algorithms. This paper presents novel expected generalization error upper bounds based on the average joint distribution between the output hypothesis and each input training sample. Multiple generalization error upper bounds based on different information measures are provided, including Wasserstein distance, total variation distance, KL divergence, and Jensen-Shannon divergence. Due to the convexity of the information measures, the proposed bounds in terms of Wasserstein distance and total variation distance are shown to be tighter than their counterparts based on individual samples in the literature. An example is provided to demonstrate the tightness of the proposed generalization error bounds.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/01/2022

Tight bounds for augmented KL divergence in terms of augmented total variation distance

We provide optimal variational upper and lower bounds for the augmented ...
research
11/08/2018

An Optimal Transport View on Generalization

We derive upper bounds on the generalization error of learning algorithm...
research
01/22/2021

Tighter expected generalization error bounds via Wasserstein distance

In this work, we introduce several expected generalization error bounds ...
research
05/30/2019

Convergence of Smoothed Empirical Measures with Applications to Entropy Estimation

This paper studies convergence of empirical measures smoothed by a Gauss...
research
02/15/2019

Asymptotic Finite Sample Information Losses in Neural Classifiers

This paper considers the subject of information losses arising from fini...
research
01/20/2014

Generalized Bhattacharyya and Chernoff upper bounds on Bayes error using quasi-arithmetic means

Bayesian classification labels observations based on given prior informa...
research
12/11/2020

Randomised Wasserstein Barycenter Computation: Resampling with Statistical Guarantees

We propose a hybrid resampling method to approximate finitely supported ...

Please sign up or login with your details

Forgot password? Click here to reset