Exactly Tight Information-Theoretic Generalization Error Bound for the Quadratic Gaussian Problem

05/01/2023
by   Ruida Zhou, et al.
0

We provide a new information-theoretic generalization error bound that is exactly tight (i.e., matching even the constant) for the canonical quadratic Gaussian mean estimation problem. Despite considerable existing efforts in deriving information-theoretic generalization error bounds, applying them to this simple setting where sample average is used as the estimate of the mean value of Gaussian data has not yielded satisfying results. In fact, most existing bounds are order-wise loose in this setting, which has raised concerns about the fundamental capability of information-theoretic bounds in reasoning the generalization behavior for machine learning. The proposed new bound adopts the individual-sample-based approach proposed by Bu et al., but also has several key new ingredients. Firstly, instead of applying the change of measure inequality on the loss function, we apply it to the generalization error function itself; secondly, the bound is derived in a conditional manner; lastly, a reference distribution, which bears a certain similarity to the prior distribution in the Bayesian setting, is introduced. The combination of these components produces a general KL-divergence-based generalization error bound. We further show that although the conditional bounding and the reference distribution can make the bound exactly tight, removing them does not significantly degrade the bound, which leads to a mutual-information-based bound that is also asymptotically tight in this setting.

READ FULL TEXT
research
12/17/2020

Individually Conditional Individual Mutual Information Bound on Generalization Error

We propose a new information-theoretic bound on generalization error bas...
research
05/18/2023

A unified framework for information-theoretic generalization bounds

This paper presents a general methodology for deriving information-theor...
research
01/28/2022

Stochastic Chaining and Strengthened Information-Theoretic Generalization Bounds

We propose a new approach to apply the chaining technique in conjunction...
research
06/23/2023

Precise Asymptotic Generalization for Multiclass Classification with Overparameterized Linear Models

We study the asymptotic generalization of an overparameterized linear mo...
research
10/12/2022

A New Family of Generalization Bounds Using Samplewise Evaluated CMI

We present a new family of information-theoretic generalization bounds, ...
research
10/26/2013

Efficient Information Theoretic Clustering on Discrete Lattices

We consider the problem of clustering data that reside on discrete, low ...
research
02/19/2021

Information-Theoretic Bounds for Integral Estimation

In this paper, we consider a zero-order stochastic oracle model of estim...

Please sign up or login with your details

Forgot password? Click here to reset