Information-Theoretic Generalization Bounds for SGLD via Data-Dependent Estimates

11/06/2019
by   Jeffrey Negrea, et al.
15

In this work, we improve upon the stepwise analysis of noisy iterative learning algorithms initiated by Pensia, Jog, and Loh (2018) and recently extended by Bu, Zou, and Veeravalli (2019). Our main contributions are significantly improved mutual information bounds for Stochastic Gradient Langevin Dynamics via data-dependent estimates. Our approach is based on the variational characterization of mutual information and the use of data-dependent priors that forecast the mini-batch gradient based on a subset of the training samples. Our approach is broadly applicable within the information-theoretic framework of Russo and Zou (2015) and Xu and Raginsky (2017). Our bound can be tied to a measure of flatness of the empirical risk surface. As compared with other bounds that depend on the squared norms of gradients, empirical investigations show that the terms in our bounds are orders of magnitude smaller.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/09/2023

Information Theoretic Lower Bounds for Information Theoretic Upper Bounds

We examine the relationship between the mutual information between the o...
research
04/27/2020

Sharpened Generalization Bounds based on Conditional Mutual Information and an Application to Noisy, Iterative Algorithms

The information-theoretic framework of Russo and J. Zou (2016) and Xu an...
research
03/09/2019

Strengthened Information-theoretic Bounds on the Generalization Error

The following problem is considered: given a joint distribution P_XY and...
research
07/04/2023

Generalization Guarantees via Algorithm-dependent Rademacher Complexity

Algorithm- and data-dependent generalization bounds are required to expl...
research
05/02/2023

Understanding the Generalization Ability of Deep Learning Algorithms: A Kernelized Renyi's Entropy Perspective

Recently, information theoretic analysis has become a popular framework ...
research
04/20/2018

Inter-Annotator Agreement Networks

This work develops a simple information theoretic framework that capture...
research
04/04/2023

Algorithm-Dependent Bounds for Representation Learning of Multi-Source Domain Adaptation

We use information-theoretic tools to derive a novel analysis of Multi-s...

Please sign up or login with your details

Forgot password? Click here to reset