Instance-Dependent Generalization Bounds via Optimal Transport

11/02/2022
by   Songyan Hou, et al.
0

Existing generalization bounds fail to explain crucial factors that drive generalization of modern neural networks. Since such bounds often hold uniformly over all parameters, they suffer from over-parametrization, and fail to account for the strong inductive bias of initialization and stochastic gradient descent. As an alternative, we propose a novel optimal transport interpretation of the generalization problem. This allows us to derive instance-dependent generalization bounds that depend on the local Lipschitz regularity of the earned prediction function in the data space. Therefore, our bounds are agnostic to the parametrization of the model and work well when the number of training samples is much smaller than the number of parameters. With small modifications, our approach yields accelerated rates for data on low-dimensional manifolds, and guarantees under distribution shifts. We empirically analyze our generalization bounds for neural networks, showing that the bound values are meaningful and capture the effect of popular regularization methods during training.

READ FULL TEXT
research
06/07/2021

Measuring Generalization with Optimal Transport

Understanding the generalization of deep neural networks is one of the m...
research
09/17/2020

A Principle of Least Action for the Training of Neural Networks

Neural networks have been achieving high generalization performance on m...
research
02/12/2018

Computational Optimal Transport: Complexity by Accelerated Gradient Descent Is Better Than by Sinkhorn's Algorithm

We analyze two algorithms for approximating the general optimal transpor...
research
02/16/2022

GAN Estimation of Lipschitz Optimal Transport Maps

This paper introduces the first statistically consistent estimator of th...
research
06/17/2022

How You Start Matters for Generalization

Characterizing the remarkable generalization properties of over-paramete...
research
12/08/2021

Generalization Error Bounds for Iterative Recovery Algorithms Unfolded as Neural Networks

Motivated by the learned iterative soft thresholding algorithm (LISTA), ...
research
09/29/2022

Improving Generative Flow Networks with Path Regularization

Generative Flow Networks (GFlowNets) are recently proposed models for le...

Please sign up or login with your details

Forgot password? Click here to reset