High probability generalization bounds for uniformly stable algorithms with nearly optimal rate

02/27/2019
by   Vitaly Feldman, et al.
2

Algorithmic stability is a classical approach to understanding and analysis of the generalization error of learning algorithms. A notable weakness of most stability-based generalization bounds is that they hold only in expectation. Generalization with high probability has been established in a landmark paper of Bousquet and Elisseeff (2002) albeit at the expense of an additional √(n) factor in the bound. Specifically, their bound on the estimation error of any γ-uniformly stable learning algorithm on n samples and range in [0,1] is O(γ√(n (1/δ)) + √((1/δ)/n)) with probability ≥ 1-δ. The √(n) overhead makes the bound vacuous in the common settings where γ≥ 1/√(n). A stronger bound was recently proved by the authors (Feldman and Vondrak, 2018) that reduces the overhead to at most O(n^1/4). Still, both of these results give optimal generalization bounds only when γ = O(1/n). We prove a nearly tight bound of O(γ(n)(n/δ) + √((1/δ)/n)) on the estimation error of any γ-uniformly stable algorithm. It implies that algorithms that are uniformly stable with γ = O(1/√(n)) have essentially the same estimation error as algorithms that output a fixed function. Our result leads to the first high-probability generalization bounds for multi-pass stochastic gradient descent and regularized ERM for stochastic convex problems with nearly optimal rate --- resolving open problems in prior work. Our proof technique is new and we introduce several analysis tools that might find additional applications.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/24/2018

Generalization Bounds for Uniformly Stable Algorithms

Uniform stability of a learning algorithm is a classical notion of algor...
research
03/22/2021

Stability and Deviation Optimal Risk Bounds with Convergence Rate O(1/n)

The sharpest known high probability generalization bounds for uniformly ...
research
10/17/2019

Sharper bounds for uniformly stable algorithms

The generalization bounds for stable algorithms is a classical question ...
research
02/20/2023

On the Stability and Generalization of Triplet Learning

Triplet learning, i.e. learning from triplet data, has attracted much at...
research
05/27/2022

Generalization Bounds for Gradient Methods via Discrete and Continuous Prior

Proving algorithm-dependent generalization error bounds for gradient-typ...
research
12/19/2022

The One-Inclusion Graph Algorithm is not Always Optimal

The one-inclusion graph algorithm of Haussler, Littlestone, and Warmuth ...
research
05/31/2021

Fine-grained Generalization Analysis of Structured Output Prediction

In machine learning we often encounter structured output prediction prob...

Please sign up or login with your details

Forgot password? Click here to reset