A Tight Excess Risk Bound via a Unified PAC-Bayesian-Rademacher-Shtarkov-MDL Complexity

10/21/2017
by   Peter D. Grünwald, et al.
0

We present a novel notion of complexity that interpolates between and generalizes some classic existing complexity notions in learning theory: for estimators like empirical risk minimization (ERM) with arbitrary bounded losses, it is upper bounded in terms of data-independent Rademacher complexity; for generalized Bayesian estimators, it is upper bounded by the data-dependent information complexity (also known as stochastic or PAC-Bayesian, KL(posteriorprior) complexity. For (penalized) ERM, the new complexity reduces to (generalized) normalized maximum likelihood (NML) complexity, i.e. a minimax log-loss individual-sequence regret. Our first main result bounds excess risk in terms of the new complexity. Our second main result links the new complexity via Rademacher complexity to L_2(P) entropy, thereby generalizing earlier results of Opper, Haussler, Lugosi, and Cesa-Bianchi who did the log-loss case with L_∞. Together, these results recover optimal bounds for VC- and large (polynomial entropy) classes, replacing localized Rademacher complexity by a simpler analysis which almost completely separates the two aspects that determine the achievable rates: 'easiness' (Bernstein) conditions and model complexity.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/01/2016

Fast Rates for General Unbounded Loss Functions: from ERM to Generalized Bayes

We present new excess risk bounds for general unbounded loss functions i...
research
01/18/2022

On the minimax rate of the Gaussian sequence model under bounded convex constraints

We determine the exact minimax rate of a Gaussian sequence model under b...
research
05/27/2016

PAC-Bayesian Theory Meets Bayesian Inference

We exhibit a strong link between frequentist PAC-Bayesian risk bounds an...
research
01/14/2021

New bounds for k-means and information k-means

In this paper, we derive a new dimension-free non-asymptotic upper bound...
research
05/19/2015

Risk and Regret of Hierarchical Bayesian Learners

Common statistical practice has shown that the full power of Bayesian me...
research
01/19/2007

Algorithmic Complexity Bounds on Future Prediction Errors

We bound the future loss when predicting any (computably) stochastic seq...
research
01/29/2021

Sequential prediction under log-loss and misspecification

We consider the question of sequential prediction under the log-loss in ...

Please sign up or login with your details

Forgot password? Click here to reset