The Interplay Between Stability and Regret in Online Learning

11/26/2012
by   Ankan Saha, et al.
0

This paper considers the stability of online learning algorithms and its implications for learnability (bounded regret). We introduce a novel quantity called forward regret that intuitively measures how good an online learning algorithm is if it is allowed a one-step look-ahead into the future. We show that given stability, bounded forward regret is equivalent to bounded regret. We also show that the existence of an algorithm with bounded regret implies the existence of a stable algorithm with bounded regret and bounded forward regret. The equivalence results apply to general, possibly non-convex problems. To the best of our knowledge, our analysis provides the first general connection between stability and regret in the online setting that is not restricted to a particular class of algorithms. Our stability-regret connection provides a simple recipe for analyzing regret incurred by any online learning algorithm. Using our framework, we analyze several existing online learning algorithms as well as the "approximate" versions of algorithms like RDA that solve an optimization problem at each iteration. Our proofs are simpler than existing analysis for the respective algorithms, show a clear trade-off between stability and forward regret, and provide tighter regret bounds in some cases. Furthermore, using our recipe, we analyze "approximate" versions of several algorithms such as follow-the-regularized-leader (FTRL) that requires solving an optimization problem at each step.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/31/2023

Online-to-PAC Conversions: Generalization Bounds via Regret Analysis

We present a new framework for deriving bounds on the generalization bou...
research
05/28/2022

History-Restricted Online Learning

We introduce the concept of history-restricted no-regret online learning...
research
02/12/2013

Competing With Strategies

We study the problem of online learning with a notion of regret defined ...
research
05/31/2022

AdaTask: Adaptive Multitask Online Learning

We introduce and analyze AdaTask, a multitask online learning algorithm ...
research
08/16/2011

Stability Conditions for Online Learnability

Stability is a general notion that quantifies the sensitivity of a learn...
research
09/08/2017

A Modular Analysis of Adaptive (Non-)Convex Optimization: Optimism, Composite Objectives, and Variational Bounds

Recently, much work has been done on extending the scope of online learn...
research
09/25/2018

Fully Implicit Online Learning

Regularized online learning is widely used in machine learning. In this ...

Please sign up or login with your details

Forgot password? Click here to reset