The Many Faces of Exponential Weights in Online Learning

02/21/2018
by   Dirk van der Hoeven, et al.
0

A standard introduction to online learning might place Online Gradient Descent at its center and then proceed to develop generalizations and extensions like Online Mirror Descent and second-order methods. Here we explore the alternative approach of putting exponential weights (EW) first. We show that many standard methods and their regret bounds then follow as a special case by plugging in suitable surrogate losses and playing the EW posterior mean. For instance, we easily recover Online Gradient Descent by using EW with a Gaussian prior on linearized losses, and, more generally, all instances of Online Mirror Descent based on regular Bregman divergences also correspond to EW with a prior that depends on the mirror map. Furthermore, appropriate quadratic surrogate losses naturally give rise to Online Gradient Descent for strongly convex losses and to Online Newton Step. We further interpret several recent adaptive methods (iProd, Squint, and a variation of Coin Betting for experts) as a series of closely related reductions to exp-concave surrogate losses that are then handled by Exponential Weights. Finally, a benefit of our EW interpretation is that it opens up the possibility of sampling from the EW posterior distribution instead of playing the mean. As already observed by Bubeck and Eldan, this recovers the best-known rate in Online Bandit Linear Optimization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/27/2018

Online learning with kernel losses

We present a generalization of the adversarial linear bandits framework,...
research
11/25/2018

Online Newton Step Algorithm with Estimated Gradient

Online learning with limited information feedback (bandit) tries to solv...
research
02/17/2020

Last iterate convergence in no-regret learning: constrained min-max optimization for convex-concave landscapes

In a recent series of papers it has been established that variants of Gr...
research
02/17/2018

Black-Box Reductions for Parameter-free Online Learning in Banach Spaces

We introduce several new black-box reductions that significantly improve...
research
04/23/2021

Optimal Dynamic Regret in Exp-Concave Online Learning

We consider the problem of the Zinkevich (2003)-style dynamic regret min...
research
07/03/2018

On the Computational Power of Online Gradient Descent

We prove that the evolution of weight vectors in online gradient descent...
research
04/04/2014

Optimal learning with Bernstein Online Aggregation

We introduce a new recursive aggregation procedure called Bernstein Onli...

Please sign up or login with your details

Forgot password? Click here to reset