Mirror Descent Meets Fixed Share (and feels no regret)

02/15/2012
by   Nicolò Cesa-Bianchi, et al.
0

Mirror descent with an entropic regularizer is known to achieve shifting regret bounds that are logarithmic in the dimension. This is done using either a carefully designed projection or by a weight sharing technique. Via a novel unified analysis, we show that these two approaches deliver essentially equivalent bounds on a notion of regret generalizing shifting, adaptive, discounted, and other related regrets. Our analysis also captures and extends the generalized weight sharing technique of Bousquet and Warmuth, and can be refined in several ways, including improvements for small losses and adaptive tuning of parameters.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/27/2019

Adaptive Hedging under Delayed Feedback

The article is devoted to investigating the application of hedging strat...
research
06/24/2021

Improved Regret Bounds for Tracking Experts with Memory

We address the problem of sequential prediction with expert advice in a ...
research
10/09/2018

Adaptive Minimax Regret against Smooth Logarithmic Losses over High-Dimensional ℓ_1-Balls via Envelope Complexity

We develop a new theoretical framework, the envelope complexity, to anal...
research
04/26/2019

Adaptive Regret of Convex and Smooth Functions

We investigate online convex optimization in changing environments, and ...
research
05/09/2019

Non-Asymptotic Gap-Dependent Regret Bounds for Tabular MDPs

This paper establishes that optimistic algorithms attain gap-dependent a...
research
12/22/2021

A Unified Analysis Method for Online Optimization in Normed Vector Space

We present a unified analysis method that relies on the generalized cosi...
research
11/26/2015

Gains and Losses are Fundamentally Different in Regret Minimization: The Sparse Case

We demonstrate that, in the classical non-stochastic regret minimization...

Please sign up or login with your details

Forgot password? Click here to reset