Optimality of the Subgradient Algorithm in the Stochastic Setting

09/10/2019
by   Daron Anderson, et al.
0

Recently Jaouad Mourtada and Stéphane Gaïffas showed the anytime hedge algorithm has pseudo-regret O( (d) / Δ) if the cost vectors are generated by an i.i.d sequence in the cube [0,1]^d. Here d is the dimension and Δ the suboptimality gap. This is remarkable because the Hedge algorithm was designed for the antagonistic setting. We prove a similar result for the anytime subgradient algorithm on the simplex. Given i.i.d cost vectors in the unit ball our pseudo-regret bound is O(1/Δ) and does not depend on the dimension of the problem.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/03/2020

Universal Algorithms: Beyond the Simplex

The bulk of universal algorithms in the online convex optimisation liter...
research
03/16/2023

On the Interplay Between Misspecification and Sub-optimality Gap in Linear Contextual Bandits

We study linear contextual bandits in the misspecified setting, where th...
research
05/06/2023

An improved regret analysis for UCB-N and TS-N

In the setting of stochastic online learning with undirected feedback gr...
research
03/10/2021

Linear Bandits on Uniformly Convex Sets

Linear bandit algorithms yield 𝒪̃(n√(T)) pseudo-regret bounds on compact...
research
07/24/2020

Exploiting the Surrogate Gap in Online Multiclass Classification

We present Gaptron, a randomized first-order algorithm for online multic...
research
06/15/2022

Corruption-Robust Contextual Search through Density Updates

We study the problem of contextual search in the adversarial noise model...
research
11/02/2020

Stochastic Linear Bandits with Protected Subspace

We study a variant of the stochastic linear bandit problem wherein we op...

Please sign up or login with your details

Forgot password? Click here to reset