Between Stochastic and Adversarial Online Convex Optimization: Improved Regret Bounds via Smoothness

02/15/2022
by   Sarah Sachs, et al.
0

Stochastic and adversarial data are two widely studied settings in online learning. But many optimization tasks are neither i.i.d. nor fully adversarial, which makes it of fundamental interest to get a better theoretical understanding of the world between these extremes. In this work we establish novel regret bounds for online convex optimization in a setting that interpolates between stochastic i.i.d. and fully adversarial losses. By exploiting smoothness of the expected losses, these bounds replace a dependence on the maximum gradient length by the variance of the gradients, which was previously known only for linear losses. In addition, they weaken the i.i.d. assumption by allowing adversarially poisoned rounds or shifts in the data distribution. To accomplish this goal, we introduce two key quantities associated with the loss sequence, that we call the cumulative stochastic variance and the adversarial variation. Our upper bounds are attained by instances of optimistic follow the regularized leader, and we design adaptive learning rates that automatically adapt to the cumulative stochastic variance and adversarial variation. In the fully i.i.d. case, our bounds match the rates one would expect from results in stochastic acceleration, and in the fully adversarial case they gracefully deteriorate to match the minimax regret. We further provide lower bounds showing that our regret upper bounds are tight for all intermediate regimes for the cumulative stochastic variance and the adversarial variation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/06/2023

Accelerated Rates between Stochastic and Adversarial Online Convex Optimization

Stochastic and adversarial data are two widely studied settings in onlin...
research
06/23/2021

Best-Case Lower Bounds in Online Learning

Much of the work in online learning focuses on the study of sublinear up...
research
02/09/2023

Optimistic Online Mirror Descent for Bridging Stochastic and Adversarial Online Convex Optimization

Stochastically Extended Adversarial (SEA) model is introduced by Sachs e...
research
02/24/2023

Best-of-Three-Worlds Linear Bandit Algorithm with Variance-Adaptive Regret Bounds

This paper proposes a linear bandit algorithm that is adaptive to enviro...
research
05/23/2023

Data-Dependent Bounds for Online Portfolio Selection Without Lipschitzness and Smoothness

This work introduces the first small-loss and gradual-variation regret b...
research
12/03/2020

Online learning with dynamics: A minimax perspective

We study the problem of online learning with dynamics, where a learner i...
research
07/20/2021

Best-of-All-Worlds Bounds for Online Learning with Feedback Graphs

We study the online learning with feedback graphs framework introduced b...

Please sign up or login with your details

Forgot password? Click here to reset