Online Non-Convex Learning: Following the Perturbed Leader is Optimal

03/19/2019
by   Arun Sai Suggala, et al.
2

We study the problem of online learning with non-convex losses, where the learner has access to an offline optimization oracle. We show that the classical Follow the Perturbed Leader (FTPL) algorithm achieves optimal regret rate of O(T^-1/2) in this setting. This improves upon the previous best-known regret rate of O(T^-1/3) for FTPL. We further show that an optimistic variant of FTPL achieves better regret bounds when the sequence of losses encountered by the learner is `predictable'.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/17/2018

Learning in Non-convex Games with an Optimization Oracle

We consider adversarial online learning in a non-convex setting under th...
research
06/13/2020

Follow the Perturbed Leader: Optimism and Fast Parallel Algorithms for Smooth Minimax Games

We consider the problem of online learning and its application to solvin...
research
06/18/2022

Optimal Dynamic Regret in LQR Control

We consider the problem of nonstochastic control with a sequence of quad...
research
12/11/2019

Near-optimal Oracle-efficient Algorithms for Stationary and Non-Stationary Stochastic Linear Bandits

We investigate the design of two algorithms that enjoy not only computat...
research
02/10/2023

Oracle-Efficient Smoothed Online Learning for Piecewise Continuous Decision Making

Smoothed online learning has emerged as a popular framework to mitigate ...
research
01/06/2017

Follow the Compressed Leader: Faster Online Learning of Eigenvectors and Faster MMWU

The online problem of computing the top eigenvector is fundamental to ma...
research
02/24/2020

Prediction with Corrupted Expert Advice

We revisit the fundamental problem of prediction with expert advice, in ...

Please sign up or login with your details

Forgot password? Click here to reset