Multi-Point Bandit Algorithms for Nonstationary Online Nonconvex Optimization

07/31/2019
by   Abhishek Roy, et al.
0

Bandit algorithms have been predominantly analyzed in the convex setting with function-value based stationary regret as the performance measure. In this paper, we propose and analyze bandit algorithms for both general and structured nonconvex problems with nonstationary (or dynamic) regret as the performance measure, in both stochastic and non-stochastic settings. First, for general nonconvex functions, we consider nonstationary versions of first-order and second-order stationary solutions as a regret measure, motivated by similar performance measures for offline nonconvex optimization. In the case of second-order stationary solution based regret, we propose and analyze online and bandit versions of the cubic regularized Newton's method. The bandit version is based on estimating the Hessian matrices in the bandit setting, based on second-order Gaussian Stein's identity. Our nonstationary regret bounds in terms of second-order stationary solutions have interesting consequences for avoiding saddle points in the bandit setting. Next, for weakly quasi convex functions and monotone weakly submodular functions we consider nonstationary regret measures in terms of function-values; such structured classes of nonconvex functions enable one to consider regret measure defined in terms of function values, similar to convex functions. For this case of function-value, and first-order stationary solution based regret measures, we provide regret bounds in both the low- and high-dimensional settings, for some scenarios.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/03/2019

Online and Bandit Algorithms for Nonstationary Stochastic Saddle-Point Optimization

Saddle-point optimization problems are an important class of optimizatio...
research
05/01/2019

Stabilized SVRG: Simple Variance Reduction for Nonconvex Optimization

Variance reduction techniques like SVRG provide simple and fast algorith...
research
07/12/2022

A Newton-CG based barrier method for finding a second-order stationary point of nonconvex conic optimization with complexity guarantees

In this paper we consider finding an approximate second-order stationary...
research
02/25/2017

Efficient Online Bandit Multiclass Learning with Õ(√(T)) Regret

We present an efficient second-order algorithm with Õ(1/η√(T)) regret fo...
research
06/12/2018

Sparse Stochastic Zeroth-Order Optimization with an Application to Bandit Structured Prediction

Stochastic zeroth-order (SZO), or gradient-free, optimization allows to ...
research
05/30/2015

Saddle-free Hessian-free Optimization

Nonconvex optimization problems such as the ones in training deep neural...
research
09/18/2022

Online Regenerative Learning

We study a type of Online Linear Programming (OLP) problem that maximize...

Please sign up or login with your details

Forgot password? Click here to reset