Bandit Convex Optimisation Revisited: FTRL Achieves Õ(t^1/2) Regret

02/01/2023
by   David Young, et al.
0

We show that a kernel estimator using multiple function evaluations can be easily converted into a sampling-based bandit estimator with expectation equal to the original kernel estimate. Plugging such a bandit estimator into the standard FTRL algorithm yields a bandit convex optimisation algorithm that achieves Õ(t^1/2) regret against adversarial time-varying convex loss functions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/31/2020

Improved Regret for Zeroth-Order Adversarial Bandit Convex Optimisation

We prove that the information-theoretic upper bound on the minimax regre...
research
06/01/2021

Minimax Regret for Bandit Convex Optimisation of Ridge Functions

We analyse adversarial bandit convex optimisation with an adversary that...
research
02/10/2023

A Second-Order Method for Stochastic Bandit Convex Optimisation

We introduce a simple and efficient algorithm for unconstrained zeroth-o...
research
10/27/2022

Lifelong Bandit Optimization: No Prior and No Regret

In practical applications, machine learning algorithms are often repeate...
research
01/28/2020

Bandit optimisation of functions in the Matérn kernel RKHS

We consider the problem of optimising functions in the Reproducing kerne...
research
07/11/2016

Kernel-based methods for bandit convex optimization

We consider the adversarial convex bandit problem and we build the first...
research
05/18/2018

Projection-Free Bandit Convex Optimization

In this paper, we propose the first computationally efficient projection...

Please sign up or login with your details

Forgot password? Click here to reset