Minimax Regret for Bandit Convex Optimisation of Ridge Functions

06/01/2021
by   Tor Lattimore, et al.
0

We analyse adversarial bandit convex optimisation with an adversary that is restricted to playing functions of the form f(x) = g(⟨ x, θ⟩) for convex g : ℝ→ℝ and θ∈ℝ^d. We provide a short information-theoretic proof that the minimax regret is at most O(d√(n)log(diam𝒦)) where n is the number of interactions, d the dimension and diam(𝒦) is the diameter of the constraint set. Hence, this class of functions is at most logarithmically harder than the linear case.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/31/2020

Improved Regret for Zeroth-Order Adversarial Bandit Convex Optimisation

We prove that the information-theoretic upper bound on the minimax regre...
research
02/01/2023

Bandit Convex Optimisation Revisited: FTRL Achieves Õ(t^1/2) Regret

We show that a kernel estimator using multiple function evaluations can ...
research
05/28/2019

Connections Between Mirror Descent, Thompson Sampling and the Information Ratio

The information-theoretic analysis by Russo and Van Roy (2014) in combin...
research
10/24/2019

Minimax Regret of Switching-Constrained Online Convex Optimization: No Phase Transition

We study the problem of switching-constrained online convex optimization...
research
02/10/2023

A Second-Order Method for Stochastic Bandit Convex Optimisation

We introduce a simple and efficient algorithm for unconstrained zeroth-o...
research
06/17/2020

The Influence of Shape Constraints on the Thresholding Bandit Problem

We investigate the stochastic Thresholding Bandit problem (TBP) under se...
research
04/15/2020

Online Multiserver Convex Chasing and Optimization

We introduce the problem of k-chasing of convex functions, a simultaneou...

Please sign up or login with your details

Forgot password? Click here to reset