Comparator-adaptive Convex Bandits

07/16/2020
by   Dirk van der Hoeven, et al.
8

We study bandit convex optimization methods that adapt to the norm of the comparator, a topic that has only been studied before for its full-information counterpart. Specifically, we develop convex bandit algorithms with regret bounds that are small whenever the norm of the comparator is small. We first use techniques from the full-information setting to develop comparator-adaptive algorithms for linear bandits. Then, we extend the ideas to convex bandits with Lipschitz or smooth loss functions, using a new single-point gradient estimator and carefully designed surrogate losses.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/31/2015

An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback

We consider the closely related problems of bandit convex optimization w...
research
03/10/2021

Linear Bandits on Uniformly Convex Sets

Linear bandit algorithms yield 𝒪̃(n√(T)) pseudo-regret bounds on compact...
research
02/12/2022

Adaptive Bandit Convex Optimization with Heterogeneous Curvature

We consider the problem of adversarial bandit convex optimization, that ...
research
01/29/2019

Improved Path-length Regret Bounds for Bandits

We study adaptive regret bounds in terms of the variation of the losses ...
research
06/28/2018

Contextual bandits with surrogate losses: Margin bounds and efficient algorithms

We introduce a new family of margin-based regret guarantees for adversar...
research
06/01/2020

Locally Differentially Private (Contextual) Bandits Learning

We study locally differentially private (LDP) bandits learning in this p...
research
08/29/2023

Exploiting Problem Geometry in Safe Linear Bandits

The safe linear bandit problem is a version of the classic linear bandit...

Please sign up or login with your details

Forgot password? Click here to reset