Polynomial Cost of Adaptation for X -Armed Bandits

05/24/2019
by   Hédi Hadiji, et al.
0

In the context of stochastic continuum-armed bandits, we present an algorithm that adapts to the unknown smoothness of the objective function. We exhibit and compute a polynomial cost of adaptation to the Hölder regularity for regret minimization. To do this, we first reconsider the recent lower bound of Locatelli and Carpentier [20], and define and characterize admissible rate functions. Our new algorithm matches any of these minimal rate functions. We provide a finite-time analysis and a thorough discussion about asymptotic optimality.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/29/2016

Regret Analysis of the Anytime Optimally Confident UCB Algorithm

I introduce and analyse an anytime version of the Optimally Confident UC...
research
10/15/2020

Continuum-Armed Bandits: A Function Space Perspective

Continuum-armed bandits (a.k.a., black-box or 0^th-order optimization) i...
research
12/11/2020

Smooth Bandit Optimization: Generalization to Hölder Space

We consider bandit optimization of a smooth reward function, where the g...
research
04/26/2023

Adaptation to Misspecified Kernel Regularity in Kernelised Bandits

In continuum-armed bandit problems where the underlying function resides...
research
10/14/2016

The End of Optimism? An Asymptotic Analysis of Finite-Armed Linear Bandits

Stochastic linear bandits are a natural and simple generalisation of fin...
research
02/11/2013

Adaptive-treed bandits

We describe a novel algorithm for noisy global optimisation and continuu...
research
03/27/2017

A Scale Free Algorithm for Stochastic Bandits with Bounded Kurtosis

Existing strategies for finite-armed stochastic bandits mostly depend on...

Please sign up or login with your details

Forgot password? Click here to reset