Algorithmic Chaining and the Role of Partial Feedback in Online Nonparametric Learning

02/27/2017
by   Nicolò Cesa-Bianchi, et al.
0

We investigate contextual online learning with nonparametric (Lipschitz) comparison classes under different assumptions on losses and feedback information. For full information feedback and Lipschitz losses, we design the first explicit algorithm achieving the minimax regret rate (up to log factors). In a partial feedback model motivated by second-price auctions, we obtain algorithms for Lipschitz and semi-Lipschitz losses with regret bounds improving on the known bounds for standard bandit feedback. Our analysis combines novel results for contextual second-price auctions with a novel algorithmic approach based on chaining. When the context space is Euclidean, our chaining approach is efficient and delivers an even better regret bound.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2023

Unconstrained Online Learning with Unbounded Losses

Algorithms for online learning typically require one or more boundedness...
research
03/09/2023

Improved Regret Bounds for Online Kernel Selection under Bandit Feedback

In this paper, we improve the regret bound for online kernel selection u...
research
10/19/2019

On Adaptivity in Information-constrained Online Learning

We study how to adapt to smoothly-varying (`easy') environments in well-...
research
06/28/2018

Contextual bandits with surrogate losses: Margin bounds and efficient algorithms

We introduce a new family of margin-based regret guarantees for adversar...
research
02/05/2019

Contextual Bandits with Continuous Actions: Smoothing, Zooming, and Adapting

We study contextual bandit learning with an abstract policy class and co...
research
02/17/2023

Graph Feedback via Reduction to Regression

When feedback is partial, leveraging all available information is critic...
research
09/03/2022

Sharp bounds on the price of bandit feedback for several models of mistake-bounded online learning

We determine sharp bounds on the price of bandit feedback for several va...

Please sign up or login with your details

Forgot password? Click here to reset