Online Boosting with Bandit Feedback

07/23/2020
by   Nataly Brukhim, et al.
0

We consider the problem of online boosting for regression tasks, when only limited information is available to the learner. We give an efficient regret minimization method that has two implications: an online boosting algorithm with noisy multi-point bandit feedback, and a new projection-free online convex optimization algorithm with stochastic gradient, that improves state-of-the-art guarantees in terms of efficiency.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/18/2021

Boosting for Online Convex Optimization

We consider the decision-making framework of online convex optimization ...
research
09/12/2017

Setpoint Tracking with Partially Observed Loads

We use online convex optimization (OCO) for setpoint tracking with uncer...
research
10/11/2018

Online Multiclass Boosting with Bandit Feedback

We present online boosting algorithms for multiclass classification with...
research
03/02/2020

Online Agnostic Boosting via Regret Minimization

Boosting is a widely used machine learning approach based on the idea of...
research
05/15/2022

Online Nonsubmodular Minimization with Delayed Costs: From Full Information to Bandit Feedback

Motivated by applications to online learning in sparse estimation and Ba...
research
10/01/2018

Risk-Averse Stochastic Convex Bandit

Motivated by applications in clinical trials and finance, we study the p...
research
10/14/2020

Boosting One-Point Derivative-Free Online Optimization via Residual Feedback

Zeroth-order optimization (ZO) typically relies on two-point feedback to...

Please sign up or login with your details

Forgot password? Click here to reset