Learning to Optimize under Non-Stationarity

10/06/2018
by   Wang Chi Cheung, et al.
0

We introduce algorithms that achieve state-of-the-art dynamic regret bounds for non-stationary linear stochastic bandits setting. It captures natural applications such as dynamic pricing and ads allocation in a changing environment. We show how the difficulty posed by the (possibly adversarial) non-stationarity can be overcome by a novel marriage between stochastic and adversarial bandits learning algorithms. Defining d,B_T, and T as the problem dimension, the variation budget, and the total time horizon, respectively, our main contributions are the tuned Sliding Window UCB (SW-UCB) algorithm with optimal O(d^2/3(B_T+1)^1/3T^2/3) dynamic regret, and the tuning free bandits-over-bandits (BOB) framework built on top of the SW-UCB algorithm with best O(d^2/3(B_T+1)^1/4T^3/4) dynamic regret.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/04/2019

Hedging the Drift: Learning to Optimize under Non-Stationarity

We introduce general data-driven decision-making algorithms that achieve...
research
03/05/2023

Revisiting Weighted Strategy for Non-stationary Parametric Bandits

Non-stationary parametric bandits have attracted much attention recently...
research
03/09/2021

Non-stationary Linear Bandits Revisited

In this note, we revisit non-stationary linear bandits, a variant of sto...
research
02/28/2023

Approximately Stationary Bandits with Knapsacks

Bandits with Knapsacks (BwK), the generalization of the Multi-Armed Band...
research
02/10/2021

Non-stationary Reinforcement Learning without Prior Knowledge: An Optimal Black-box Approach

We propose a black-box reduction that turns a certain reinforcement lear...
research
11/28/2018

Adversarial Bandits with Knapsacks

We consider Bandits with Knapsacks (henceforth, BwK), a general model fo...
research
10/24/2021

Online estimation and control with optimal pathlength regret

A natural goal when designing online learning algorithms for non-station...

Please sign up or login with your details

Forgot password? Click here to reset