On Abruptly-Changing and Slowly-Varying Multiarmed Bandit Problems

02/23/2018
by   Lai Wei, et al.
0

We study the non-stationary stochastic multiarmed bandit (MAB) problem and propose two generic algorithms, namely, the limited memory deterministic sequencing of exploration and exploitation (LM-DSEE) and the Sliding-Window Upper Confidence Bound# (SW-UCB#). We rigorously analyze these algorithms in abruptly-changing and slowly-varying environments and characterize their performance. We show that the expected cumulative regret for these algorithms under either of the environments is upper bounded by sublinear functions of time, i.e., the time average of the regret asymptotically converges to zero. We complement our analytic results with numerical illustrations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/12/2018

On Distributed Multi-player Multiarmed Bandit Problems in Abruptly Changing Environment

We study the multi-player stochastic multiarmed bandit (MAB) problem in ...
research
02/26/2017

Kiefer Wolfowitz Algorithm is Asymptotically Optimal for a Class of Non-Stationary Bandit Problems

We consider the problem of designing an allocation rule or an "online le...
research
03/04/2019

Hedging the Drift: Learning to Optimize under Non-Stationarity

We introduce general data-driven decision-making algorithms that achieve...
research
06/07/2019

Reinforcement Learning under Drift

We propose algorithms with state-of-the-art dynamic regret bounds for un...
research
05/18/2023

Discounted Thompson Sampling for Non-Stationary Bandit Problems

Non-stationary multi-armed bandit (NS-MAB) problems have recently receiv...
research
05/25/2018

A Sliding-Window Algorithm for Markov Decision Processes with Arbitrarily Changing Rewards and Transitions

We consider reinforcement learning in changing Markov Decision Processes...
research
02/25/2016

Thompson Sampling is Asymptotically Optimal in General Environments

We discuss a variant of Thompson sampling for nonparametric reinforcemen...

Please sign up or login with your details

Forgot password? Click here to reset