Multiscale Non-stationary Stochastic Bandits

02/13/2020
by   Qin Ding, et al.
0

Classic contextual bandit algorithms for linear models, such as LinUCB, assume that the reward distribution for an arm is modeled by a stationary linear regression. When the linear regression model is non-stationary over time, the regret of LinUCB can scale linearly with time. In this paper, we propose a novel multiscale changepoint detection method for the non-stationary linear bandit problems, called Multiscale-LinUCB, which actively adapts to the changing environment. We also provide theoretical analysis of regret bound for Multiscale-LinUCB algorithm. Experimental results show that our proposed Multiscale-LinUCB algorithm outperforms other state-of-the-art algorithms in non-stationary contextual environments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/19/2019

Weighted Linear Bandits for Non-Stationary Environments

We consider a stochastic linear bandit model in which the available acti...
research
08/05/2017

Efficient Contextual Bandits in Non-stationary Worlds

Most contextual bandit algorithms minimize regret to the best fixed poli...
research
05/18/2023

Discounted Thompson Sampling for Non-Stationary Bandit Problems

Non-stationary multi-armed bandit (NS-MAB) problems have recently receiv...
research
07/09/2020

Recurrent Neural-Linear Posterior Sampling for Non-Stationary Contextual Bandits

An agent in a non-stationary contextual bandit problem should balance be...
research
04/28/2020

A Linear Bandit for Seasonal Environments

Contextual bandit algorithms are extremely popular and widely used in re...
research
09/19/2022

A Multi-Layer Regression based Predicable Function Fitting Network

Function plays an important role in mathematics and many science branche...
research
07/07/2023

BOF-UCB: A Bayesian-Optimistic Frequentist Algorithm for Non-Stationary Contextual Bandits

We propose a novel Bayesian-Optimistic Frequentist Upper Confidence Boun...

Please sign up or login with your details

Forgot password? Click here to reset