MNL-Bandit in non-stationary environments

03/04/2023
by   Ayoub Foussoul, et al.
0

In this paper, we study the MNL-Bandit problem in a non-stationary environment and present an algorithm with worst-case dynamic regret of Õ( min{√(NTL) , N^1/3(Δ_∞^K)^1/3 T^2/3 + √(NT)}). Here N is the number of arms, L is the number of switches and Δ_∞^K is a variation measure of the unknown parameters. We also show that our algorithm is near-optimal (up to logarithmic factors). Our algorithm builds upon the epoch-based algorithm for stationary MNL-Bandit in Agrawal et al. 2016. However, non-stationarity poses several challenges and we introduce new techniques and ideas to address these. In particular, we give a tight characterization for the bias introduced in the estimators due to non stationarity and derive new concentration bounds.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/25/2022

Non-stationary Bandits and Meta-Learning with a Small Set of Optimal Arms

We study a sequential decision problem where the learner faces a sequenc...
research
02/23/2023

A Definition of Non-Stationary Bandits

The subject of non-stationary bandit learning has attracted much recent ...
research
02/01/2021

Generalized non-stationary bandits

In this paper, we study a non-stationary stochastic bandit problem, whic...
research
08/02/2023

Maximizing Success Rate of Payment Routing using Non-stationary Bandits

This paper discusses the system architecture design and deployment of no...
research
03/09/2021

Regret Bounds for Generalized Linear Bandits under Parameter Drift

Generalized Linear Bandits (GLBs) are powerful extensions to the Linear ...
research
02/10/2020

Combinatorial Semi-Bandit in the Non-Stationary Environment

In this paper, we investigate the non-stationary combinatorial semi-band...
research
11/06/2021

Dynamic Regret Minimization for Control of Non-stationary Linear Dynamical Systems

We consider the problem of controlling a Linear Quadratic Regulator (LQR...

Please sign up or login with your details

Forgot password? Click here to reset