Thompson Sampling for High-Dimensional Sparse Linear Contextual Bandits

11/11/2022
by   Sunrit Chakraborty, et al.
0

We consider the stochastic linear contextual bandit problem with high-dimensional features. We analyze the Thompson sampling (TS) algorithm, using special classes of sparsity-inducing priors (e.g. spike-and-slab) to model the unknown parameter, and provide a nearly optimal upper bound on the expected cumulative regret. To the best of our knowledge, this is the first work that provides theoretical guarantees of Thompson sampling in high dimensional and sparse contextual bandits. For faster computation, we use spike-and-slab prior to model the unknown parameter and variational inference instead of MCMC to approximate the posterior distribution. Extensive simulations demonstrate improved performance of our proposed algorithm over existing ones.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/27/2020

Dynamic Batch Learning in High-Dimensional Sparse Linear Contextual Bandits

We study the problem of dynamic batch learning in high-dimensional spars...
research
03/20/2019

Contextual Bandits with Random Projection

Contextual bandits with linear payoffs, which are also known as linear b...
research
06/19/2023

High-dimensional Contextual Bandit Problem without Sparsity

In this research, we investigate the high-dimensional linear contextual ...
research
07/19/2023

VITS : Variational Inference Thomson Sampling for contextual bandits

In this paper, we introduce and analyze a variant of the Thompson sampli...
research
07/16/2020

Sparsity-Agnostic Lasso Bandit

We consider a stochastic contextual bandit problem where the dimension d...
research
06/22/2022

Langevin Monte Carlo for Contextual Bandits

We study the efficiency of Thompson sampling for contextual bandits. Exi...
research
06/10/2015

On the Prior Sensitivity of Thompson Sampling

The empirically successful Thompson Sampling algorithm for stochastic ba...

Please sign up or login with your details

Forgot password? Click here to reset