PG-TS: Improved Thompson Sampling for Logistic Contextual Bandits

05/18/2018
by   Bianca Dumitrascu, et al.
0

We address the problem of regret minimization in logistic contextual bandits, where a learner decides among sequential actions or arms given their respective contexts to maximize binary rewards. Using a fast inference procedure with Polya-Gamma distributed augmentation variables, we propose an improved version of Thompson Sampling, a Bayesian formulation of contextual bandits with near-optimal performance. Our approach, Polya-Gamma augmented Thompson Sampling (PG-TS), achieves state-of-the-art performance on simulated and real data. PG-TS explores the action space efficiently and exploits high-reward arms, quickly converging to solutions of low regret. Its explicit estimation of the posterior distribution of the context feature covariance leads to substantial empirical gains over approximate approaches. PG-TS is the first approach to demonstrate the benefits of Polya-Gamma augmentation in bandits and to propose an efficient Gibbs sampler for approximating the analytically unsolvable integral of logistic contextual bandits.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/25/2018

Contextual Bandits with Cross-learning

In the classical contextual bandits problem, in each round t, a learner ...
research
04/25/2016

Double Thompson Sampling for Dueling Bandits

In this paper, we propose a Double Thompson Sampling (D-TS) algorithm fo...
research
06/15/2021

Thompson Sampling for Unimodal Bandits

In this paper, we propose a Thompson Sampling algorithm for unimodal ban...
research
09/06/2019

A Pólya-Gamma Sampler for a Generalized Logistic Regression

In this paper we introduce a novel Bayesian data augmentation approach f...
research
02/02/2022

Efficient Algorithms for Learning to Control Bandits with Unobserved Contexts

Contextual bandits are widely-used in the study of learning-based contro...
research
09/29/2021

Apple Tasting Revisited: Bayesian Approaches to Partially Monitored Online Binary Classification

We consider a variant of online binary classification where a learner se...
research
03/21/2021

UCB-based Algorithms for Multinomial Logistic Regression Bandits

Out of the rich family of generalized linear bandits, perhaps the most w...

Please sign up or login with your details

Forgot password? Click here to reset