Incentivizing Exploration with Linear Contexts and Combinatorial Actions

06/03/2023
by   Mark Sellke, et al.
0

We advance the study of incentivized bandit exploration, in which arm choices are viewed as recommendations and are required to be Bayesian incentive compatible. Recent work has shown under certain independence assumptions that after collecting enough initial samples, the popular Thompson sampling algorithm becomes incentive compatible. We give an analog of this result for linear bandits, where the independence of the prior is replaced by a natural convexity condition. This opens up the possibility of efficient and regret-optimal incentivized exploration in high-dimensional action spaces. In the semibandit model, we also improve the sample complexity for the pre-Thompson sampling phase of initial data collection.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/01/2022

Incentivizing Combinatorial Bandit Exploration

Consider a bandit algorithm that recommends actions to self-interested u...
research
02/03/2020

Sample Complexity of Incentivized Exploration

We consider incentivized exploration: a version of multi-armed bandits w...
research
07/23/2022

Exploration in Linear Bandits with Rich Action Sets and its Implications for Inference

We present a non-asymptotic lower bound on the eigenspectrum of the desi...
research
10/30/2020

The Combinatorial Multi-Bandit Problem and its Application to Energy Management

We study a Combinatorial Multi-Bandit Problem motivated by applications ...
research
05/16/2019

Fiduciary Bandits

Recommendation systems often face exploration-exploitation tradeoffs: th...
research
07/17/2013

From Bandits to Experts: A Tale of Domination and Independence

We consider the partial observability model for multi-armed bandits, int...
research
10/08/2019

Credible Sample Elicitation by Deep Learning, for Deep Learning

It is important to collect credible training samples (x,y) for building ...

Please sign up or login with your details

Forgot password? Click here to reset