Multi-task Representation Learning for Pure Exploration in Linear Bandits

02/09/2023
by   Yihan Du, et al.
0

Despite the recent success of representation learning in sequential decision making, the study of the pure exploration scenario (i.e., identify the best option and minimize the sample complexity) is still limited. In this paper, we study multi-task representation learning for best arm identification in linear bandits (RepBAI-LB) and best policy identification in contextual linear bandits (RepBPI-CLB), two popular pure exploration settings with wide applications, e.g., clinical trials and web content optimization. In these two problems, all tasks share a common low-dimensional linear representation, and our goal is to leverage this feature to accelerate the best arm (policy) identification process for all tasks. For these problems, we design computationally and sample efficient algorithms DouExpDes and C-DouExpDes, which perform double experimental designs to plan optimal sample allocations for learning the global representation. We show that by learning the common representation among tasks, our sample complexity is significantly better than that of the native approach which solves tasks independently. To the best of our knowledge, this is the first work to demonstrate the benefits of representation learning for multi-task pure exploration.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/28/2022

On the Sample Complexity of Representation Learning in Multi-task Bandits with Global and Local structure

We investigate the sample complexity of learning the optimal arm for mul...
research
02/08/2021

Near-optimal Representation Learning for Linear Bandits and Linear RL

This paper studies representation learning for multi-task linear bandits...
research
07/02/2020

Gamification of Pure Exploration for Linear Bandits

We investigate an active pure-exploration setting, that includes best-ar...
research
01/13/2022

Non-Stationary Representation Learning in Sequential Linear Bandits

In this paper, we study representation learning for multi-task decision-...
research
06/09/2022

Choosing Answers in ε-Best-Answer Identification for Linear Bandits

In pure-exploration problems, information is gathered sequentially to an...
research
06/22/2021

Pure Exploration in Kernel and Neural Bandits

We study pure exploration in bandits, where the dimension of the feature...
research
02/02/2022

Active Multi-Task Representation Learning

To leverage the power of big data from source tasks and overcome the sca...

Please sign up or login with your details

Forgot password? Click here to reset