Pure exploration in multi-armed bandits with low rank structure using oblivious sampler

06/28/2023
by   Yaxiong Liu, et al.
0

In this paper, we consider the low rank structure of the reward sequence of the pure exploration problems. Firstly, we propose the separated setting in pure exploration problem, where the exploration strategy cannot receive the feedback of its explorations. Due to this separation, it requires that the exploration strategy to sample the arms obliviously. By involving the kernel information of the reward vectors, we provide efficient algorithms for both time-varying and fixed cases with regret bound O(d√((ln N)/n)). Then, we show the lower bound to the pure exploration in multi-armed bandits with low rank sequence. There is an O(√(ln N)) gap between our upper bound and the lower bound.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/12/2019

Incentivized Exploration for Multi-Armed Bandits under Reward Drift

We study incentivized exploration for the multi-armed bandit (MAB) probl...
research
06/22/2021

A Unified Framework for Conservative Exploration

We study bandits and reinforcement learning (RL) subject to a conservati...
research
10/27/2021

Heterogeneous Multi-player Multi-armed Bandits: Closing the Gap and Generalization

Despite the significant interests and many progresses in decentralized m...
research
04/08/2021

Incentivizing Exploration in Linear Bandits under Information Gap

We study the problem of incentivizing exploration for myopic users in li...
research
01/08/2019

Bilinear Bandits with Low-rank Structure

We introduce the bilinear bandit problem with low-rank structure where a...
research
03/01/2023

Multi-Armed Bandits with Generalized Temporally-Partitioned Rewards

Decision-making problems of sequential nature, where decisions made in t...
research
06/30/2020

Forced-exploration free Strategies for Unimodal Bandits

We consider a multi-armed bandit problem specified by a set of Gaussian ...

Please sign up or login with your details

Forgot password? Click here to reset