Collaborative Pure Exploration in Kernel Bandit

10/29/2021
by   Yihan Du, et al.
0

In this paper, we formulate a Collaborative Pure Exploration in Kernel Bandit problem (CoPE-KB), which provides a novel model for multi-agent multi-task decision making under limited communication and general reward functions, and is applicable to many online learning tasks, e.g., recommendation systems and network scheduling. We consider two settings of CoPE-KB, i.e., Fixed-Confidence (FC) and Fixed-Budget (FB), and design two optimal algorithms CoopKernelFC (for FC) and CoopKernelFB (for FB). Our algorithms are equipped with innovative and efficient kernelized estimators to simultaneously achieve computation and communication efficiency. Matching upper and lower bounds under both the statistical and communication metrics are established to demonstrate the optimality of our algorithms. The theoretical bounds successfully quantify the influences of task similarities on learning acceleration and only depend on the effective dimension of the kernelized feature space. Our analytical techniques, including data dimension decomposition, linear structured instance transformation and (communication) round-speedup induction, are novel and applicable to other bandit problems. Empirical evaluations are provided to validate our theoretical results and demonstrate the performance superiority of our algorithms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/02/2020

Instance-Sensitive Algorithms for Pure Exploration in Multinomial Logit Bandit

Motivated by real-world applications such as fast fashion retailing and ...
research
05/31/2022

Near-Optimal Collaborative Learning in Bandits

This paper introduces a general multi-agent bandit model in which each a...
research
05/27/2016

An optimal algorithm for the Thresholding Bandit Problem

We study a specific combinatorial pure exploration stochastic bandit pro...
research
08/18/2022

Communication-Efficient Collaborative Best Arm Identification

We investigate top-m arm identification, a basic problem in bandit theor...
research
04/05/2019

Collaborative Learning with Limited Interaction: Tight Bounds for Distributed Exploration in Multi-Armed Bandits

Best arm identification (or, pure exploration) in multi-armed bandits is...
research
05/08/2021

Pure Exploration Bandit Problem with General Reward Functions Depending on Full Distributions

In this paper, we study the pure exploration bandit model on general dis...
research
11/21/2017

Disagreement-based combinatorial pure exploration: Efficient algorithms and an analysis with localization

We design new algorithms for the combinatorial pure exploration problem ...

Please sign up or login with your details

Forgot password? Click here to reset