Tight Lower Bounds for Combinatorial Multi-Armed Bandits

02/13/2020
by   Nadav Merlis, et al.
11

The Combinatorial Multi-Armed Bandit problem is a sequential decision-making problem in which an agent selects a set of arms on each round, observes feedback for each of these arms and aims to maximize a known reward function of the arms it chose. While previous work proved regret upper bounds in this setting for general reward functions, only a few works provided matching lower bounds, all for specific reward functions. In this work, we prove regret lower bounds for combinatorial bandits that hold under mild assumptions for all smooth reward functions. We derive both problem-dependent and problem-independent bounds and show that the recently proposed Gini-weighted smoothness parameter <cit.> also determines the lower bounds for monotone reward functions. Notably, this implies that our lower bounds are tight up to log-factors.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/08/2019

Batch-Size Independent Regret Bounds for the Combinatorial Multi-Armed Bandit Problem

We consider the combinatorial multi-armed bandit (CMAB) problem, where t...
research
10/12/2021

Dare not to Ask: Problem-Dependent Guarantees for Budgeted Bandits

We consider a stochastic multi-armed bandit setting where feedback is li...
research
07/14/2023

On Interpolating Experts and Multi-Armed Bandits

Learning with expert advice and multi-armed bandit are two classic onlin...
research
01/20/2021

Near-Optimal Regret Bounds for Contextual Combinatorial Semi-Bandits with Linear Payoff Functions

The contextual combinatorial semi-bandit problem with linear payoff func...
research
07/02/2020

Structure Adaptive Algorithms for Stochastic Bandits

We study reward maximisation in a wide class of structured stochastic mu...
research
10/25/2022

PopArt: Efficient Sparse Regression and Experimental Design for Optimal Sparse Linear Bandits

In sparse linear bandits, a learning agent sequentially selects an actio...
research
06/08/2022

Uplifting Bandits

We introduce a multi-armed bandit model where the reward is a sum of mul...

Please sign up or login with your details

Forgot password? Click here to reset