Adversarial Sleeping Bandit Problems with Multiple Plays: Algorithm and Ranking Application

07/27/2023
by   Jianjun Yuan, et al.
0

This paper presents an efficient algorithm to solve the sleeping bandit with multiple plays problem in the context of an online recommendation system. The problem involves bounded, adversarial loss and unknown i.i.d. distributions for arm availability. The proposed algorithm extends the sleeping bandit algorithm for single arm selection and is guaranteed to achieve theoretical performance with regret upper bounded by (kN^2√(Tlog T)), where k is the number of arms selected per time step, N is the total number of arms, and T is the time horizon.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/07/2020

Online Model Selection: a Rested Bandit Formulation

Motivated by a natural problem in online model selection with bandit inf...
research
11/25/2019

Minimax Optimal Algorithms for Adversarial Bandit Problem with Multiple Plays

We investigate the adversarial bandit problem with multiple plays under ...
research
01/22/2023

Doubly Adversarial Federated Bandits

We study a new non-stochastic federated multi-armed bandit problem with ...
research
07/24/2017

Combinatorial Multi-armed Bandit with Probabilistically Triggered Arms: A Case with Bounded Regret

In this paper, we study the combinatorial multi-armed bandit problem (CM...
research
02/15/2021

Top-k eXtreme Contextual Bandits with Arm Hierarchy

Motivated by modern applications, such as online advertisement and recom...
research
02/09/2021

A Multi-Arm Bandit Approach To Subset Selection Under Constraints

We explore the class of problems where a central planner needs to select...
research
05/10/2022

Risk Aversion In Learning Algorithms and an Application To Recommendation Systems

Consider a bandit learning environment. We demonstrate that popular lear...

Please sign up or login with your details

Forgot password? Click here to reset