Decoy Bandits Dueling on a Poset

02/08/2016
by   Julien Audiffren, et al.
0

We adress the problem of dueling bandits defined on partially ordered sets, or posets. In this setting, arms may not be comparable, and there may be several (incomparable) optimal arms. We propose an algorithm, UnchainedBandits, that efficiently finds the set of optimal arms of any poset even when pairs of comparable arms cannot be distinguished from pairs of incomparable arms, with a set of minimal assumptions. This algorithm relies on the concept of decoys, which stems from social psychology. For the easier case where the incomparability information may be accessible, we propose a second algorithm, SlicingBandits, which takes advantage of this information and achieves a very significant gain of performance compared to UnchainedBandits. We provide theoretical guarantees and experimental evaluation for both algorithms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/20/2014

Unimodal Bandits: Regret Lower Bounds and Optimal Algorithms

We consider stochastic multi-armed bandits where the expected reward is ...
research
09/08/2016

On Sequential Elimination Algorithms for Best-Arm Identification in Multi-Armed Bandits

We consider the best-arm identification problem in multi-armed bandits, ...
research
07/04/2019

Reducing Exploration of Dying Arms in Mortal Bandits

Mortal bandits have proven to be extremely useful for providing news art...
research
06/03/2021

Sleeping Combinatorial Bandits

In this paper, we study an interesting combination of sleeping and combi...
research
10/28/2018

Exploring k out of Top ρ Fraction of Arms in Stochastic Bandits

This paper studies the problem of identifying any k distinct arms among ...
research
02/25/2018

Model of a motion of substance in a channel of a network consisting of two arms

We study the problem of the motion of substance in a channel of a networ...
research
09/30/2021

Adapting Bandit Algorithms for Settings with Sequentially Available Arms

Although the classical version of the Multi-Armed Bandits (MAB) framewor...

Please sign up or login with your details

Forgot password? Click here to reset