Achieving Fairness in the Stochastic Multi-armed Bandit Problem

07/23/2019
by   Vishakha Patil, et al.
0

We study an interesting variant of the stochastic multi-armed bandit problem, called the Fair-SMAB problem, where each arm is required to be pulled for at least a given fraction of the total available rounds. We investigate the interplay between learning and fairness in terms of a pre-specified vector denoting the fractions of guaranteed pulls. We define a fairness-aware regret, called r-Regret, that takes into account the above fairness constraints and naturally extends the conventional notion of regret. Our primary contribution is characterizing a class of Fair-SMAB algorithms by two parameters: the unfairness tolerance and the learning algorithm used as a black-box. We provide a fairness guarantee for this class that holds uniformly over time irrespective of the choice of the learning algorithm. In particular, when the learning algorithm is UCB1, we show that our algorithm achieves O( T) r-Regret. Finally, we evaluate the cost of fairness in terms of the conventional notion of regret.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/27/2019

Stochastic Multi-armed Bandits with Arm-specific Fairness Guarantees

We study an interesting variant of the stochastic multi-armed bandit pro...
research
05/23/2016

Fairness in Learning: Classic and Contextual Bandits

We introduce the study of fairness in multi-armed bandit problems. Our f...
research
06/23/2023

Trading-off price for data quality to achieve fair online allocation

We consider the problem of online allocation subject to a long-term fair...
research
07/11/2019

Fairness without Regret

A popular approach of achieving fairness in optimization problems is by ...
research
11/15/2022

On Penalization in Stochastic Multi-armed Bandits

We study an important variant of the stochastic multi-armed bandit (MAB)...
research
09/23/2022

An Efficient Algorithm for Fair Multi-Agent Multi-Armed Bandit with Low Regret

Recently a multi-agent variant of the classical multi-armed bandit was p...
research
01/15/2019

Combinatorial Sleeping Bandits with Fairness Constraints

The multi-armed bandit (MAB) model has been widely adopted for studying ...

Please sign up or login with your details

Forgot password? Click here to reset