Risk averse non-stationary multi-armed bandits

09/28/2021
by   Leo Benac, et al.
0

This paper tackles the risk averse multi-armed bandits problem when incurred losses are non-stationary. The conditional value-at-risk (CVaR) is used as the objective function. Two estimation methods are proposed for this objective function in the presence of non-stationary losses, one relying on a weighted empirical distribution of losses and another on the dual representation of the CVaR. Such estimates can then be embedded into classic arm selection methods such as epsilon-greedy policies. Simulation experiments assess the performance of the arm selection algorithms based on the two novel estimation approaches, and such policies are shown to outperform naive benchmarks not taking non-stationarity into account.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/04/2019

Risk-aware Multi-armed Bandits Using Conditional Value-at-Risk

Traditional multi-armed bandit problems are geared towards finding the a...
research
12/27/2017

Active Search for High Recall: a Non-Stationary Extension of Thompson Sampling

We consider the problem of Active Search, where a maximum of relevant ob...
research
06/21/2019

Entropic Risk Measure in Policy Search

With the increasing pace of automation, modern robotic systems need to a...
research
03/29/2017

Bandit-Based Model Selection for Deformable Object Manipulation

We present a novel approach to deformable object manipulation that does ...
research
09/18/2023

Task Selection and Assignment for Multi-modal Multi-task Dialogue Act Classification with Non-stationary Multi-armed Bandits

Multi-task learning (MTL) aims to improve the performance of a primary t...
research
04/20/2018

Delegating via Quitting Games

Delegation allows an agent to request that another agent completes a tas...
research
09/05/2014

Simulating Non Stationary Operators in Search Algorithms

In this paper, we propose a model for simulating search operators whose ...

Please sign up or login with your details

Forgot password? Click here to reset