Exponentiated Gradient LINUCB for Contextual Multi-Armed Bandits

05/10/2013
by   Djallel Bouneffouf, et al.
0

We present Exponentiated Gradient LINUCB, an algorithm for con-textual multi-armed bandits. This algorithm uses Exponentiated Gradient to find the optimal exploration of the LINUCB. Within a deliberately designed offline simulation framework we conduct evaluations with real online event log data. The experimental results demonstrate that our algorithm outperforms surveyed algorithms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/12/2017

Value Directed Exploration in Multi-Armed Bandits with Structured Priors

Multi-armed bandits are a quintessential machine learning problem requir...
research
11/11/2018

Adapting multi-armed bandits policies to contextual bandits scenarios

This work explores adaptations of successful multi-armed bandits policie...
research
04/15/2019

Introduction to Multi-Armed Bandits

Multi-armed bandits a simple but very powerful framework for algorithms ...
research
02/15/2023

Genetic multi-armed bandits: a reinforcement learning approach for discrete optimization via simulation

This paper proposes a new algorithm, referred to as GMAB, that combines ...
research
10/23/2021

Towards the D-Optimal Online Experiment Design for Recommender Selection

Selecting the optimal recommender via online exploration-exploitation is...
research
03/22/2023

A simulation framework of procurement operations in the container logistics industry

This study proposes a simulation framework of procurement operations in ...
research
07/10/2019

Productization Challenges of Contextual Multi-Armed Bandits

Contextual Multi-Armed Bandits is a well-known and accepted online optim...

Please sign up or login with your details

Forgot password? Click here to reset