On the Robustness of Epoch-Greedy in Multi-Agent Contextual Bandit Mechanisms

07/15/2023
by   Yinglun Xu, et al.
0

Efficient learning in multi-armed bandit mechanisms such as pay-per-click (PPC) auctions typically involves three challenges: 1) inducing truthful bidding behavior (incentives), 2) using personalization in the users (context), and 3) circumventing manipulations in click patterns (corruptions). Each of these challenges has been studied orthogonally in the literature; incentives have been addressed by a line of work on truthful multi-armed bandit mechanisms, context has been extensively tackled by contextual bandit algorithms, while corruptions have been discussed via a recent line of work on bandits with adversarial corruptions. Since these challenges co-exist, it is important to understand the robustness of each of these approaches in addressing the other challenges, provide algorithms that can handle all simultaneously, and highlight inherent limitations in this combination. In this work, we show that the most prominent contextual bandit algorithm, ϵ-greedy can be extended to handle the challenges introduced by strategic arms in the contextual multi-arm bandit mechanism setting. We further show that ϵ-greedy is inherently robust to adversarial data corruption attacks and achieves performance that degrades linearly with the amount of corruption.

READ FULL TEXT
research
06/26/2020

Online learning with Corrupted context: Corrupted Contextual Bandits

We consider a novel variant of the contextual bandit problem (i.e., the ...
research
05/10/2017

Context Attentive Bandits: Contextual Bandit with Restricted Context

We consider a novel formulation of the multi-armed bandit model, which w...
research
03/23/2020

Contextual Bandit-Based Channel Selection for Wireless LANs with Interference-Driven Feature Extraction

This paper proposes a radio channel selection algorithm based on a conte...
research
07/20/2023

Decentralized Smart Charging of Large-Scale EVs using Adaptive Multi-Agent Multi-Armed Bandits

The drastic growth of electric vehicles and photovoltaics can introduce ...
research
08/08/2023

AdaptEx: A Self-Service Contextual Bandit Platform

This paper presents AdaptEx, a self-service contextual bandit platform w...
research
07/04/2022

Autonomous Drug Design with Multi-armed Bandits

Recent developments in artificial intelligence and automation could pote...
research
05/14/2023

Multi-View Interactive Collaborative Filtering

In many scenarios, recommender system user interaction data such as clic...

Please sign up or login with your details

Forgot password? Click here to reset