Corrupt Bandits for Preserving Local Privacy

08/16/2017
by   Pratik Gajane, et al.
0

We study a variant of the stochastic multi-armed bandit (MAB) problem in which the rewards are corrupted. In this framework, motivated by privacy preservation in online recommender systems, the goal is to maximize the sum of the (unobserved) rewards, based on the observation of transformation of these rewards through a stochastic corruption process with known parameters. We provide a lower bound on the expected regret of any bandit algorithm in this corrupted setting. We devise a frequentist algorithm, KLUCB-CF, and a Bayesian algorithm, TS-CF and give upper bounds on their regret. We also provide the appropriate corruption parameters to guarantee a desired level of local privacy and analyze how this impacts the regret. Finally, we present some experimental results that confirm our analysis.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/06/2020

Multi-Armed Bandits with Local Differential Privacy

This paper investigates the problem of regret minimization for multi-arm...
research
07/03/2023

Thompson Sampling under Bernoulli Rewards with Local Differential Privacy

This paper investigates the problem of regret minimization for multi-arm...
research
10/12/2019

What You See May Not Be What You Get: UCB Bandit Algorithms Robust to ε-Contamination

Motivated by applications of bandit algorithms in education, we consider...
research
10/17/2022

A Unified Algorithm for Stochastic Path Problems

We study reinforcement learning in stochastic path (SP) problems. The go...
research
04/25/2023

Communication-Constrained Bandits under Additive Gaussian Noise

We study a distributed stochastic multi-armed bandit where a client supp...
research
05/29/2019

Differential Privacy for Multi-armed Bandits: What Is It and What Is Its Cost?

We introduce a number of privacy definitions for the multi-armed bandit ...
research
10/13/2020

Local Differential Privacy for Bayesian Optimization

Motivated by the increasing concern about privacy in nowadays data-inten...

Please sign up or login with your details

Forgot password? Click here to reset