Adversarial Attacks on Stochastic Bandits

10/29/2018
by   Kwang-Sung Jun, et al.
0

We study adversarial attacks that manipulate the reward signals to control the actions chosen by a stochastic multi-armed bandit algorithm. We propose the first attack against two popular bandit algorithms: ϵ-greedy and UCB, without knowledge of the mean rewards. The attacker is able to spend only logarithmic effort, multiplied by a problem-specific parameter that becomes smaller as the bandit problem gets easier to attack. The result means the attacker can easily hijack the behavior of the bandit algorithm to promote or obstruct certain actions, say, a particular medical treatment. As bandits are seeing increasingly wide use in practice, our study exposes a significant security threat.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/16/2019

Data Poisoning Attacks on Stochastic Bandits

Stochastic multi-armed bandits form a class of online learning problems ...
research
01/30/2023

Adversarial Attacks on Adversarial Bandits

We study a security threat to adversarial multi-armed bandits, in which ...
research
08/17/2018

Data Poisoning Attacks in Contextual Bandits

We study offline data poisoning attacks in contextual bandits, a class o...
research
10/18/2021

When Are Linear Stochastic Bandits Attackable?

We study adversarial attacks on linear stochastic bandits, a sequential ...
research
02/15/2021

Secure-UCB: Saving Stochastic Bandits from Poisoning Attacks via Limited Data Verification

This paper studies bandit algorithms under data poisoning attacks in a b...
research
03/04/2020

Bandits with adversarial scaling

We study "adversarial scaling", a multi-armed bandit model where rewards...
research
02/10/2020

Adversarial Attacks on Linear Contextual Bandits

Contextual bandit algorithms are applied in a wide range of domains, fro...

Please sign up or login with your details

Forgot password? Click here to reset