Vulnerability of Deep Reinforcement Learning to Policy Induction Attacks

01/16/2017
by   Vahid Behzadan, et al.
0

Deep learning classifiers are known to be inherently vulnerable to manipulation by intentionally perturbed inputs, named adversarial examples. In this work, we establish that reinforcement learning techniques based on Deep Q-Networks (DQNs) are also vulnerable to adversarial input perturbations, and verify the transferability of adversarial examples across different DQN models. Furthermore, we present a novel class of attacks based on this vulnerability that enable policy manipulation and induction in the learning process of DQNs. We propose an attack mechanism that exploits the transferability of adversarial examples to implement policy induction attacks on DQNs, and demonstrate its efficacy and impact through experimental study of a game-learning scenario.

READ FULL TEXT

page 8

page 9

research
06/04/2018

Mitigation of Policy Manipulation Attacks on Deep Q-Networks with Parameter-Space Noise

Recent developments have established the vulnerability of deep reinforce...
research
05/18/2017

Delving into adversarial attacks on deep policies

Adversarial examples have been shown to exist for a variety of deep lear...
research
06/18/2022

Comment on Transferability and Input Transformation with Additive Noise

Adversarial attacks have verified the existence of the vulnerability of ...
research
03/31/2019

On the Vulnerability of CNN Classifiers in EEG-Based BCIs

Deep learning has been successfully used in numerous applications becaus...
research
10/02/2017

Detecting Adversarial Attacks on Neural Network Policies with Visual Foresight

Deep reinforcement learning has shown promising results in learning cont...
research
10/25/2019

Effectiveness of random deep feature selection for securing image manipulation detectors against adversarial examples

We investigate if the random feature selection approach proposed in [1] ...

Please sign up or login with your details

Forgot password? Click here to reset