Characterizing Attacks on Deep Reinforcement Learning

07/21/2019
by   Chaowei Xiao, et al.
4

Deep reinforcement learning (DRL) has achieved great success in various applications. However, recent studies show that machine learning models are vulnerable to adversarial attacks. DRL models have been attacked by adding perturbations to observations. While such observation based attack is only one aspect of potential attacks on DRL, other forms of attacks which are more practical require further analysis, such as manipulating environment dynamics. Therefore, we propose to understand the vulnerabilities of DRL from various perspectives and provide a thorough taxonomy of potential attacks. We conduct the first set of experiments on the unexplored parts within the taxonomy. In addition to current observation based attacks against DRL, we propose the first targeted attacks based on action space and environment dynamics. We also introduce the online sequential attacks based on temporal consistency information among frames. To better estimate gradient in black-box setting, we propose a sampling strategy and theoretically prove its efficiency and estimation error bound. We conduct extensive experiments to compare the effectiveness of different attacks with several baselines in various environments, including game playing, robotics control, and autonomous driving.

READ FULL TEXT

page 5

page 12

research
01/27/2020

Challenges and Countermeasures for Adversarial Attacks on Deep Reinforcement Learning

Deep Reinforcement Learning (DRL) has numerous applications in the real ...
research
11/22/2022

Don't Watch Me: A Spatio-Temporal Trojan Attack on Deep-Reinforcement-Learning-Augment Autonomous Driving

Deep reinforcement learning (DRL) is one of the most popular algorithms ...
research
04/24/2019

How You Act Tells a Lot: Privacy-Leakage Attack on Deep Reinforcement Learning

Machine learning has been widely applied to various applications, some o...
research
06/09/2020

Stealing Deep Reinforcement Learning Models for Fun and Profit

In this paper, we present the first attack methodology to extract black-...
research
06/14/2022

Defending Observation Attacks in Deep Reinforcement Learning via Detection and Denoising

Neural network policies trained using Deep Reinforcement Learning (DRL) ...
research
06/03/2019

Adversarial Exploitation of Policy Imitation

This paper investigates a class of attacks targeting the confidentiality...
research
09/16/2021

Targeted Attack on Deep RL-based Autonomous Driving with Learned Visual Patterns

Recent studies demonstrated the vulnerability of control policies learne...

Please sign up or login with your details

Forgot password? Click here to reset