Understanding Adversarial Attacks on Observations in Deep Reinforcement Learning

06/30/2021
by   You Qiaoben, et al.
0

Recent works demonstrate that deep reinforcement learning (DRL) models are vulnerable to adversarial attacks which can decrease the victim's total reward by manipulating the observations. Compared with adversarial attacks in supervised learning, it is much more challenging to deceive a DRL model since the adversary has to infer the environmental dynamics. To address this issue, we reformulate the problem of adversarial attacks in function space and separate the previous gradient based attacks into several subspace. Following the analysis of the function space, we design a generic two-stage framework in the subspace where the adversary lures the agent to a target trajectory or a deceptive policy. In the first stage, we train a deceptive policy by hacking the environment, and discover a set of trajectories routing to the lowest reward. The adversary then misleads the victim to imitate the deceptive policy by perturbing the observations. Our method provides a tighter theoretical upper bound for the attacked agent's performance than the existing approaches. Extensive experiments demonstrate the superiority of our method and we achieve the state-of-the-art performance on both Atari and MuJoCo environments.

READ FULL TEXT

page 15

page 16

research
01/21/2021

Robust Reinforcement Learning on State Observations with Learned Optimal Adversary

We study the robustness of reinforcement learning (RL) with adversariall...
research
10/30/2020

Adversarial Attacks on Optimization based Planners

Trajectory planning is a key piece in the algorithmic architecture of a ...
research
10/07/2022

Improving Robustness of Deep Reinforcement Learning Agents: Environment Attack based on the Critic Network

To improve policy robustness of deep reinforcement learning agents, a li...
research
07/16/2018

Online Robust Policy Learning in the Presence of Unknown Adversaries

The growing prospect of deep reinforcement learning (DRL) being used in ...
research
11/21/2022

Backdoor Attacks on Multiagent Collaborative Systems

Backdoor attacks on reinforcement learning implant a backdoor in a victi...
research
10/13/2022

Observed Adversaries in Deep Reinforcement Learning

In this work, we point out the problem of observed adversaries for deep ...
research
09/29/2021

Mitigation of Adversarial Policy Imitation via Constrained Randomization of Policy (CRoP)

Deep reinforcement learning (DRL) policies are vulnerable to unauthorize...

Please sign up or login with your details

Forgot password? Click here to reset