Robust Deep Reinforcement Learning against Adversarial Perturbations on Observations

03/19/2020 ∙ by Huan Zhang, et al. ∙ 0

Deep Reinforcement Learning (DRL) is vulnerable to small adversarial perturbations on state observations. These perturbations do not alter the environment directly but can mislead the agent into making suboptimal decisions. We analyze the Markov Decision Process (MDP) under this threat model and utilize tools from the neural net-work verification literature to enable robust train-ing for DRL under observational perturbations. Our techniques are general and can be applied to both Deep Q Networks (DQN) and Deep Deterministic Policy Gradient (DDPG) algorithms for discrete and continuous action control problems. We demonstrate that our proposed training procedure significantly improves the robustness of DQN and DDPG agents under a suite of strong white-box attacks on observations, including a few novel attacks we specifically craft. Additionally, our training procedure can produce provable certificates for the robustness of a Deep RL agent.

READ FULL TEXT
POST COMMENT

Comments

There are no comments yet.

Authors

page 14

Code Repositories

StateAdvDRL

Code for "Robust Deep Reinforcement Learning against Adversarial Perturbations on Observations" http://arxiv.org/abs/2003.08938


view repo
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.