Mitigation of Adversarial Policy Imitation via Constrained Randomization of Policy (CRoP)

09/29/2021
by   Nancirose Piazza, et al.
0

Deep reinforcement learning (DRL) policies are vulnerable to unauthorized replication attacks, where an adversary exploits imitation learning to reproduce target policies from observed behavior. In this paper, we propose Constrained Randomization of Policy (CRoP) as a mitigation technique against such attacks. CRoP induces the execution of sub-optimal actions at random under performance loss constraints. We present a parametric analysis of CRoP, address the optimality of CRoP, and establish theoretical bounds on the adversarial budget and the expectation of loss. Furthermore, we report the experimental evaluation of CRoP in Atari environments under adversarial imitation, which demonstrate the efficacy and feasibility of our proposed method against policy replication attacks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/03/2019

Adversarial Exploitation of Policy Imitation

This paper investigates a class of attacks targeting the confidentiality...
research
10/27/2019

BAIL: Best-Action Imitation Learning for Batch Deep Reinforcement Learning

The field of Deep Reinforcement Learning (DRL) has recently seen a surge...
research
08/21/2020

Adversarial Imitation Learning via Random Search

Developing agents that can perform challenging complex tasks is the goal...
research
06/30/2021

Understanding Adversarial Attacks on Observations in Deep Reinforcement Learning

Recent works demonstrate that deep reinforcement learning (DRL) models a...
research
01/31/2020

Preventing Imitation Learning with Adversarial Policy Ensembles

Imitation learning can reproduce policies by observing experts, which po...
research
06/04/2018

Mitigation of Policy Manipulation Attacks on Deep Q-Networks with Parameter-Space Noise

Recent developments have established the vulnerability of deep reinforce...

Please sign up or login with your details

Forgot password? Click here to reset