Privacy Preserving Off-Policy Evaluation

02/01/2019
by   Tengyang Xie, et al.
0

Many reinforcement learning applications involve the use of data that is sensitive, such as medical records of patients or financial information. However, most current reinforcement learning methods can leak information contained within the (possibly sensitive) data on which they are trained. To address this problem, we present the first differentially private approach for off-policy evaluation. We provide a theoretical analysis of the privacy-preserving properties of our algorithm and analyze its utility (speed of convergence). After describing some results of this theoretical analysis, we show empirically that our method outperforms previous methods (which are restricted to the on-policy setting).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/07/2016

Differentially Private Policy Evaluation

We present the first differentially private algorithms for reinforcement...
research
03/25/2021

Differentially Private Normalizing Flows for Privacy-Preserving Density Estimation

Normalizing flow models have risen as a popular solution to the problem ...
research
08/19/2023

DPMAC: Differentially Private Communication for Cooperative Multi-Agent Reinforcement Learning

Communication lays the foundation for cooperation in human society and i...
research
03/03/2023

Exploring Machine Learning Privacy/Utility trade-off from a hyperparameters Lens

Machine Learning (ML) architectures have been applied to several applica...
research
02/02/2022

Improved Regret for Differentially Private Exploration in Linear MDP

We study privacy-preserving exploration in sequential decision-making fo...
research
12/30/2020

Privacy-Constrained Policies via Mutual Information Regularized Policy Gradients

As reinforcement learning techniques are increasingly applied to real-wo...

Please sign up or login with your details

Forgot password? Click here to reset