Learning with Differential Privacy: Stability, Learnability and the Sufficiency and Necessity of ERM Principle

02/23/2015
by   Yu-Xiang Wang, et al.
0

While machine learning has proven to be a powerful data-driven solution to many real-life problems, its use in sensitive domains has been limited due to privacy concerns. A popular approach known as **differential privacy** offers provable privacy guarantees, but it is often observed in practice that it could substantially hamper learning accuracy. In this paper we study the learnability (whether a problem can be learned by any algorithm) under Vapnik's general learning setting with differential privacy constraint, and reveal some intricate relationships between privacy, stability and learnability. In particular, we show that a problem is privately learnable **if an only if** there is a private algorithm that asymptotically minimizes the empirical risk (AERM). In contrast, for non-private learning AERM alone is not sufficient for learnability. This result suggests that when searching for private learning algorithms, we can restrict the search to algorithms that are AERM. In light of this, we propose a conceptual procedure that always finds a universally consistent algorithm whenever the problem is learnable under privacy constraint. We also propose a generic and practical algorithm and show that under very general conditions it privately learns a wide class of learning problems. Lastly, we extend some of the results to the more practical (ϵ,δ)-differential privacy and establish the existence of a phase-transition on the class of problems that are approximately privately learnable with respect to how small δ needs to be.

READ FULL TEXT
research
06/25/2020

Stability Enhanced Privacy and Applications in Private Stochastic Gradient Descent

Private machine learning involves addition of noise while training, resu...
research
07/31/2018

Subsampled Rényi Differential Privacy and Analytical Moments Accountant

We study the problem of subsampling in differential privacy (DP), a ques...
research
11/19/2018

How to Use Heuristics for Differential Privacy

We develop theory for using heuristics to solve computationally hard pro...
research
01/04/2017

Private Incremental Regression

Data is continuously generated by modern data sources, and a recent chal...
research
02/21/2020

Privately Learning Markov Random Fields

We consider the problem of learning Markov Random Fields (including the ...
research
02/11/2021

On Deep Learning with Label Differential Privacy

In many machine learning applications, the training data can contain hig...
research
09/12/2023

Private Distribution Testing with Heterogeneous Constraints: Your Epsilon Might Not Be Mine

Private closeness testing asks to decide whether the underlying probabil...

Please sign up or login with your details

Forgot password? Click here to reset