Efficient differentially private learning improves drug sensitivity prediction

06/07/2016
by   Antti Honkela, et al.
0

Users of a personalised recommendation system face a dilemma: recommendations can be improved by learning from data, but only if the other users are willing to share their private information. Good personalised predictions are vitally important in precision medicine, but genomic information on which the predictions are based is also particularly sensitive, as it directly identifies the patients and hence cannot easily be anonymised. Differential privacy has emerged as a potentially promising solution: privacy is considered sufficient if presence of individual patients cannot be distinguished. However, differentially private learning with current methods does not improve predictions with feasible data sizes and dimensionalities. Here we show that useful predictors can be learned under powerful differential privacy guarantees, and even from moderately-sized data sets, by demonstrating significant improvements with a new robust private regression method in the accuracy of private drug sensitivity prediction. The method combines two key properties not present even in recent proposals, which can be generalised to other predictors: we prove it is asymptotically consistently and efficiently private, and demonstrate that it performs well on finite data. Good finite data performance is achieved by limiting the sharing of private information by decreasing the dimensionality and by projecting outliers to fit tighter bounds, therefore needing to add less noise for equal privacy. As already the simple-to-implement method shows promise on the challenging genomic data, we anticipate rapid progress towards practical applications in many fields, such as mobile sensing and social media, in addition to the badly needed precision medicine solutions.

READ FULL TEXT

page 7

page 24

research
01/29/2019

Representation Transfer for Differentially Private Drug Sensitivity Prediction

Motivation: Human genomic datasets often contain sensitive information t...
research
03/31/2020

Differentially Private Naïve Bayes Classifier using Smooth Sensitivity

With the increasing collection of users' data, protecting individual pri...
research
07/21/2022

Widespread Underestimation of Sensitivity in Differentially Private Libraries and How to Fix It

We identify a new class of vulnerabilities in implementations of differe...
research
10/20/2022

Private Algorithms with Private Predictions

When applying differential privacy to sensitive data, a common way of ge...
research
03/01/2020

Federating Recommendations Using Differentially Private Prototypes

Machine learning methods allow us to make recommendations to users in ap...
research
07/10/2020

Differentially Private Simple Linear Regression

Economics and social science research often require analyzing datasets o...
research
02/15/2022

Private Quantiles Estimation in the Presence of Atoms

We address the differentially private estimation of multiple quantiles (...

Please sign up or login with your details

Forgot password? Click here to reset