Locally Differentially Private Naive Bayes Classification

05/03/2019
by   Emre Yilmaz, et al.
0

In machine learning, classification models need to be trained in order to predict class labels. When the training data contains personal information about individuals, collecting training data becomes difficult due to privacy concerns. Local differential privacy is a definition to measure the individual privacy when there is no trusted data curator. Individuals interact with an untrusted data aggregator who obtains statistical information about the population without learning personal data. In order to train a Naive Bayes classifier in an untrusted setting, we propose to use methods satisfying local differential privacy. Individuals send their perturbed inputs that keep the relationship between the feature values and class labels. The data aggregator estimates all probabilities needed by the Naive Bayes classifier. Then, new instances can be classified based on the estimated probabilities. We propose solutions for both discrete and continuous data. In order to eliminate high amount of noise and decrease communication cost in multi-dimensional data, we propose utilizing dimensionality reduction techniques which can be applied by individuals before perturbing their inputs. Our experimental results show that the accuracy of the Naive Bayes classifier is maintained even when the individual privacy is guaranteed under local differential privacy, and that using dimensionality reduction enhances the accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/31/2020

Differentially Private Naïve Bayes Classifier using Smooth Sensitivity

With the increasing collection of users' data, protecting individual pri...
research
10/20/2020

DuetSGX: Differential Privacy with Secure Hardware

Differential privacy offers a formal privacy guarantee for individuals, ...
research
11/22/2019

Privacy-preserving parametric inference: a case for robust statistics

Differential privacy is a cryptographically-motivated approach to privac...
research
06/19/2023

Prior-itizing Privacy: A Bayesian Approach to Setting the Privacy Budget in Differential Privacy

When releasing outputs from confidential data, agencies need to balance ...
research
02/11/2021

On Deep Learning with Label Differential Privacy

In many machine learning applications, the training data can contain hig...
research
12/08/2022

Structure of Classifier Boundaries: Case Study for a Naive Bayes Classifier

Whether based on models, training data or a combination, classifiers pla...
research
10/12/2017

An Improved Naive Bayes Classifier-based Noise Detection Technique for Classifying User Phone Call Behavior

The presence of noisy instances in mobile phone data is a fundamental is...

Please sign up or login with your details

Forgot password? Click here to reset