Privacy-preserving Active Learning on Sensitive Data for User Intent Classification

03/26/2019
by   Oluwaseyi Feyisetan, et al.
0

Active learning holds promise of significantly reducing data annotation costs while maintaining reasonable model performance. However, it requires sending data to annotators for labeling. This presents a possible privacy leak when the training set includes sensitive user data. In this paper, we describe an approach for carrying out privacy preserving active learning with quantifiable guarantees. We evaluate our approach by showing the tradeoff between privacy, utility and annotation budget on a binary classification task in a active learning setting.

READ FULL TEXT
research
04/18/2022

Active Learning with Weak Labels for Gaussian Processes

Annotating data for supervised learning can be costly. When the annotati...
research
12/02/2015

Active Learning for Delineation of Curvilinear Structures

Many recent delineation techniques owe much of their increased effective...
research
11/05/2018

Model Extraction and Active Learning

Machine learning is being increasingly used by individuals, research ins...
research
11/20/2022

Finding active galactic nuclei through Fink

We present the Active Galactic Nuclei (AGN) classifier as currently impl...
research
03/03/2017

Active Learning for Cost-Sensitive Classification

We design an active learning algorithm for cost-sensitive multiclass cla...
research
08/07/2020

Deep Active Learning with Crowdsourcing Data for Privacy Policy Classification

Privacy policies are statements that notify users of the services' data ...
research
08/17/2019

ED2: Two-stage Active Learning for Error Detection – Technical Report

Traditional error detection approaches require user-defined parameters a...

Please sign up or login with your details

Forgot password? Click here to reset