InfoShape: Task-Based Neural Data Shaping via Mutual Information

10/26/2022
by   Homa Esfahanizadeh, et al.
0

The use of mutual information as a tool in private data sharing has remained an open challenge due to the difficulty of its estimation in practice. In this paper, we propose InfoShape, a task-based encoder that aims to remove unnecessary sensitive information from training data while maintaining enough relevant information for a particular ML training task. We achieve this goal by utilizing mutual information estimators that are based on neural networks, in order to measure two performance metrics, privacy and utility. Using these together in a Lagrangian optimization, we train a separate neural network as a lossy encoder. We empirically show that InfoShape is capable of shaping the encoded samples to be informative for a specific downstream task while eliminating unnecessary sensitive information. Moreover, we demonstrate that the classification accuracy of downstream models has a meaningful connection with our utility and privacy measures.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/17/2021

Data Privacy and Utility Trade-Off Based on Mutual Information Neural Estimator

In the era of big data and the Internet of Things (IoT), data owners nee...
research
04/22/2019

Learning gradient-based ICA by neurally estimating mutual information

Several methods of estimating the mutual information of random variables...
research
03/31/2022

Mutual information estimation for graph convolutional neural networks

Measuring model performance is a key issue for deep learning practitione...
research
07/03/2021

Privacy-Preserving Representation Learning on Graphs: A Mutual Information Perspective

Learning with graphs has attracted significant attention recently. Exist...
research
06/29/2020

Neural Mutual Information Estimation for Channel Coding: State-of-the-Art Estimators, Analysis, and Performance Comparison

Deep learning based physical layer design, i.e., using dense neural netw...
research
09/04/2018

Pointwise HSIC: A Linear-Time Kernelized Co-occurrence Norm for Sparse Linguistic Expressions

In this paper, we propose a new kernel-based co-occurrence measure that ...
research
05/13/2018

Doing the impossible: Why neural networks can be trained at all

As deep neural networks grow in size, from thousands to millions to bill...

Please sign up or login with your details

Forgot password? Click here to reset