A Comprehensive Study on Dataset Distillation: Performance, Privacy, Robustness and Fairness

05/05/2023
by   Zongxiong Chen, et al.
0

The aim of dataset distillation is to encode the rich features of an original dataset into a tiny dataset. It is a promising approach to accelerate neural network training and related studies. Different approaches have been proposed to improve the informativeness and generalization performance of distilled images. However, no work has comprehensively analyzed this technique from a security perspective and there is a lack of systematic understanding of potential risks. In this work, we conduct extensive experiments to evaluate current state-of-the-art dataset distillation methods. We successfully use membership inference attacks to show that privacy risks still remain. Our work also demonstrates that dataset distillation can cause varying degrees of impact on model robustness and amplify model unfairness across classes when making predictions. This work offers a large-scale benchmarking framework for dataset distillation evaluation.

READ FULL TEXT

page 3

page 7

page 11

page 12

research
03/06/2023

Students Parrot Their Teachers: Membership Inference on Model Distillation

Model distillation is frequently proposed as a technique to reduce the p...
research
11/02/2021

Knowledge Cross-Distillation for Membership Privacy

A membership inference attack (MIA) poses privacy risks on the training ...
research
10/15/2021

Mitigating Membership Inference Attacks by Self-Distillation Through a Novel Ensemble Architecture

Membership inference attacks are a key measure to evaluate privacy leaka...
research
05/03/2023

A Survey on Dataset Distillation: Approaches, Applications and Future Directions

Dataset distillation is attracting more attention in machine learning as...
research
05/14/2023

Improving Defensive Distillation using Teacher Assistant

Adversarial attacks pose a significant threat to the security and safety...
research
04/09/2018

Large scale distributed neural network training through online distillation

Techniques such as ensembling and distillation promise model quality imp...
research
02/02/2023

Dataset Distillation Fixes Dataset Reconstruction Attacks

Modern deep learning requires large volumes of data, which could contain...

Please sign up or login with your details

Forgot password? Click here to reset