Dataset Distillation using Parameter Pruning

09/29/2022
by   Guang Li, et al.
0

The acquisition of advanced models relies on large datasets in many fields, which makes storing datasets and training models expensive. As a solution, dataset distillation can synthesize a small dataset such that models trained on it achieve high performance on par with the original large dataset. The recently proposed dataset distillation method by matching network parameters has been proved effective for several datasets. However, a few parameters in the distillation process are difficult to match, which harms the distillation performance. Based on this observation, this paper proposes a new method to solve the problem using parameter pruning. The proposed method can synthesize more robust distilled datasets and improve the distillation performance by pruning difficult-to-match parameters in the distillation process. Experimental results on three datasets show that the proposed method outperformed other SOTA dataset distillation methods.

READ FULL TEXT

page 3

page 4

research
09/29/2022

Dataset Distillation for Medical Dataset Sharing

Sharing medical datasets between hospitals is challenging because of the...
research
08/03/2023

Baby Llama: knowledge distillation from an ensemble of teachers trained on a small dataset with no performance penalty

We present our proposed solution to the BabyLM challenge [arXiv:2301.117...
research
05/28/2023

Distill Gold from Massive Ores: Efficient Dataset Distillation via Critical Samples Selection

Data-efficient learning has drawn significant attention, especially give...
research
01/13/2023

A Comprehensive Survey to Dataset Distillation

Deep learning technology has unprecedentedly developed in the last decad...
research
03/22/2022

Dataset Distillation by Matching Training Trajectories

Dataset distillation is the task of synthesizing a small dataset such th...
research
01/11/2023

Data Distillation: A Survey

The popularity of deep learning has led to the curation of a vast number...
research
06/15/2020

Flexible Dataset Distillation: Learn Labels Instead of Images

We study the problem of dataset distillation - creating a small set of s...

Please sign up or login with your details

Forgot password? Click here to reset