Dataset Distillation by Matching Training Trajectories

03/22/2022
by   George Cazenavette, et al.
16

Dataset distillation is the task of synthesizing a small dataset such that a model trained on the synthetic set will match the test accuracy of the model trained on the full dataset. In this paper, we propose a new formulation that optimizes our distilled data to guide networks to a similar state as those trained on real data across many training steps. Given a network, we train it for several iterations on our distilled data and optimize the distilled data with respect to the distance between the synthetically trained parameters and the parameters trained on real data. To efficiently obtain the initial and target network parameters for large-scale datasets, we pre-compute and store training trajectories of expert networks trained on the real dataset. Our method handily outperforms existing methods and also allows us to distill higher-resolution visual data.

READ FULL TEXT

page 15

page 16

page 17

page 18

page 19

page 20

page 21

page 22

research
08/03/2023

Baby Llama: knowledge distillation from an ensemble of teachers trained on a small dataset with no performance penalty

We present our proposed solution to the BabyLM challenge [arXiv:2301.117...
research
07/24/2023

Rethinking Data Distillation: Do Not Overlook Calibration

Neural networks trained on distilled data often produce over-confident o...
research
09/29/2022

Dataset Distillation using Parameter Pruning

The acquisition of advanced models relies on large datasets in many fiel...
research
06/18/2020

Photometric Data-driven Classification of Type Ia Supernovae in the Open Supernova Catalog

We propose a novel approach for a machine-learning-based detection of th...
research
03/08/2023

DiM: Distilling Dataset into Generative Model

Dataset distillation reduces the network training cost by synthesizing s...
research
11/20/2022

Minimizing the Accumulated Trajectory Error to Improve Dataset Distillation

Model-based deep learning has achieved astounding successes due in part ...
research
10/13/2021

Data Incubation – Synthesizing Missing Data for Handwriting Recognition

In this paper, we demonstrate how a generative model can be used to buil...

Please sign up or login with your details

Forgot password? Click here to reset