Flexible Dataset Distillation: Learn Labels Instead of Images

06/15/2020
by   Ondrej Bohdal, et al.
0

We study the problem of dataset distillation - creating a small set of synthetic examples capable of training a good model. In particular, we study the problem of label distillation - creating synthetic labels for a small set of real images, and show it to be more effective than the prior image-based approach to dataset distillation. Interestingly, label distillation can be applied across datasets, for example enabling learning Japanese character recognition by training only on synthetically labeled English letters. Methodologically, we introduce a more robust and flexible meta-learning algorithm for distillation, as well as an effective first-order strategy based on convex optimization layers. Distilling labels with our new algorithm leads to improved results over prior image-based distillation. More importantly, it leads to clear improvements in flexibility of the distilled dataset in terms of compatibility with off-the-shelf optimizers and diverse neural architectures.

READ FULL TEXT

page 11

page 18

research
10/06/2019

Improving Dataset Distillation

Dataset distillation is a method for reducing dataset sizes: the goal is...
research
06/06/2022

Remember the Past: Distilling Datasets into Addressable Memories for Neural Networks

We propose an algorithm that compresses the critical information of a la...
research
05/02/2023

Generalizing Dataset Distillation via Deep Generative Prior

Dataset Distillation aims to distill an entire dataset's knowledge into ...
research
11/27/2018

Dataset Distillation

Model distillation aims to distill the knowledge of a complex model into...
research
09/29/2022

Dataset Distillation using Parameter Pruning

The acquisition of advanced models relies on large datasets in many fiel...
research
01/13/2023

A Comprehensive Survey to Dataset Distillation

Deep learning technology has unprecedentedly developed in the last decad...
research
11/20/2022

Minimizing the Accumulated Trajectory Error to Improve Dataset Distillation

Model-based deep learning has achieved astounding successes due in part ...

Please sign up or login with your details

Forgot password? Click here to reset