A Comprehensive Survey to Dataset Distillation

01/13/2023
by   Shiye Lei, et al.
0

Deep learning technology has unprecedentedly developed in the last decade and has become the primary choice in many application domains. This progress is mainly attributed to a systematic collaboration that rapidly growing computing resources encourage advanced algorithms to deal with massive data. However, it gradually becomes challenging to cope with the unlimited growth of data with limited computing power. To this end, diverse approaches are proposed to improve data processing efficiency. Dataset distillation, one of the dataset reduction methods, tackles the problem via synthesising a small typical dataset from giant data and has attracted a lot of attention from the deep learning community. Existing dataset distillation methods can be taxonomised into meta-learning and data match framework according to whether explicitly mimic target data. Albeit dataset distillation has shown a surprising performance in compressing datasets, it still possesses several limitations such as distilling high-resolution data. This paper provides a holistic understanding of dataset distillation from multiple aspects, including distillation frameworks and algorithms, disentangled dataset distillation, performance comparison, and applications. Finally, we discuss challenges and promising directions to further promote future studies about dataset distillation.

READ FULL TEXT

page 3

page 4

research
05/03/2023

A Survey on Dataset Distillation: Approaches, Applications and Future Directions

Dataset distillation is attracting more attention in machine learning as...
research
09/29/2022

Dataset Distillation using Parameter Pruning

The acquisition of advanced models relies on large datasets in many fiel...
research
05/28/2023

Distill Gold from Massive Ores: Efficient Dataset Distillation via Critical Samples Selection

Data-efficient learning has drawn significant attention, especially give...
research
06/19/2023

Categories of Response-Based, Feature-Based, and Relation-Based Knowledge Distillation

Deep neural networks have achieved remarkable performance for artificial...
research
01/17/2023

Dataset Distillation: A Comprehensive Review

Recent success of deep learning can be largely attributed to the huge am...
research
01/11/2023

Data Distillation: A Survey

The popularity of deep learning has led to the curation of a vast number...
research
06/15/2020

Flexible Dataset Distillation: Learn Labels Instead of Images

We study the problem of dataset distillation - creating a small set of s...

Please sign up or login with your details

Forgot password? Click here to reset