Data Distillation for Text Classification

04/17/2021
by   Yongqi Li, et al.
13

Deep learning techniques have achieved great success in many fields, while at the same time deep learning models are getting more complex and expensive to compute. It severely hinders the wide applications of these models. In order to alleviate this problem, model distillation emerges as an effective means to compress a large model into a smaller one without a significant drop in accuracy. In this paper, we study a related but orthogonal issue, data distillation, which aims to distill the knowledge from a large training dataset down to a smaller and synthetic one. It has the potential to address the large and growing neural network training problem based on the small dataset. We develop a novel data distillation method for text classification. We evaluate our method on eight benchmark datasets. The results that the distilled data with the size of 0.1 performance of the original is rather impressive.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/27/2018

Dataset Distillation

Model distillation aims to distill the knowledge of a complex model into...
research
07/16/2023

Dataset Distillation Meets Provable Subset Selection

Deep learning has grown tremendously over recent years, yielding state-o...
research
08/15/2023

Multimodal Dataset Distillation for Image-Text Retrieval

Dataset distillation methods offer the promise of reducing a large-scale...
research
06/13/2019

Linear Distillation Learning

Deep Linear Networks do not have expressive power but they are mathemati...
research
05/28/2019

Adaptive Region Embedding for Text Classification

Deep learning models such as convolutional neural networks and recurrent...
research
07/27/2021

Dataset Distillation with Infinitely Wide Convolutional Networks

The effectiveness of machine learning algorithms arises from being able ...
research
04/10/2022

Reducing Model Jitter: Stable Re-training of Semantic Parsers in Production Environments

Retraining modern deep learning systems can lead to variations in model ...

Please sign up or login with your details

Forgot password? Click here to reset