Prompting to Distill: Boosting Data-Free Knowledge Distillation via Reinforced Prompt

05/16/2022
by   Xinyin Ma, et al.
0

Data-free knowledge distillation (DFKD) conducts knowledge distillation via eliminating the dependence of original training data, and has recently achieved impressive results in accelerating pre-trained language models. At the heart of DFKD is to reconstruct a synthetic dataset by inverting the parameters of the uncompressed model. Prior DFKD approaches, however, have largely relied on hand-crafted priors of the target data distribution for the reconstruction, which can be inevitably biased and often incompetent to capture the intrinsic distributions. To address this problem, we propose a prompt-based method, termed as PromptDFD, that allows us to take advantage of learned language priors, which effectively harmonizes the synthetic sentences to be semantically and grammatically correct. Specifically, PromptDFD leverages a pre-trained generative model to provide language priors and introduces a reinforced topic prompter to control data synthesis, making the generated samples thematically relevant and semantically plausible, and thus friendly to downstream tasks. As shown in our experiments, the proposed method substantially improves the synthesis quality and achieves considerable improvements on distillation performance. In some cases, PromptDFD even gives rise to results on par with those from the data-driven knowledge distillation with access to the original training data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/04/2023

Revisiting Data-Free Knowledge Distillation with Poisoned Teachers

Data-free knowledge distillation (KD) helps transfer knowledge from a pr...
research
12/12/2021

Up to 100x Faster Data-free Knowledge Distillation

Data-free knowledge distillation (DFKD) has recently been attracting inc...
research
06/03/2023

Deep Classifier Mimicry without Data Access

Access to pre-trained models has recently emerged as a standard across n...
research
06/29/2023

NaturalInversion: Data-Free Image Synthesis Improving Real-World Consistency

We introduce NaturalInversion, a novel model inversion-based method to s...
research
03/30/2023

KD-DLGAN: Data Limited Image Generation via Knowledge Distillation

Generative Adversarial Networks (GANs) rely heavily on large-scale train...
research
05/18/2021

Contrastive Model Inversion for Data-Free Knowledge Distillation

Model inversion, whose goal is to recover training data from a pre-train...
research
09/06/2023

A deep Natural Language Inference predictor without language-specific training data

In this paper we present a technique of NLP to tackle the problem of inf...

Please sign up or login with your details

Forgot password? Click here to reset