ZeroGen: Efficient Zero-shot Learning via Dataset Generation

02/16/2022
by   Jiacheng Ye, et al.
18

There is a growing interest in dataset generation recently due to the superior generative capacity of large pre-trained language models (PLMs). In this paper, we study a flexible and efficient zero-short learning method, ZeroGen. Given a zero-shot task, we first generate a dataset from scratch using PLMs in an unsupervised manner. Then, we train a tiny task model (e.g., LSTM) under the supervision of the synthesized dataset. This approach allows highly efficient inference as the final task model only has orders of magnitude fewer parameters comparing to PLMs (e.g., GPT2-XL). Apart from being annotation-free and efficient, we argue that ZeroGen can also provide useful insights from the perspective of data-free model-agnostic knowledge distillation, and unreferenced text generation evaluation. Experiments and analysis on different NLP tasks, namely, text classification, question answering, and natural language inference), show the effectiveness of ZeroGen.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/22/2022

ProGen: Progressive Zero-shot Dataset Generation via In-context Feedback

Recently, dataset-generation-based zero-shot learning has shown promisin...
research
11/27/2022

Understanding BLOOM: An empirical study on diverse NLP tasks

In this work, we present an evaluation of smaller BLOOM model variants (...
research
09/18/2023

Fabricator: An Open Source Toolkit for Generating Labeled Training Data with Teacher LLMs

Most NLP tasks are modeled as supervised learning and thus require label...
research
05/25/2022

ZeroGen^+: Self-Guided High-Quality Data Generation in Efficient Zero-Shot Learning

Nowadays, owing to the superior capacity of the large pre-trained langua...
research
08/14/2023

Approximating Human-Like Few-shot Learning with GPT-based Compression

In this work, we conceptualize the learning process as information compr...
research
03/28/2023

On Codex Prompt Engineering for OCL Generation: An Empirical Study

The Object Constraint Language (OCL) is a declarative language that adds...
research
04/17/2023

Testing the Reliability of ChatGPT for Text Annotation and Classification: A Cautionary Remark

Recent studies have demonstrated promising potential of ChatGPT for vari...

Please sign up or login with your details

Forgot password? Click here to reset