Learning Instructions with Unlabeled Data for Zero-Shot Cross-Task Generalization

10/17/2022
by   Yuxian Gu, et al.
0

Training language models to learn from human instructions for zero-shot cross-task generalization has attracted much attention in NLP communities. Recently, instruction tuning (IT), which fine-tunes a pre-trained language model on a massive collection of tasks described via human-craft instructions, has been shown effective in instruction learning for unseen tasks. However, IT relies on a large amount of human-annotated samples, which restricts its generalization. Unlike labeled data, unlabeled data are often massive and cheap to obtain. In this work, we study how IT can be improved with unlabeled data. We first empirically explore the IT performance trends versus the number of labeled data, instructions, and training tasks. We find it critical to enlarge the number of training instructions, and the instructions can be underutilized due to the scarcity of labeled data. Then, we propose Unlabeled Data Augmented Instruction Tuning (UDIT) to take better advantage of the instructions during IT by constructing pseudo-labeled data from unlabeled plain texts. We conduct extensive experiments to show UDIT's effectiveness in various scenarios of tasks and datasets. We also comprehensively analyze the key factors of UDIT to investigate how to better improve IT with unlabeled data. The code is publicly available at https://github.com/thu-coai/UDIT.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/21/2022

MultiInstruct: Improving Multi-Modal Zero-Shot Learning via Instruction Tuning

Instruction tuning, a new learning paradigm that fine-tunes pre-trained ...
research
05/19/2023

Zero-Shot Text Classification via Self-Supervised Tuning

Existing solutions to zero-shot text classification either conduct promp...
research
11/09/2022

Zero-Label Prompt Selection

Natural language prompts have been shown to facilitate cross-task genera...
research
12/20/2022

Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good movie, and a good prompt too?

Large language models can perform new tasks in a zero-shot fashion, give...
research
06/02/2023

Enhancing CLIP with CLIP: Exploring Pseudolabeling for Limited-Label Prompt Tuning

Fine-tuning vision-language models (VLMs) like CLIP to downstream tasks ...
research
06/16/2023

Differentiable Instruction Optimization for Cross-Task Generalization

Instruction tuning has been attracting much attention to achieve general...
research
08/14/2023

EcomGPT: Instruction-tuning Large Language Model with Chain-of-Task Tasks for E-commerce

Recently, instruction-following Large Language Models (LLMs) , represent...

Please sign up or login with your details

Forgot password? Click here to reset