Improving In-Context Few-Shot Learning via Self-Supervised Training

05/03/2022
by   Mingda Chen, et al.
1

Self-supervised pretraining has made few-shot learning possible for many NLP tasks. But the pretraining objectives are not typically adapted specifically for in-context few-shot learning. In this paper, we propose to use self-supervision in an intermediate training stage between pretraining and downstream few-shot usage with the goal to teach the model to perform in-context few shot learning. We propose and evaluate four self-supervised objectives on two benchmarks. We find that the intermediate self-supervision stage produces models that outperform strong baselines. Ablation study shows that several factors affect the downstream performance, such as the amount of training data and the diversity of the self-supervised objectives. Human-annotated cross-task supervision and self-supervision are complementary. Qualitative analysis suggests that the self-supervised-trained models are better at following task requirements.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/17/2019

Boosting Supervision with Self-Supervision for Few-shot Learning

We present a technique to improve the transferability of deep representa...
research
06/07/2023

GPT Self-Supervision for a Better Data Annotator

The task of annotating data into concise summaries poses a significant c...
research
05/22/2022

Self-supervised U-net for few-shot learning of object segmentation in microscopy images

State-of-the-art segmentation performances are achieved by deep neural n...
research
10/14/2020

Function Contrastive Learning of Transferable Representations

Few-shot-learning seeks to find models that are capable of fast-adaptati...
research
05/27/2023

Instance-based Max-margin for Practical Few-shot Recognition

In order to mimic the human few-shot learning (FSL) ability better and t...
research
10/02/2020

Long-Tail Zero and Few-Shot Learning via Contrastive Pretraining on and for Small Data

For natural language processing (NLP) tasks such as sentiment or topic c...
research
05/18/2023

How does the task complexity of masked pretraining objectives affect downstream performance?

Masked language modeling (MLM) is a widely used self-supervised pretrain...

Please sign up or login with your details

Forgot password? Click here to reset