Learning to Few-Shot Learn Across Diverse Natural Language Classification Tasks

11/10/2019
by   Trapit Bansal, et al.
0

Self-supervised pre-training of transformer models has shown enormous success in improving performance on a number of downstream tasks. However, fine-tuning on a new task still requires large amounts of task-specific labelled data to achieve good performance. We consider this problem of learning to generalize to new tasks with few examples as a meta-learning problem. While meta-learning has shown tremendous progress in recent years, its application is still limited to simulated problems or problems with limited diversity across tasks. We develop a novel method, LEOPARD, which enables optimization-based meta-learning across tasks with different number of classes, and evaluate existing methods on generalization to diverse NLP classification tasks. LEOPARD is trained with the state-of-the-art transformer architecture and shows strong generalization to tasks not seen at all during training, with as few as 8 examples per label. On 16 NLP datasets, across a diverse task-set such as entity typing, relation extraction, natural language inference, sentiment analysis, and several other text categorization tasks, we show that LEOPARD learns better initial parameters for few-shot learning than self-supervised pre-training or multi-task training, outperforming many strong baselines, for example, increasing F1 from 49

READ FULL TEXT
research
09/17/2020

Self-Supervised Meta-Learning for Few-Shot Natural Language Classification Tasks

Self-supervised pre-training of transformer models has revolutionized NL...
research
11/02/2021

Diverse Distributions of Self-Supervised Tasks for Meta-Learning in NLP

Meta-learning considers the problem of learning an efficient learning pr...
research
10/29/2021

MetaICL: Learning to Learn In Context

We introduce MetaICL (Meta-training for In-Context Learning), a new meta...
research
10/22/2022

Meta-learning Pathologies from Radiology Reports using Variance Aware Prototypical Networks

Large pretrained Transformer-based language models like BERT and GPT hav...
research
12/22/2022

Robust Meta-Representation Learning via Global Label Inference and Classification

Few-shot learning (FSL) is a central problem in meta-learning, where lea...
research
05/19/2018

Diverse Few-Shot Text Classification with Multiple Metrics

We study few-shot learning in natural language domains. Compared to many...
research
12/22/2021

Meta-Learning and Self-Supervised Pretraining for Real World Image Translation

Recent advances in deep learning, in particular enabled by hardware adva...

Please sign up or login with your details

Forgot password? Click here to reset