Meta-learning for Few-shot Natural Language Processing: A Survey

07/19/2020
by   Wenpeng Yin, et al.
0

Few-shot natural language processing (NLP) refers to NLP tasks that are accompanied with merely a handful of labeled examples. This is a real-world challenge that an AI system must learn to handle. Usually we rely on collecting more auxiliary information or developing a more efficient learning algorithm. However, the general gradient-based optimization in high capacity models, if training from scratch, requires many parameter-updating steps over a large number of labeled examples to perform well (Snell et al., 2017). If the target task itself cannot provide more information, how about collecting more tasks equipped with rich annotations to help the model learning? The goal of meta-learning is to train a model on a variety of tasks with rich annotations, such that it can solve a new task using only a few labeled samples. The key idea is to train the model's initial parameters such that the model has maximal performance on a new task after the parameters have been updated through zero or a couple of gradient steps. There are already some surveys for meta-learning, such as (Vilalta and Drissi, 2002; Vanschoren, 2018; Hospedales et al., 2020). Nevertheless, this paper focuses on NLP domain, especially few-shot applications. We try to provide clearer definitions, progress summary and some common datasets of applying meta-learning to few-shot NLP.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/03/2022

Meta Learning for Natural Language Processing: A Survey

Deep learning has been the mainstream technique in natural language proc...
research
01/27/2022

Grad2Task: Improved Few-shot Text Classification Using Gradients for Task Representation

Large pretrained language models (LMs) like BERT have improved performan...
research
12/14/2018

Online gradient-based mixtures for transfer modulation in meta-learning

Learning-to-learn or meta-learning leverages data-driven inductive bias ...
research
05/18/2022

Persian Natural Language Inference: A Meta-learning approach

Incorporating information from other languages can improve the results o...
research
02/17/2021

Few-shot Conformal Prediction with Auxiliary Tasks

We develop a novel approach to conformal prediction when the target task...
research
06/01/2023

Effective Structured Prompting by Meta-Learning and Representative Verbalizer

Prompt tuning for pre-trained masked language models (MLM) has shown pro...
research
05/24/2019

MatchZoo: A Learning, Practicing, and Developing System for Neural Text Matching

Text matching is the core problem in many natural language processing (N...

Please sign up or login with your details

Forgot password? Click here to reset