Improving Few-shot Text Classification via Pretrained Language Representations

08/22/2019
by   Ningyu Zhang, et al.
0

Text classification tends to be difficult when the data is deficient or when it is required to adapt to unseen classes. In such challenging scenarios, recent studies have often used meta-learning to simulate the few-shot task, thus negating explicit common linguistic features across tasks. Deep language representations have proven to be very effective forms of unsupervised pretraining, yielding contextualized features that capture linguistic properties and benefit downstream natural language understanding tasks. However, the effect of pretrained language representation for few-shot learning on text classification tasks is still not well understood. In this study, we design a few-shot learning model with pretrained language representations and report the empirical results. We show that our approach is not only simple but also produces state-of-the-art performance on a well-studied sentiment classification dataset. It can thus be further suggested that pretraining could be a promising solution for few shot learning of many other NLP tasks. The code and the dataset to replicate the experiments are made available at https://github.com/zxlzr/FewShotNLP.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/16/2022

Zero-Shot Learners for Natural Language Understanding via a Unified Multiple Choice Perspective

We propose a new paradigm for zero-shot learners that is format agnostic...
research
02/27/2019

Few-Shot Text Classification with Induction Network

Text classification tends to struggle when data is deficient or when it ...
research
05/21/2023

Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers

This paper explores the effectiveness of model-generated signals in impr...
research
05/31/2023

Analyzing Text Representations by Measuring Task Alignment

Textual representations based on pre-trained language models are key, es...
research
08/18/2021

AdapterHub Playground: Simple and Flexible Few-Shot Learning with Adapters

The open-access dissemination of pretrained language models through onli...
research
02/03/2023

Towards Few-Shot Identification of Morality Frames using In-Context Learning

Data scarcity is a common problem in NLP, especially when the annotation...
research
11/05/2018

Evolutionary Data Measures: Understanding the Difficulty of Text Classification Tasks

Classification tasks are usually analysed and improved through new model...

Please sign up or login with your details

Forgot password? Click here to reset