Text Classification via Large Language Models

05/15/2023
by   Xiaofei Sun, et al.
0

Despite the remarkable success of large-scale Language Models (LLMs) such as GPT-3, their performances still significantly underperform fine-tuned models in the task of text classification. This is due to (1) the lack of reasoning ability in addressing complex linguistic phenomena (e.g., intensification, contrast, irony etc); (2) limited number of tokens allowed in in-context learning. In this paper, we introduce Clue And Reasoning Prompting (CARP). CARP adopts a progressive reasoning strategy tailored to addressing the complex linguistic phenomena involved in text classification: CARP first prompts LLMs to find superficial clues (e.g., keywords, tones, semantic relations, references, etc), based on which a diagnostic reasoning process is induced for final decisions. To further address the limited-token issue, CARP uses a fine-tuned model on the supervised dataset for kNN demonstration search in the in-context learning, allowing the model to take the advantage of both LLM's generalization ability and the task-specific evidence provided by the full labeled dataset. Remarkably, CARP yields new SOTA performances on 4 out of 5 widely-used text-classification benchmarks, 97.39 (+1.24) on SST-2, 96.40 (+0.72) on AGNews, 98.78 (+0.25) on R8 and 96.95 (+0.6) on R52, and a performance comparable to SOTA on MR (92.39 v.s. 93.3). More importantly, we find that CARP delivers impressive abilities on low-resource and domain-adaptation setups. Specifically, Specifically, using 16 examples per class, CARP achieves comparable performances to supervised models with 1,024 examples per class.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/03/2023

GPT-RE: In-context Learning for Relation Extraction using Large Language Models

In spite of the potential for ground-breaking achievements offered by la...
research
05/23/2022

Many-Class Text Classification with Matching

In this work, we formulate Text Classification as a Matching problem bet...
research
05/24/2023

EXnet: Efficient In-context Learning for Data-less Text classification

Large pre-trained language models (PLMs) have made significant progress ...
research
11/03/2022

PINTO: Faithful Language Reasoning Using Prompt-Generated Rationales

Neural language models (LMs) have achieved impressive results on various...
research
11/20/2022

Artificial Interrogation for Attributing Language Models

This paper presents solutions to the Machine Learning Model Attribution ...
research
06/28/2020

Progressive Generation of Long Text

Large-scale language models pretrained on massive corpora of text, such ...
research
05/11/2022

Building for Tomorrow: Assessing the Temporal Persistence of Text Classifiers

Where performance of text classification models drops over time due to c...

Please sign up or login with your details

Forgot password? Click here to reset