Conformal Predictor for Improving Zero-shot Text Classification Efficiency

10/23/2022
by   Prafulla Kumar Choubey, et al.
0

Pre-trained language models (PLMs) have been shown effective for zero-shot (0shot) text classification. 0shot models based on natural language inference (NLI) and next sentence prediction (NSP) employ cross-encoder architecture and infer by making a forward pass through the model for each label-text pair separately. This increases the computational cost to make inferences linearly in the number of labels. In this work, we improve the efficiency of such cross-encoder-based 0shot models by restricting the number of likely labels using another fast base classifier-based conformal predictor (CP) calibrated on samples labeled by the 0shot model. Since a CP generates prediction sets with coverage guarantees, it reduces the number of target labels without excluding the most probable label based on the 0shot model. We experiment with three intent and two topic classification datasets. With a suitable CP for each dataset, we reduce the average inference time for NLI- and NSP-based models by 25.6 error rate of 1

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/03/2023

The Benefits of Label-Description Training for Zero-Shot Text Classification

Large language models have improved zero-shot text classification by all...
research
05/25/2023

Label Agnostic Pre-training for Zero-shot Text Classification

Conventional approaches to text classification typically assume the exis...
research
07/24/2023

Leveraging Label Variation in Large Language Models for Zero-Shot Text Classification

The zero-shot learning capabilities of large language models (LLMs) make...
research
09/02/2021

Do Prompt-Based Models Really Understand the Meaning of their Prompts?

Recently, a boom of papers have shown extraordinary progress in few-shot...
research
04/04/2021

Improving Pretrained Models for Zero-shot Multi-label Text Classification through Reinforced Label Hierarchy Reasoning

Exploiting label hierarchies has become a promising approach to tackling...
research
07/06/2020

Relaxed Conformal Prediction Cascades for Efficient Inference Over Many Labels

Providing a small set of promising candidates in place of a single predi...
research
04/24/2023

Generation-driven Contrastive Self-training for Zero-shot Text Classification with Instruction-tuned GPT

Moreover, GPT-based zero-shot classification models tend to make indepen...

Please sign up or login with your details

Forgot password? Click here to reset