Don't Prompt, Search! Mining-based Zero-Shot Learning with Language Models

10/26/2022
by   Mozes van de Kar, et al.
0

Masked language models like BERT can perform text classification in a zero-shot fashion by reformulating downstream tasks as text infilling. However, this approach is highly sensitive to the template used to prompt the model, yet practitioners are blind when designing them in strict zero-shot settings. In this paper, we propose an alternative mining-based approach for zero-shot learning. Instead of prompting language models, we use regular expressions to mine labeled examples from unlabeled corpora, which can optionally be filtered through prompting, and used to finetune a pretrained model. Our method is more flexible and interpretable than prompting, and outperforms it on a wide range of tasks when using comparable templates. Our results suggest that the success of prompting can partly be explained by the model being exposed to similar examples during pretraining, which can be directly retrieved through regular expressions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/23/2021

Detecting Hate Speech with GPT-3

Sophisticated language models such as OpenAI's GPT-3 can generate hatefu...
research
02/12/2022

Semantic-Oriented Unlabeled Priming for Large-Scale Language Models

Due to the high costs associated with finetuning large language models, ...
research
12/08/2022

Demystifying Prompts in Language Models via Perplexity Estimation

Language models can be prompted to perform a wide variety of zero- and f...
research
05/24/2023

EXnet: Efficient In-context Learning for Data-less Text classification

Large pre-trained language models (PLMs) have made significant progress ...
research
05/20/2022

Prototypical Calibration for Few-shot Learning of Language Models

In-context learning of GPT-like models has been recognized as fragile ac...
research
03/21/2023

Large Language Models Can Be Used to Scale the Ideologies of Politicians in a Zero-Shot Learning Setting

The aggregation of knowledge embedded in large language models (LLMs) ho...
research
01/14/2022

Eliciting Knowledge from Pretrained Language Models for Prototypical Prompt Verbalizer

Recent advances on prompt-tuning cast few-shot classification tasks as a...

Please sign up or login with your details

Forgot password? Click here to reset