Avoiding Inference Heuristics in Few-shot Prompt-based Finetuning

09/09/2021
by   Prasetya Ajie Utama, et al.
16

Recent prompt-based approaches allow pretrained language models to achieve strong performances on few-shot finetuning by reformulating downstream tasks as a language modeling problem. In this work, we demonstrate that, despite its advantages on low data regimes, finetuned prompt-based models for sentence pair classification tasks still suffer from a common pitfall of adopting inference heuristics based on lexical overlap, e.g., models incorrectly assuming a sentence pair is of the same meaning because they consist of the same set of words. Interestingly, we find that this particular inference heuristic is significantly less present in the zero-shot evaluation of the prompt-based model, indicating how finetuning can be destructive to useful knowledge learned during the pretraining. We then show that adding a regularization that preserves pretraining weights is effective in mitigating this destructive tendency of few-shot finetuning. Our evaluation on three datasets demonstrates promising improvements on the three corresponding challenge datasets used to diagnose the inference heuristics.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/08/2023

Zero-Shot Robustification of Zero-Shot Models With Foundation Models

Zero-shot inference is a powerful paradigm that enables the use of large...
research
10/07/2020

A Mathematical Exploration of Why Language Models Help Solve Downstream Tasks

Autoregressive language models pretrained on large corpora have been suc...
research
04/07/2020

Byte Pair Encoding is Suboptimal for Language Model Pretraining

The success of pretrained transformer language models in natural languag...
research
10/23/2020

Multilingual BERT Post-Pretraining Alignment

We propose a simple method to align multilingual contextual embeddings a...
research
07/15/2021

FLEX: Unifying Evaluation for Few-Shot NLP

Few-shot NLP research is highly active, yet conducted in disjoint resear...
research
10/23/2022

Lexical Generalization Improves with Larger Models and Longer Training

While fine-tuned language models perform well on many tasks, they were a...
research
02/04/2019

Right for the Wrong Reasons: Diagnosing Syntactic Heuristics in Natural Language Inference

Machine learning systems can often achieve high performance on a test se...

Please sign up or login with your details

Forgot password? Click here to reset