Prompt Engineering and Calibration for Zero-Shot Commonsense Reasoning

04/14/2023
by   Chenkai Ma, et al.
0

Prompt engineering and calibration make large language models excel at reasoning tasks, including multiple choice commonsense reasoning. From a practical perspective, we investigate and evaluate these strategies on smaller language models. Through experiments on five commonsense reasoning benchmarks, we find that each strategy favors certain models, but their joint effects are mostly negative.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/10/2021

Towards Zero-shot Commonsense Reasoning with Self-supervised Refinement of Language Models

Can we get existing language models and refine them for zero-shot common...
research
10/31/2021

A Systematic Investigation of Commonsense Understanding in Large Language Models

Large language models have shown impressive performance on many natural ...
research
04/22/2023

Dialectical language model evaluation: An initial appraisal of the commonsense spatial reasoning abilities of LLMs

Language models have become very popular recently and many claims have b...
research
06/16/2022

PInKS: Preconditioned Commonsense Inference with Minimal Supervision

Reasoning with preconditions such as "glass can be used for drinking wat...
research
11/18/2020

Do Fine-tuned Commonsense Language Models Really Generalize?

Recently, transformer-based methods such as RoBERTa and GPT-3 have led t...
research
06/04/2023

Probing Physical Reasoning with Counter-Commonsense Context

In this study, we create a CConS (Counter-commonsense Contextual Size co...
research
01/31/2022

Causal Inference Principles for Reasoning about Commonsense Causality

Commonsense causality reasoning (CCR) aims at identifying plausible caus...

Please sign up or login with your details

Forgot password? Click here to reset