Navigating Prompt Complexity for Zero-Shot Classification: A Study of Large Language Models in Computational Social Science

05/23/2023
by   Yida Mu, et al.
0

Instruction-tuned Large Language Models (LLMs) have exhibited impressive language understanding and the capacity to generate responses that follow specific instructions. However, due to the computational demands associated with training these models, their applications often rely on zero-shot settings. In this paper, we evaluate the zero-shot performance of two publicly accessible LLMs, ChatGPT and OpenAssistant, in the context of Computational Social Science classification tasks, while also investigating the effects of various prompting strategies. Our experiment considers the impact of prompt complexity, including the effect of incorporating label definitions into the prompt, using synonyms for label names, and the influence of integrating past memories during the foundation model training. The findings indicate that in a zero-shot setting, the current LLMs are unable to match the performance of smaller, fine-tuned baseline transformer models (such as BERT). Additionally, we find that different prompting strategies can significantly affect classification accuracy, with variations in accuracy and F1 scores exceeding 10

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/26/2023

Zero is Not Hero Yet: Benchmarking Zero-Shot Performance of LLMs for Financial Tasks

Recently large language models (LLMs) like ChatGPT have shown impressive...
research
10/03/2022

Hypothesis Engineering for Zero-Shot Hate Speech Detection

Standard approaches to hate speech detection rely on sufficient availabl...
research
05/24/2023

EXnet: Efficient In-context Learning for Data-less Text classification

Large pre-trained language models (PLMs) have made significant progress ...
research
05/19/2023

Do Models Really Learn to Follow Instructions? An Empirical Study of Instruction Tuning

Recent works on instruction tuning (IT) have achieved great performance ...
research
10/17/2022

Zero-Shot Ranking Socio-Political Texts with Transformer Language Models to Reduce Close Reading Time

We approach the classification problem as an entailment problem and appl...
research
05/18/2023

A Generalist Dynamics Model for Control

We investigate the use of transformer sequence models as dynamics models...
research
07/19/2023

Can Instruction Fine-Tuned Language Models Identify Social Bias through Prompting?

As the breadth and depth of language model applications continue to expa...

Please sign up or login with your details

Forgot password? Click here to reset