Metacognitive Prompting Improves Understanding in Large Language Models

08/10/2023
by   Yuqing Wang, et al.
0

In Large Language Models (LLMs), there have been consistent advancements in task-specific performance, largely influenced by effective prompt design. While recent research on prompting has enhanced the reasoning capabilities of LLMs, a gap remains in further improving their understanding abilities. In this study, we introduce Metacognitive Prompting (MP), a strategy inspired by human introspective reasoning processes. Using MP, LLMs undergo a systematic series of structured, self-aware evaluations, drawing on both their vast inherent knowledge and new insights. Our experiments involve five prevalent LLMs: Llama2, Vicuna, PaLM, GPT-3.5, and GPT-4, all of which span various general natural language understanding (NLU) tasks from the GLUE and SuperGLUE benchmarks. Results indicate that, although GPT-4 consistently excels in most tasks, PaLM, when equipped with MP, approaches its performance level. Furthermore, across models and datasets, MP consistently outperforms existing prompting methods, including standard and chain-of-thought prompting. This study underscores the potential to amplify the understanding abilities of LLMs and highlights the benefits of mirroring human introspective reasoning in NLU tasks.

READ FULL TEXT

page 3

page 7

research
08/29/2023

Large Language Models on the Chessboard: A Study on ChatGPT's Formal Language Comprehension and Complex Reasoning Skills

While large language models have made strides in natural language proces...
research
09/12/2023

Re-Reading Improves Reasoning in Language Models

Reasoning presents a significant and challenging issue for Large Languag...
research
06/21/2023

Understanding Social Reasoning in Language Models with Language Models

As Large Language Models (LLMs) become increasingly integrated into our ...
research
09/16/2022

Psychologically-informed chain-of-thought prompts for metaphor understanding in large language models

Probabilistic models of language understanding are interpretable and str...
research
05/23/2023

Let's Think Frame by Frame: Evaluating Video Chain of Thought with Video Infilling and Prediction

Despite constituting 65 underrepresented in generative AI research. Mean...
research
07/31/2023

Deception Abilities Emerged in Large Language Models

Large language models (LLMs) are currently at the forefront of intertwin...
research
08/29/2023

Where Would I Go Next? Large Language Models as Human Mobility Predictors

Accurate human mobility prediction underpins many important applications...

Please sign up or login with your details

Forgot password? Click here to reset