Metric-Based In-context Learning: A Case Study in Text Simplification

07/27/2023
by   Subha Vadlamannati, et al.
0

In-context learning (ICL) for large language models has proven to be a powerful approach for many natural language processing tasks. However, determining the best method to select examples for ICL is nontrivial as the results can vary greatly depending on the quality, quantity, and order of examples used. In this paper, we conduct a case study on text simplification (TS) to investigate how to select the best and most robust examples for ICL. We propose Metric-Based in-context Learning (MBL) method that utilizes commonly used TS metrics such as SARI, compression ratio, and BERT-Precision for selection. Through an extensive set of experiments with various-sized GPT models on standard TS benchmarks such as TurkCorpus and ASSET, we show that examples selected by the top SARI scores perform the best on larger models such as GPT-175B, while the compression ratio generally performs better on smaller models such as GPT-13B and GPT-6.7B. Furthermore, we demonstrate that MBL is generally robust to example orderings and out-of-domain test sets, and outperforms strong baselines and state-of-the-art finetuned language models. Finally, we show that the behaviour of large GPT models can be implicitly controlled by the chosen metric. Our research provides a new framework for selecting examples in ICL, and demonstrates its effectiveness in text simplification tasks, breaking new ground for more accurate and efficient NLG systems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/08/2022

Active Example Selection for In-Context Learning

With a handful of demonstration examples, large-scale language models sh...
research
05/23/2023

RetICL: Sequential Retrieval of In-Context Examples with Reinforcement Learning

Many recent developments in large language models focus on prompting the...
research
02/21/2023

In-context Example Selection with Influences

In-context learning (ICL) is a powerful paradigm emerged from large lang...
research
05/23/2023

In-context Example Selection for Machine Translation Using Multiple Features

Large language models have demonstrated the capability to perform well o...
research
09/19/2023

Language Modeling Is Compression

It has long been established that predictive models can be transformed i...
research
01/17/2022

Evaluation of HTR models without Ground Truth Material

The evaluation of Handwritten Text Recognition (HTR) models during their...
research
10/22/2022

Meta-learning Pathologies from Radiology Reports using Variance Aware Prototypical Networks

Large pretrained Transformer-based language models like BERT and GPT hav...

Please sign up or login with your details

Forgot password? Click here to reset