Stabilized In-Context Learning with Pre-trained Language Models for Few Shot Dialogue State Tracking

02/12/2023
by   Derek Chen, et al.
0

Prompt-based methods with large pre-trained language models (PLMs) have shown impressive unaided performance across many NLP tasks. These models improve even further with the addition of a few labeled in-context exemplars to guide output generation. However, for more complex tasks such as dialogue state tracking (DST), designing prompts that reliably convey the desired intent is nontrivial, leading to unstable results. Furthermore, building in-context exemplars for dialogue tasks is difficult because conversational contexts are long while model input lengths are relatively short. To overcome these issues we first adapt a meta-learning scheme to the dialogue domain which stabilizes the ability of the model to perform well under various prompts. We additionally design a novel training method to improve upon vanilla retrieval mechanisms to find ideal in-context examples. Finally, we introduce a saliency model to limit dialogue text length, allowing us to include more exemplars per query. In effect, we are able to achieve highly competitive results for few-shot DST on MultiWOZ.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/16/2022

In-Context Learning for Few-Shot Dialogue State Tracking

Collecting and annotating task-oriented dialogues is time-consuming and ...
research
04/18/2022

A Study on Prompt-based Few-Shot Learning Methods for Belief State Tracking in Task-oriented Dialog Systems

We tackle the Dialogue Belief State Tracking(DST) problem of task-orient...
research
08/14/2023

Dialogue for Prompting: a Policy-Gradient-Based Discrete Prompt Optimization for Few-shot Learning

Prompt-based pre-trained language models (PLMs) paradigm have succeeded ...
research
04/22/2021

A Short Survey of Pre-trained Language Models for Conversational AI-A NewAge in NLP

Building a dialogue system that can communicate naturally with humans is...
research
06/08/2022

Few-shot Prompting Towards Controllable Response Generation

Much literature has shown that prompt-based learning is an efficient met...
research
10/12/2022

SUMBot: Summarizing Context in Open-Domain Dialogue Systems

In this paper, we investigate the problem of including relevant informat...
research
04/24/2023

ChatLLM Network: More brains, More intelligence

Dialogue-based language models mark a huge milestone in the field of art...

Please sign up or login with your details

Forgot password? Click here to reset