Few-shot Prompting Towards Controllable Response Generation

06/08/2022
by   Hsuan Su, et al.
0

Much literature has shown that prompt-based learning is an efficient method to make use of the large pre-trained language model. Recent works also exhibit the possibility of steering a chatbot's output by plugging in an appropriate prompt. Gradient-based methods are often used to perturb the prompts. However, some language models are not even available to the public. In this work, we first explored the combination of prompting and reinforcement learning (RL) to steer models' generation without accessing any of the models' parameters. Second, to reduce the training effort and enhance the generalizability to the unseen task, we apply multi-task learning to make the model learn to generalize to new tasks better. The experiment results show that our proposed method can successfully control several state-of-the-art (SOTA) dialogue models without accessing their parameters. Furthermore, the model demonstrates the strong ability to quickly adapt to an unseen task in fewer steps than the baseline model.

READ FULL TEXT
research
02/01/2023

Collaborating with language models for embodied reasoning

Reasoning in a complex and ambiguous environment is a key goal for Reinf...
research
09/05/2021

SideControl: Controlled Open-domain Dialogue Generation via Additive Side Networks

Transformer-based pre-trained language models boost the performance of o...
research
02/12/2023

Stabilized In-Context Learning with Pre-trained Language Models for Few Shot Dialogue State Tracking

Prompt-based methods with large pre-trained language models (PLMs) have ...
research
05/04/2023

Task-Optimized Adapters for an End-to-End Task-Oriented Dialogue System

Task-Oriented Dialogue (TOD) systems are designed to carry out specific ...
research
08/14/2023

Dialogue for Prompting: a Policy-Gradient-Based Discrete Prompt Optimization for Few-shot Learning

Prompt-based pre-trained language models (PLMs) paradigm have succeeded ...
research
08/09/2023

Answering Unseen Questions With Smaller Language Models Using Rationale Generation and Dense Retrieval

When provided with sufficient explanatory context, smaller Language Mode...
research
05/17/2023

Using a Large Language Model to Control Speaking Style for Expressive TTS

Appropriate prosody is critical for successful spoken communication. Con...

Please sign up or login with your details

Forgot password? Click here to reset