Dialogue Transformers

10/01/2019
by   Vladimir Vlasov, et al.
0

We introduce a dialogue policy based on a transformer architecture, where the self-attention mechanism operates over the sequence of dialogue turns. Recent work has used hierarchical recurrent neural networks to encode multiple utterances in a dialogue context, but we argue that a pure self-attention mechanism is more suitable. By default, an RNN assumes that every item in a sequence is relevant for producing an encoding of the full sequence, but a single conversation can consist of multiple overlapping discourse segments as speakers interleave multiple topics. A transformer picks which turns to include in its encoding of the current dialogue state, and is naturally suited to selectively ignoring or attending to dialogue history. We compare the performance of the Transformer Embedding Dialogue (TED) policy to an LSTM and to the REDP, which was specifically designed to overcome this limitation of RNNs. We show that the TED policy's behaviour compares favourably, both in terms of accuracy and speed.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/16/2019

Memory-Augmented Recurrent Networks for Dialogue Coherence

Recent dialogue approaches operate by reading each word in a conversatio...
research
04/04/2019

Dialogue Act Classification with Context-Aware Self-Attention

Recent work in Dialogue Act classification has treated the task as a seq...
research
06/08/2021

Staircase Attention for Recurrent Processing of Sequences

Attention mechanisms have become a standard tool for sequence modeling t...
research
08/24/2020

End to End Dialogue Transformer

Dialogue systems attempt to facilitate conversations between humans and ...
research
03/21/2019

Learning Multi-Level Information for Dialogue Response Selection by Highway Recurrent Transformer

With the increasing research interest in dialogue response generation, t...
research
10/27/2022

Working Alliance Transformer for Psychotherapy Dialogue Classification

As a predictive measure of the treatment outcome in psychotherapy, the w...
research
11/21/2016

Coherent Dialogue with Attention-based Language Models

We model coherent conversation continuation via RNN-based dialogue model...

Please sign up or login with your details

Forgot password? Click here to reset