DeepAI AI Chat
Log In Sign Up

End-to-End Learning of Flowchart Grounded Task-Oriented Dialogs

by   Dinesh Raghu, et al.
Indian Institute of Technology Delhi

We propose a novel problem within end-to-end learning of task-oriented dialogs (TOD), in which the dialog system mimics a troubleshooting agent who helps a user by diagnosing their problem (e.g., car not starting). Such dialogs are grounded in domain-specific flowcharts, which the agent is supposed to follow during the conversation. Our task exposes novel technical challenges for neural TOD, such as grounding an utterance to the flowchart without explicit annotation, referring to additional manual pages when user asks a clarification question, and ability to follow unseen flowcharts at test time. We release a dataset (FloDial) consisting of 2,738 dialogs grounded on 12 different troubleshooting flowcharts. We also design a neural model, FloNet, which uses a retrieval-augmented generation architecture to train the dialog agent. Our experiments find that FloNet can do zero-shot transfer to unseen flowcharts, and sets a strong baseline for future research.


CookDial: A dataset for task-oriented dialogs grounded in procedural documents

This work presents a new dialog dataset, CookDial, that facilitates rese...

AnyTOD: A Programmable Task-Oriented Dialog System

We propose AnyTOD, an end-to-end task-oriented dialog (TOD) system with ...

End-to-end Conversation Modeling Track in DSTC6

End-to-end training of neural networks is a promising approach to automa...

Zero-Shot Dialog Generation with Cross-Domain Latent Actions

This paper introduces zero-shot dialog generation (ZSDG), as a step towa...

Navigating Connected Memories with a Task-oriented Dialog System

Recent years have seen an increasing trend in the volume of personal med...

Cascaded Span Extraction and Response Generation for Document-Grounded Dialog

This paper summarizes our entries to both subtasks of the first DialDoc ...

SIMMC 2.0: A Task-oriented Dialog Dataset for Immersive Multimodal Conversations

We present a new corpus for the Situated and Interactive Multimodal Conv...