DeepAI AI Chat
Log In Sign Up

Re-framing Incremental Deep Language Models for Dialogue Processing with Multi-task Learning

by   Morteza Rohanian, et al.

We present a multi-task learning framework to enable the training of one universal incremental dialogue processing model with four tasks of disfluency detection, language modelling, part-of-speech tagging, and utterance segmentation in a simple deep recurrent setting. We show that these tasks provide positive inductive biases to each other with the optimal contribution of each one relying on the severity of the noise from the task. Our live multi-task model outperforms similar individual tasks, delivers competitive performance, and is beneficial for future use in conversational agents in psychiatric treatment.


page 1

page 2

page 3

page 4


Language Modelling as a Multi-Task Problem

In this paper, we propose to study language modelling as a multi-task pr...

SpeechNet: A Universal Modularized Model for Speech Processing Tasks

There is a wide variety of speech processing tasks ranging from extracti...

Multi-Task Learning for Domain-General Spoken Disfluency Detection in Dialogue Systems

Spontaneous spoken dialogue is often disfluent, containing pauses, hesit...

Multi-task learning for Joint Language Understanding and Dialogue State Tracking

This paper presents a novel approach for multi-task learning of language...

Scheduled Multi-task Learning for Neural Chat Translation

Neural Chat Translation (NCT) aims to translate conversational text into...

Modular Universal Reparameterization: Deep Multi-task Learning Across Diverse Domains

As deep learning applications continue to become more diverse, an intere...

CaSiNo: A Corpus of Campsite Negotiation Dialogues for Automatic Negotiation Systems

Automated systems that negotiate with humans have broad applications in ...