Multi-Task Deep Neural Networks for Natural Language Understanding

01/31/2019
by   Xiaodong Liu, et al.
0

In this paper, we present a Multi-Task Deep Neural Network (MT-DNN) for learning representations across multiple natural language understanding (NLU) tasks. MT-DNN not only leverages large amounts of cross-task data, but also benefits from a regularization effect that leads to more general representations in order to adapt to new tasks and domains. MT-DNN extends the model proposed in Liu et al. (2015) by incorporating a pre-trained bidirectional transformer language model, known as BERT (Devlin et al., 2018). MT-DNN obtains new state-of-the-art results on ten NLU tasks, including SNLI, SciTail, and eight out of nine GLUE tasks, pushing the GLUE benchmark to 82.2 (1.8 datasets that the representations learned by MT-DNN allow domain adaptation with substantially fewer in-domain labels than the pre-trained BERT representations. Our code and pre-trained models will be made publicly available.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/20/2019

Improving Multi-Task Deep Neural Networks via Knowledge Distillation for Natural Language Understanding

This paper explores the use of knowledge distillation to improve a Multi...
research
01/11/2020

Exploring and Improving Robustness of Multi Task Deep Neural Networks via Domain Agnostic Defenses

In this paper, we explore the robustness of the Multi-Task Deep Neural N...
research
10/31/2019

LIMIT-BERT : Linguistic Informed Multi-Task BERT

In this paper, we present a Linguistic Informed Multi-Task BERT (LIMIT-B...
research
10/25/2022

Evaluating Parameter Efficient Learning for Generation

Parameter efficient learning methods (PERMs) have recently gained signif...
research
03/31/2017

Comparison of multi-task convolutional neural network (MT-CNN) and a few other methods for toxicity prediction

Toxicity analysis and prediction are of paramount importance to human he...
research
02/19/2020

The Microsoft Toolkit of Multi-Task Deep Neural Networks for Natural Language Understanding

We present MT-DNN, an open-source natural language understanding (NLU) t...
research
05/16/2019

Latent Universal Task-Specific BERT

This paper describes a language representation model which combines the ...

Please sign up or login with your details

Forgot password? Click here to reset