Multi-Task Learning in Natural Language Processing: An Overview

09/19/2021
by   Shijie Chen, et al.
0

Deep learning approaches have achieved great success in the field of Natural Language Processing (NLP). However, deep neural models often suffer from overfitting and data scarcity problems that are pervasive in NLP tasks. In recent years, Multi-Task Learning (MTL), which can leverage useful information of related tasks to achieve simultaneous performance improvement on multiple related tasks, has been used to handle these problems. In this paper, we give an overview of the use of MTL in NLP tasks. We first review MTL architectures used in NLP tasks and categorize them into four classes, including the parallel architecture, hierarchical architecture, modular architecture, and generative adversarial architecture. Then we present optimization techniques on loss construction, data sampling, and task scheduling to properly train a multi-task model. After presenting applications of MTL in a variety of NLP tasks, we introduce some benchmark datasets. Finally, we make a conclusion and discuss several possible research directions in this field.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/16/2019

An Empirical Evaluation of Multi-task Learning in Deep Neural Networks for Natural Language Processing

Multi-Task Learning (MTL) aims at boosting the overall performance of ea...
research
11/18/2020

Out-of-Task Training for Dialog State Tracking Models

Dialog state tracking (DST) suffers from severe data sparsity. While man...
research
09/26/2021

Paradigm Shift in Natural Language Processing

In the era of deep learning, modeling for most NLP tasks has converged t...
research
01/03/2022

Robust Natural Language Processing: Recent Advances, Challenges, and Future Directions

Recent natural language processing (NLP) techniques have accomplished hi...
research
07/22/2020

Multi-task learning for natural language processing in the 2020s: where are we going?

Multi-task learning (MTL) significantly pre-dates the deep learning era,...
research
08/16/2023

Challenges and Opportunities of Using Transformer-Based Multi-Task Learning in NLP Through ML Lifecycle: A Survey

The increasing adoption of natural language processing (NLP) models acro...
research
07/30/2023

Text Analysis Using Deep Neural Networks in Digital Humanities and Information Science

Combining computational technologies and humanities is an ongoing effort...

Please sign up or login with your details

Forgot password? Click here to reset