When Does Aggregating Multiple Skills with Multi-Task Learning Work? A Case Study in Financial NLP

05/23/2023
by   Jingwei Ni, et al.
0

Multi-task learning (MTL) aims at achieving a better model by leveraging data and knowledge from multiple tasks. However, MTL does not always work – sometimes negative transfer occurs between tasks, especially when aggregating loosely related skills, leaving it an open question when MTL works. Previous studies show that MTL performance can be improved by algorithmic tricks. However, what tasks and skills should be included is less well explored. In this work, we conduct a case study in Financial NLP where multiple datasets exist for skills relevant to the domain, such as numeric reasoning and sentiment analysis. Due to the task difficulty and data scarcity in the Financial NLP domain, we explore when aggregating such diverse skills from multiple datasets with MTL can work. Our findings suggest that the key to MTL success lies in skill diversity, relatedness between tasks, and choice of aggregation size and shared capacity. Specifically, MTL works well when tasks are diverse but related, and when the size of the task aggregation and the shared capacity of the model are balanced to avoid overwhelming certain tasks.

READ FULL TEXT
research
01/30/2023

ForkMerge: Overcoming Negative Transfer in Multi-Task Learning

The goal of multi-task learning is to utilize useful knowledge from mult...
research
12/13/2022

Do Text-to-Text Multi-Task Learners Suffer from Task Conflict?

Traditional multi-task learning architectures train a single model acros...
research
09/27/2018

Multi-task Learning for Financial Forecasting

Financial forecasting is challenging and attractive in machine learning....
research
02/14/2022

ASC me to Do Anything: Multi-task Training for Embodied AI

Embodied AI has seen steady progress across a diverse set of independent...
research
05/25/2022

Eliciting Transferability in Multi-task Learning with Task-level Mixture-of-Experts

Recent work suggests that transformer models are capable of multi-task l...
research
05/28/2021

Efficient and robust multi-task learning in the brain with modular task primitives

In a real-world setting biological agents do not have infinite resources...
research
11/15/2022

Teaching Algorithmic Reasoning via In-context Learning

Large language models (LLMs) have shown increasing in-context learning c...

Please sign up or login with your details

Forgot password? Click here to reset