A Scale-Invariant Task Balancing Approach for Multi-Task Learning

08/23/2023
by   Baijiong Lin, et al.
0

Multi-task learning (MTL), a learning paradigm to learn multiple related tasks simultaneously, has achieved great success in various fields. However, task-balancing remains a significant challenge in MTL, with the disparity in loss/gradient scales often leading to performance compromises. In this paper, we propose a Scale-Invariant Multi-Task Learning (SI-MTL) method to alleviate the task-balancing problem from both loss and gradient perspectives. Specifically, SI-MTL contains a logarithm transformation which is performed on all task losses to ensure scale-invariant at the loss level, and a gradient balancing method, SI-G, which normalizes all task gradients to the same magnitude as the maximum gradient norm. Extensive experiments conducted on several benchmark datasets consistently demonstrate the effectiveness of SI-G and the state-of-the-art performance of SI-MTL.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/22/2022

Mitigating Negative Transfer in Multi-Task Learning with Exponential Moving Average Loss Weighting Strategies

Multi-Task Learning (MTL) is a growing subject of interest in deep learn...
research
06/20/2021

Heterogeneous Multi-task Learning with Expert Diversity

Predicting multiple heterogeneous biological and medical targets is a ch...
research
06/03/2021

Adjacency List Oriented Relational Fact Extraction via Adaptive Multi-task Learning

Relational fact extraction aims to extract semantic triplets from unstru...
research
05/10/2023

iLab at SemEval-2023 Task 11 Le-Wi-Di: Modelling Disagreement or Modelling Perspectives?

There are two competing approaches for modelling annotator disagreement:...
research
08/02/2022

Curvature-informed multi-task learning for graph networks

Properties of interest for crystals and molecules, such as band gap, ela...
research
09/21/2023

Multi-Task Cooperative Learning via Searching for Flat Minima

Multi-task learning (MTL) has shown great potential in medical image ana...
research
06/25/2020

MTAdam: Automatic Balancing of Multiple Training Loss Terms

When training neural models, it is common to combine multiple loss terms...

Please sign up or login with your details

Forgot password? Click here to reset