Regularizing Deep Multi-Task Networks using Orthogonal Gradients

12/14/2019
by   Mihai Suteu, et al.
0

Deep neural networks are a promising approach towards multi-task learning because of their capability to leverage knowledge across domains and learn general purpose representations. Nevertheless, they can fail to live up to these promises as tasks often compete for a model's limited resources, potentially leading to lower overall performance. In this work we tackle the issue of interfering tasks through a comprehensive analysis of their training, derived from looking at the interaction between gradients within their shared parameters. Our empirical results show that well-performing models have low variance in the angles between task gradients and that popular regularization methods implicitly reduce this measure. Based on this observation, we propose a novel gradient regularization term that minimizes task interference by enforcing near orthogonal gradients. Updating the shared parameters using this property encourages task specific decoders to optimize different parts of the feature extractor, thus reducing competition. We evaluate our method with classification and regression tasks on the multiDigitMNIST, NYUv2 and SUN RGB-D datasets where we obtain competitive results.

READ FULL TEXT
research
01/31/2023

GDOD: Effective Gradient Descent using Orthogonal Decomposition for Multi-Task Learning

Multi-task learning (MTL) aims at solving multiple related tasks simulta...
research
05/06/2022

Explaining the Effectiveness of Multi-Task Learning for Efficient Knowledge Extraction from Spine MRI Reports

Pretrained Transformer based models finetuned on domain specific corpora...
research
11/19/2019

Adaptive Activation Network and Functional Regularization for Efficient and Flexible Deep Multi-Task Learning

Multi-task learning (MTL) is a common paradigm that seeks to improve the...
research
01/11/2022

In Defense of the Unitary Scalarization for Deep Multi-Task Learning

Recent multi-task learning research argues against unitary scalarization...
research
04/18/2019

Attentive Single-Tasking of Multiple Tasks

In this work we address task interference in universal networks by consi...
research
05/31/2023

Learning Task-preferred Inference Routes for Gradient De-conflict in Multi-output DNNs

Multi-output deep neural networks(MONs) contain multiple task branches, ...
research
07/10/2023

Gradient Surgery for One-shot Unlearning on Generative Model

Recent regulation on right-to-be-forgotten emerges tons of interest in u...

Please sign up or login with your details

Forgot password? Click here to reset