Reparameterizing Convolutions for Incremental Multi-Task Learning without Task Interference

07/24/2020
by   Menelaos Kanakis, et al.
14

Multi-task networks are commonly utilized to alleviate the need for a large number of highly specialized single-task networks. However, two common challenges in developing multi-task models are often overlooked in literature. First, enabling the model to be inherently incremental, continuously incorporating information from new tasks without forgetting the previously learned ones (incremental learning). Second, eliminating adverse interactions amongst tasks, which has been shown to significantly degrade the single-task performance in a multi-task setup (task interference). In this paper, we show that both can be achieved simply by reparameterizing the convolutions of standard neural network architectures into a non-trainable shared part (filter bank) and task-specific parts (modulators), where each modulator has a fraction of the filter bank parameters. Thus, our reparameterization enables the model to learn new tasks without adversely affecting the performance of existing ones. The results of our ablation study attest the efficacy of the proposed reparameterization. Moreover, our method achieves state-of-the-art on two challenging multi-task learning benchmarks, PASCAL-Context and NYUD, and also demonstrates superior incremental learning capability as compared to its close competitors.

READ FULL TEXT
research
04/16/2022

Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners

Traditional multi-task learning (MTL) methods use dense networks that us...
research
08/03/2023

Mitigating Task Interference in Multi-Task Learning via Explicit Task Routing with Non-Learnable Primitives

Multi-task learning (MTL) seeks to learn a single model to accomplish mu...
research
04/18/2019

Attentive Single-Tasking of Multiple Tasks

In this work we address task interference in universal networks by consi...
research
06/22/2018

Continuous Learning in Single-Incremental-Task Scenarios

It was recently shown that architectural, regularization and rehearsal s...
research
06/04/2023

Top-Down Processing: Top-Down Network Combines Back-Propagation with Attention

Early neural network models relied exclusively on bottom-up processing g...
research
04/28/2021

Exploring Relational Context for Multi-Task Dense Prediction

The timeline of computer vision research is marked with advances in lear...
research
04/06/2020

A Generalized Multi-Task Learning Approach to Stereo DSM Filtering in Urban Areas

City models and height maps of urban areas serve as a valuable data sour...

Please sign up or login with your details

Forgot password? Click here to reset