TaskExpert: Dynamically Assembling Multi-Task Representations with Memorial Mixture-of-Experts

07/28/2023
by   Hanrong Ye, et al.
0

Learning discriminative task-specific features simultaneously for multiple distinct tasks is a fundamental problem in multi-task learning. Recent state-of-the-art models consider directly decoding task-specific features from one shared task-generic feature (e.g., feature from a backbone layer), and utilize carefully designed decoders to produce multi-task features. However, as the input feature is fully shared and each task decoder also shares decoding parameters for different input samples, it leads to a static feature decoding process, producing less discriminative task-specific representations. To tackle this limitation, we propose TaskExpert, a novel multi-task mixture-of-experts model that enables learning multiple representative task-generic feature spaces and decoding task-specific features in a dynamic manner. Specifically, TaskExpert introduces a set of expert networks to decompose the backbone feature into several representative task-generic features. Then, the task-specific features are decoded by using dynamic task-specific gating networks operating on the decomposed task-generic features. Furthermore, to establish long-range modeling of the task-specific representations from different layers of TaskExpert, we design a multi-task feature memory that updates at each layer and acts as an additional feature expert for dynamic task-specific feature decoding. Extensive experiments demonstrate that our TaskExpert clearly outperforms previous best-performing methods on all 9 metrics of two competitive multi-task learning benchmarks for visual scene understanding (i.e., PASCAL-Context and NYUD-v2). Codes and models will be made publicly available at https://github.com/prismformore/Multi-Task-Transformer

READ FULL TEXT

page 4

page 5

page 6

page 7

page 10

page 11

research
07/14/2020

Knowledge Distillation for Multi-task Learning

Multi-task learning (MTL) is to learn one single model that performs mul...
research
05/28/2022

Automatic Expert Selection for Multi-Scenario and Multi-Task Search

Multi-scenario learning (MSL) enables a service provider to cater for us...
research
04/17/2023

AdaMTL: Adaptive Input-dependent Inference for Efficient Multi-Task Learning

Modern Augmented reality applications require performing multiple tasks ...
research
05/31/2022

Compressed Hierarchical Representations for Multi-Task Learning and Task Clustering

In this paper, we frame homogeneous-feature multi-task learning (MTL) as...
research
04/03/2020

Context-Aware Multi-Task Learning for Traffic Scene Recognition in Autonomous Vehicles

Traffic scene recognition, which requires various visual classification ...
research
03/25/2019

Predicting Multiple Demographic Attributes with Task Specific Embedding Transformation and Attention Network

Most companies utilize demographic information to develop their strategy...
research
10/26/2022

M^3ViT: Mixture-of-Experts Vision Transformer for Efficient Multi-task Learning with Model-Accelerator Co-design

Multi-task learning (MTL) encapsulates multiple learned tasks in a singl...

Please sign up or login with your details

Forgot password? Click here to reset