Network Clustering for Multi-task Learning

01/22/2021
by   Dehong Gao, et al.
0

The Multi-Task Learning (MTL) technique has been widely studied by word-wide researchers. The majority of current MTL studies adopt the hard parameter sharing structure, where hard layers tend to learn general representations over all tasks and specific layers are prone to learn specific representations for each task. Since the specific layers directly follow the hard layers, the MTL model needs to estimate this direct change (from general to specific) as well. To alleviate this problem, we introduce the novel cluster layer, which groups tasks into clusters during training procedures. In a cluster layer, the tasks in the same cluster are further required to share the same network. By this way, the cluster layer produces the general presentation for the same cluster, while produces relatively specific presentations for different clusters. As transitions the cluster layers are used between the hard layers and the specific layers. The MTL model thus learns general representations to specific representations gradually. We evaluate our model with MTL document classification and the results demonstrate the cluster layer is quite efficient in MTL.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/23/2021

Rethinking Hard-Parameter Sharing in Multi-Task Learning

Hard parameter sharing in multi-task learning (MTL) allows tasks to shar...
research
04/05/2019

Branched Multi-Task Networks: Deciding What Layers To Share

In the context of deep learning, neural networks with multiple branches ...
research
08/26/2019

Stochastic Filter Groups for Multi-Task CNNs: Learning Specialist and Generalist Convolution Kernels

The performance of multi-task learning in Convolutional Neural Networks ...
research
03/23/2020

Learned Weight Sharing for Deep Multi-Task Learning by Natural Evolution Strategy and Stochastic Gradient Descent

In deep multi-task learning, weights of task-specific networks are share...
research
04/26/2023

Multi-Task Learning Regression via Convex Clustering

Multi-task learning (MTL) is a methodology that aims to improve the gene...
research
05/23/2017

Sluice networks: Learning what to share between loosely related tasks

Multi-task learning is partly motivated by the observation that humans b...
research
03/08/2023

Provable Pathways: Learning Multiple Tasks over Multiple Paths

Constructing useful representations across a large number of tasks is a ...

Please sign up or login with your details

Forgot password? Click here to reset