Learned Weight Sharing for Deep Multi-Task Learning by Natural Evolution Strategy and Stochastic Gradient Descent

03/23/2020
by   Jonas Prellberg, et al.
0

In deep multi-task learning, weights of task-specific networks are shared between tasks to improve performance on each single one. Since the question, which weights to share between layers, is difficult to answer, human-designed architectures often share everything but a last task-specific layer. In many cases, this simplistic approach severely limits performance. Instead, we propose an algorithm to learn the assignment between a shared set of weights and task-specific layers. To optimize the non-differentiable assignment and at the same time train the differentiable weights, learning takes place via a combination of natural evolution strategy and stochastic gradient descent. The end result are task-specific networks that share weights but allow independent inference. They achieve lower test errors than baselines and methods from literature on three multi-task learning datasets.

READ FULL TEXT

page 1

page 7

research
11/27/2019

AdaShare: Learning What To Share For Efficient Deep Multi-Task Learning

Multi-task learning is an open and challenging problem in computer visio...
research
07/23/2021

Rethinking Hard-Parameter Sharing in Multi-Task Learning

Hard parameter sharing in multi-task learning (MTL) allows tasks to shar...
research
08/26/2019

Stochastic Filter Groups for Multi-Task CNNs: Learning Specialist and Generalist Convolution Kernels

The performance of multi-task learning in Convolutional Neural Networks ...
research
09/21/2021

Optimization Strategies in Multi-Task Learning: Averaged or Separated Losses?

In Multi-Task Learning (MTL), it is a common practice to train multi-tas...
research
10/11/2021

Mining the Weights Knowledge for Optimizing Neural Network Structures

Knowledge embedded in the weights of the artificial neural network can b...
research
01/22/2021

Network Clustering for Multi-task Learning

The Multi-Task Learning (MTL) technique has been widely studied by word-...
research
05/31/2019

Modular Universal Reparameterization: Deep Multi-task Learning Across Diverse Domains

As deep learning applications continue to become more diverse, an intere...

Please sign up or login with your details

Forgot password? Click here to reset