Compositional Models: Multi-Task Learning and Knowledge Transfer with Modular Networks

07/23/2021
by   Andrey Zhmoginov, et al.
0

Conditional computation and modular networks have been recently proposed for multitask learning and other problems as a way to decompose problem solving into multiple reusable computational blocks. We propose a new approach for learning modular networks based on the isometric version of ResNet with all residual blocks having the same configuration and the same number of parameters. This architectural choice allows adding, removing and changing the order of residual blocks. In our method, the modules can be invoked repeatedly and allow knowledge transfer to novel tasks by adjusting the order of computation. This allows soft weight sharing between tasks with only a small increase in the number of parameters. We show that our method leads to interpretable self-organization of modules in case of multi-task learning, transfer learning and domain adaptation while achieving competitive results on those tasks. From practical perspective, our approach allows to: (a) reuse existing modules for learning new task by adjusting the computation order, (b) use it for unsupervised multi-source domain adaptation to illustrate that adaptation to unseen data can be achieved by only manipulating the order of pretrained modules, (c) show how our approach can be used to increase accuracy of existing architectures for image classification tasks such as ImageNet, without any parameter increase, by reusing the same block multiple times.

READ FULL TEXT

page 6

page 12

research
10/25/2018

K For The Price Of 1: Parameter Efficient Multi-task And Transfer Learning

We introduce a novel method that enables parameter-efficient transfer an...
research
11/13/2018

Modular Networks: Learning to Decompose Neural Computation

Scaling model capacity has been vital in the success of deep learning. F...
research
05/30/2017

Joint auto-encoders: a flexible multi-task learning framework

The incorporation of prior knowledge into learning is essential in achie...
research
03/16/2023

Efficient Computation Sharing for Multi-Task Visual Scene Understanding

Solving multiple visual tasks using individual models can be resource-in...
research
02/06/2023

Multipath agents for modular multitask ML systems

A standard ML model is commonly generated by a single method that specif...
research
04/11/2023

Conditional Adapters: Parameter-efficient Transfer Learning with Fast Inference

We propose Conditional Adapter (CoDA), a parameter-efficient transfer le...
research
06/01/2020

Latent Domain Learning with Dynamic Residual Adapters

A practical shortcoming of deep neural networks is their specialization ...

Please sign up or login with your details

Forgot password? Click here to reset