Joint auto-encoders: a flexible multi-task learning framework

05/30/2017
by   Baruch Epstein. Ron Meir, et al.
0

The incorporation of prior knowledge into learning is essential in achieving good performance based on small noisy samples. Such knowledge is often incorporated through the availability of related data arising from domains and tasks similar to the one of current interest. Ideally one would like to allow both the data for the current task and for previous related tasks to self-organize the learning system in such a way that commonalities and differences between the tasks are learned in a data-driven fashion. We develop a framework for learning multiple tasks simultaneously, based on sharing features that are common to all tasks, achieved through the use of a modular deep feedforward neural network consisting of shared branches, dealing with the common features of all tasks, and private branches, learning the specific unique aspects of each task. Once an appropriate weight sharing architecture has been established, learning takes place through standard algorithms for feedforward networks, e.g., stochastic gradient descent and its variations. The method deals with domain adaptation and multi-task learning in a unified fashion, and can easily deal with data arising from different types of sources. Numerical experiments demonstrate the effectiveness of learning in domain adaptation and transfer learning setups, and provide evidence for the flexible and task-oriented representations arising in the network.

READ FULL TEXT
research
08/09/2016

Multi-task Domain Adaptation for Sequence Tagging

Many domain adaptation approaches rely on learning cross domain shared r...
research
07/23/2021

Compositional Models: Multi-Task Learning and Knowledge Transfer with Modular Networks

Conditional computation and modular networks have been recently proposed...
research
11/03/2017

Lifelong Learning by Adjusting Priors

In representational lifelong learning an agent aims to continually learn...
research
01/04/2023

Multi-Task Learning with Prior Information

Multi-task learning aims to boost the generalization performance of mult...
research
11/20/2021

Safe Multi-Task Learning

In recent years, Multi-Task Learning (MTL) attracts much attention due t...
research
11/24/2017

Cross-Domain Self-supervised Multi-task Feature Learning using Synthetic Imagery

In human learning, it is common to use multiple sources of information j...
research
05/22/2017

Learning multiple visual domains with residual adapters

There is a growing interest in learning data representations that work w...

Please sign up or login with your details

Forgot password? Click here to reset