Towards Accurate Knowledge Transfer via Target-awareness Representation Disentanglement

10/16/2020
by   Xingjian Li, et al.
11

Fine-tuning deep neural networks pre-trained on large scale datasets is one of the most practical transfer learning paradigm given limited quantity of training samples. To obtain better generalization, using the starting point as the reference, either through weights or features, has been successfully applied to transfer learning as a regularizer. However, due to the domain discrepancy between the source and target tasks, there exists obvious risk of negative transfer. In this paper, we propose a novel transfer learning algorithm, introducing the idea of Target-awareness REpresentation Disentanglement (TRED), where the relevant knowledge with respect to the target task is disentangled from the original source model and used as a regularizer during fine-tuning the target model. Experiments on various real world datasets show that our method stably improves the standard fine-tuning by more than 2 in average. TRED also outperforms other state-of-the-art transfer learning regularizers such as L2-SP, AT, DELTA and BSS.

READ FULL TEXT

page 3

page 7

research
11/21/2018

SpotTune: Transfer Learning through Adaptive Fine-tuning

Transfer learning, which allows a source task to affect the inductive bi...
research
10/30/2018

Gated Transfer Network for Transfer Learning

Deep neural networks have led to a series of breakthroughs in computer v...
research
07/20/2020

XMixup: Efficient Transfer Learning with Auxiliary Samples by Cross-domain Mixup

Transferring knowledge from large source datasets is an effective way to...
research
08/29/2019

Learning to Transfer Learn

We propose a novel framework, learning to transfer learn (L2TL), to impr...
research
04/12/2017

Representation Stability as a Regularizer for Improved Text Analytics Transfer Learning

Although neural networks are well suited for sequential transfer learnin...
research
10/15/2018

Deep Transfer Reinforcement Learning for Text Summarization

Deep neural networks are data hungry models and thus they face difficult...
research
08/16/2019

Pseudo-task Regularization for ConvNet Transfer Learning

This paper is about regularizing deep convolutional networks (ConvNets) ...

Please sign up or login with your details

Forgot password? Click here to reset