DELTA: DEep Learning Transfer using Feature Map with Attention for Convolutional Networks

01/26/2019
by   Xingjian Li, et al.
0

Transfer learning through fine-tuning a pre-trained neural network with an extremely large dataset, such as ImageNet, can significantly accelerate training while the accuracy is frequently bottlenecked by the limited dataset size of the new target task. To solve the problem, some regularization methods, constraining the outer layer weights of the target network using the starting point as references (SPAR), have been studied. In this paper, we propose a novel regularized transfer learning framework DELTA, namely DEep Learning Transfer using Feature Map with Attention. Instead of constraining the weights of neural network, DELTA aims to preserve the outer layer outputs of the target network. Specifically, in addition to minimizing the empirical loss, DELTA intends to align the outer layer outputs of two networks, through constraining a subset of feature maps that are precisely selected by attention that has been learned in an supervised learning manner. We evaluate DELTA with the state-of-the-art algorithms, including L2 and L2-SP. The experiment results show that our proposed method outperforms these baselines with higher accuracy for new tasks.

READ FULL TEXT
research
11/18/2019

Towards Making Deep Transfer Learning Never Hurt

Transfer learning have been frequently used to improve deep neural netwo...
research
07/20/2020

XMixup: Efficient Transfer Learning with Auxiliary Samples by Cross-domain Mixup

Transferring knowledge from large source datasets is an effective way to...
research
07/19/2021

Non-binary deep transfer learning for imageclassification

The current standard for a variety of computer vision tasks using smalle...
research
03/22/2021

Channel Scaling: A Scale-and-Select Approach for Transfer Learning

Transfer learning with pre-trained neural networks is a common strategy ...
research
07/22/2020

Tiny Transfer Learning: Towards Memory-Efficient On-Device Learning

We present Tiny-Transfer-Learning (TinyTL), an efficient on-device learn...
research
03/25/2021

SMILE: Self-Distilled MIxup for Efficient Transfer LEarning

To improve the performance of deep learning, mixup has been proposed to ...
research
08/08/2020

Using UNet and PSPNet to explore the reusability principle of CNN parameters

How to reduce the requirement on training dataset size is a hot topic in...

Please sign up or login with your details

Forgot password? Click here to reset