TransTailor: Pruning the Pre-trained Model for Improved Transfer Learning

03/02/2021
by   Bingyan Liu, et al.
0

The increasing of pre-trained models has significantly facilitated the performance on limited data tasks with transfer learning. However, progress on transfer learning mainly focuses on optimizing the weights of pre-trained models, which ignores the structure mismatch between the model and the target task. This paper aims to improve the transfer performance from another angle - in addition to tuning the weights, we tune the structure of pre-trained models, in order to better match the target task. To this end, we propose TransTailor, targeting at pruning the pre-trained model for improved transfer learning. Different from traditional pruning pipelines, we prune and fine-tune the pre-trained model according to the target-aware weight importance, generating an optimal sub-model tailored for a specific target task. In this way, we transfer a more suitable sub-structure that can be applied during fine-tuning to benefit the final performance. Extensive experiments on multiple pre-trained models and datasets demonstrate that TransTailor outperforms the traditional pruning methods and achieves competitive or even better performance than other state-of-the-art transfer learning methods while using a smaller model. Notably, on the Stanford Dogs dataset, TransTailor can achieve 2.7 improvement over other transfer methods with 20

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2022

Hub-Pathway: Transfer Learning from A Hub of Pre-trained Models

Transfer learning aims to leverage knowledge from pre-trained models to ...
research
06/12/2022

PAC-Net: A Model Pruning Approach to Inductive Transfer Learning

Inductive transfer learning aims to learn from a small amount of trainin...
research
07/19/2022

Revealing Secrets From Pre-trained Models

With the growing burden of training deep learning models with large data...
research
02/22/2021

LogME: Practical Assessment of Pre-trained Models for Transfer Learning

This paper studies task adaptive pre-trained model selection, an underex...
research
07/19/2021

Non-binary deep transfer learning for imageclassification

The current standard for a variety of computer vision tasks using smalle...
research
08/19/2023

Disposable Transfer Learning for Selective Source Task Unlearning

Transfer learning is widely used for training deep neural networks (DNN)...
research
07/22/2020

Tiny Transfer Learning: Towards Memory-Efficient On-Device Learning

We present Tiny-Transfer-Learning (TinyTL), an efficient on-device learn...

Please sign up or login with your details

Forgot password? Click here to reset