Target Aware Network Architecture Search and Compression for Efficient Knowledge Transfer

05/12/2022
by   S. H. Shabbeer Basha, et al.
0

Transfer Learning enables Convolutional Neural Networks (CNN) to acquire knowledge from a source domain and transfer it to a target domain, where collecting large-scale annotated examples is both time-consuming and expensive. Conventionally, while transferring the knowledge learned from one task to another task, the deeper layers of a pre-trained CNN are finetuned over the target dataset. However, these layers that are originally designed for the source task are over-parameterized for the target task. Thus, finetuning these layers over the target dataset reduces the generalization ability of the CNN due to high network complexity. To tackle this problem, we propose a two-stage framework called TASCNet which enables efficient knowledge transfer. In the first stage, the configuration of the deeper layers is learned automatically and finetuned over the target dataset. Later, in the second stage, the redundant filters are pruned from the fine-tuned CNN to decrease the network's complexity for the target task while preserving the performance. This two-stage mechanism finds a compact version of the pre-trained CNN with optimal structure (number of filters in a convolutional layer, number of neurons in a dense layer, and so on) from the hypothesis space. The efficacy of the proposed method is evaluated using VGG-16, ResNet-50, and DenseNet-121 on CalTech-101, CalTech-256, and Stanford Dogs datasets. The proposed TASCNet reduces the computational complexity of pre-trained CNNs over the target task by reducing both trainable parameters and FLOPs which enables resource-efficient knowledge transfer.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/09/2021

AutoFT: Automatic Fine-Tune for Parameters Transfer Learning in Click-Through Rate Prediction

Recommender systems are often asked to serve multiple recommendation sce...
research
10/02/2018

Target Aware Network Adaptation for Efficient Representation Learning

This paper presents an automatic network adaptation method that finds a ...
research
02/14/2019

Transfer Learning for Sequence Labeling Using Source Model and Target Data

In this paper, we propose an approach for transferring the knowledge of ...
research
12/27/2017

Learning More Universal Representations for Transfer-Learning

Transfer learning is commonly used to address the problem of the prohibi...
research
03/01/2023

Speeding Up EfficientNet: Selecting Update Blocks of Convolutional Neural Networks using Genetic Algorithm in Transfer Learning

The performance of convolutional neural networks (CNN) depends heavily o...
research
03/24/2021

Transfer Learning for Piano Sustain-Pedal Detection

Detecting piano pedalling techniques in polyphonic music remains a chall...
research
08/08/2020

Using UNet and PSPNet to explore the reusability principle of CNN parameters

How to reduce the requirement on training dataset size is a hot topic in...

Please sign up or login with your details

Forgot password? Click here to reset