KTAN: Knowledge Transfer Adversarial Network

10/18/2018
by   Peiye Liu, et al.
0

To reduce the large computation and storage cost of a deep convolutional neural network, the knowledge distillation based methods have pioneered to transfer the generalization ability of a large (teacher) deep network to a light-weight (student) network. However, these methods mostly focus on transferring the probability distribution of the softmax layer in a teacher network and thus neglect the intermediate representations. In this paper, we propose a knowledge transfer adversarial network to better train a student network. Our technique holistically considers both intermediate representations and probability distributions of a teacher network. To transfer the knowledge of intermediate representations, we set high-level teacher feature maps as a target, toward which the student feature maps are trained. Specifically, we arrange a Teacher-to-Student layer for enabling our framework suitable for various student structures. The intermediate representation helps the student network better understand the transferred generalization as compared to the probability distribution only. Furthermore, we infuse an adversarial learning process by employing a discriminator network, which can fully exploit the spatial correlation of feature maps in training a student network. The experimental results demonstrate that the proposed method can significantly improve the performance of a student network on both image classification and object detection tasks.

READ FULL TEXT
research
03/31/2021

Knowledge Distillation By Sparse Representation Matching

Knowledge Distillation refers to a class of methods that transfers the k...
research
02/14/2018

Paraphrasing Complex Network: Network Compression via Factor Transfer

Deep neural networks (DNN) have recently shown promising performances in...
research
09/03/2021

Towards Learning Spatially Discriminative Feature Representations

The backbone of traditional CNN classifier is generally considered as a ...
research
10/16/2022

AttTrack: Online Deep Attention Transfer for Multi-object Tracking

Multi-object tracking (MOT) is a vital component of intelligent video an...
research
11/22/2021

Adaptive Transfer Learning: a simple but effective transfer learning

Transfer learning (TL) leverages previously obtained knowledge to learn ...
research
04/10/2019

Knowledge Squeezed Adversarial Network Compression

Deep network compression has been achieved notable progress via knowledg...
research
06/26/2022

Knowledge Distillation with Representative Teacher Keys Based on Attention Mechanism for Image Classification Model Compression

With the improvement of AI chips (e.g., GPU, TPU, and NPU) and the fast ...

Please sign up or login with your details

Forgot password? Click here to reset