DMKD: Improving Feature-based Knowledge Distillation for Object Detection Via Dual Masking Augmentation

09/06/2023
by   Guang Yang, et al.
0

Recent mainstream masked distillation methods function by reconstructing selectively masked areas of a student network from the feature map of its teacher counterpart. In these methods, the masked regions need to be properly selected, such that reconstructed features encode sufficient discrimination and representation capability like the teacher feature. However, previous masked distillation methods only focus on spatial masking, making the resulting masked areas biased towards spatial importance without encoding informative channel clues. In this study, we devise a Dual Masked Knowledge Distillation (DMKD) framework which can capture both spatially important and channel-wise informative clues for comprehensive masked feature reconstruction. More specifically, we employ dual attention mechanism for guiding the respective masking branches, leading to reconstructed feature encoding dual significance. Furthermore, fusing the reconstructed features is achieved by self-adjustable weighting strategy for effective feature distillation. Our experiments on object detection task demonstrate that the student networks achieve performance gains of 4.1 Mask R-CNN are respectively used as the teacher networks, while outperforming the other state-of-the-art distillation methods.

READ FULL TEXT

page 1

page 2

page 3

research
01/31/2023

AMD: Adaptive Masked Distillation for Object

As a general model compression paradigm, feature-based knowledge distill...
research
03/07/2022

Enhance Language Identification using Dual-mode Model with Knowledge Distillation

In this paper, we propose to employ a dual-mode framework on the x-vecto...
research
03/10/2022

Prediction-Guided Distillation for Dense Object Detection

Real-world object detection models should be cheap and accurate. Knowled...
research
12/17/2021

Weakly Supervised Semantic Segmentation via Alternative Self-Dual Teaching

Current weakly supervised semantic segmentation (WSSS) frameworks usuall...
research
05/31/2022

itKD: Interchange Transfer-based Knowledge Distillation for 3D Object Detection

Recently, point-cloud based 3D object detectors have achieved remarkable...
research
02/11/2023

Dual Relation Knowledge Distillation for Object Detection

Knowledge distillation is an effective method for model compression. How...
research
06/28/2023

Hybrid Distillation: Connecting Masked Autoencoders with Contrastive Learners

Representation learning has been evolving from traditional supervised tr...

Please sign up or login with your details

Forgot password? Click here to reset