Bridging Cross-task Protocol Inconsistency for Distillation in Dense Object Detection

08/28/2023
by   Longrong Yang, et al.
0

Knowledge distillation (KD) has shown potential for learning compact models in dense object detection. However, the commonly used softmax-based distillation ignores the absolute classification scores for individual categories. Thus, the optimum of the distillation loss does not necessarily lead to the optimal student classification scores for dense object detectors. This cross-task protocol inconsistency is critical, especially for dense object detectors, since the foreground categories are extremely imbalanced. To address the issue of protocol differences between distillation and classification, we propose a novel distillation method with cross-task consistent protocols, tailored for the dense object detection. For classification distillation, we address the cross-task protocol inconsistency problem by formulating the classification logit maps in both teacher and student models as multiple binary-classification maps and applying a binary-classification distillation loss to each map. For localization distillation, we design an IoU-based Localization Distillation Loss that is free from specific network structures and can be compared with existing localization distillation losses. Our proposed method is simple but effective, and experimental results demonstrate its superiority over existing methods. Code is available at https://github.com/TinyTigerPan/BCKD.

READ FULL TEXT

page 8

page 14

research
03/26/2021

Distilling Object Detectors via Decoupled Features

Knowledge distillation is a widely used paradigm for inheriting informat...
research
09/16/2021

Label Assignment Distillation for Object Detection

Knowledge distillation methods are proved to be promising in improving t...
research
04/19/2021

Distilling Knowledge via Knowledge Review

Knowledge distillation transfers knowledge from the teacher network to t...
research
08/05/2022

Task-Balanced Distillation for Object Detection

Mainstream object detectors are commonly constituted of two sub-tasks, i...
research
11/30/2022

Attention-based Depth Distillation with 3D-Aware Positional Encoding for Monocular 3D Object Detection

Monocular 3D object detection is a low-cost but challenging task, as it ...
research
04/24/2023

Function-Consistent Feature Distillation

Feature distillation makes the student mimic the intermediate features o...
research
08/20/2023

Representation Disparity-aware Distillation for 3D Object Detection

In this paper, we focus on developing knowledge distillation (KD) for co...

Please sign up or login with your details

Forgot password? Click here to reset