Distilling Object Detectors with Fine-grained Feature Imitation

06/09/2019
by   PetsTime, et al.
4

State-of-the-art CNN based recognition models are often computationally prohibitive to deploy on low-end devices. A promising high level approach tackling this limitation is knowledge distillation, which let small student model mimic cumbersome teacher model's output to get improved generalization. However, related methods mainly focus on simple task of classification while do not consider complex tasks like object detection. We show applying the vanilla knowledge distillation to detection model gets minor gain. To address the challenge of distilling knowledge in detection model, we propose a fine-grained feature imitation method exploiting the cross-location discrepancy of feature response. Our intuition is that detectors care more about local near object regions. Thus the discrepancy of feature response on the near object anchor locations reveals important information of how teacher model tends to generalize. We design a novel mechanism to estimate those locations and let student model imitate the teacher on them to get enhanced performance. We first validate the idea on a developed lightweight toy detector which carries simplest notion of current state-of-the-art anchor based detection models on challenging KITTI dataset, our method generates up to 15 student model compared to the non-imitated counterpart. We then extensively evaluate the method with Faster R-CNN model under various scenarios with common object detection benchmark of Pascal VOC and COCO, imitation alleviates up to 74 https://github.com/twangnh/Distilling-Object-Detectors

READ FULL TEXT

page 1

page 6

page 7

research
03/07/2023

Gradient-Guided Knowledge Distillation for Object Detectors

Deep learning models have demonstrated remarkable success in object dete...
research
10/17/2022

Distilling Object Detectors With Global Knowledge

Knowledge distillation learns a lightweight student model that mimics a ...
research
03/26/2021

Distilling Object Detectors via Decoupled Features

Knowledge distillation is a widely used paradigm for inheriting informat...
research
10/07/2022

IDa-Det: An Information Discrepancy-aware Distillation for 1-bit Detectors

Knowledge distillation (KD) has been proven to be useful for training co...
research
01/04/2023

StereoDistill: Pick the Cream from LiDAR for Distilling Stereo-based 3D Object Detection

In this paper, we propose a cross-modal distillation method named Stereo...
research
03/02/2022

SEA: Bridging the Gap Between One- and Two-stage Detector Distillation via SEmantic-aware Alignment

We revisit the one- and two-stage detector distillation tasks and presen...
research
08/18/2023

Small Object Detection via Coarse-to-fine Proposal Generation and Imitation Learning

The past few years have witnessed the immense success of object detectio...

Please sign up or login with your details

Forgot password? Click here to reset