Disentangle and Remerge: Interventional Knowledge Distillation for Few-Shot Object Detection from A Conditional Causal Perspective

08/26/2022
by   Jiangmeng Li, et al.
0

Few-shot learning models learn representations with limited human annotations, and such a learning paradigm demonstrates practicability in various tasks, e.g., image classification, object detection, etc. However, few-shot object detection methods suffer from an intrinsic defect that the limited training data makes the model cannot sufficiently explore semantic information. To tackle this, we introduce knowledge distillation to the few-shot object detection learning paradigm. We further run a motivating experiment, which demonstrates that in the process of knowledge distillation the empirical error of the teacher model degenerates the prediction performance of the few-shot object detection model, as the student. To understand the reasons behind this phenomenon, we revisit the learning paradigm of knowledge distillation on the few-shot object detection task from the causal theoretic standpoint, and accordingly, develop a Structural Causal Model. Following the theoretical guidance, we propose a backdoor adjustment-based knowledge distillation method for the few-shot object detection task, namely Disentangle and Remerge (D R), to perform conditional causal intervention toward the corresponding Structural Causal Model. Theoretically, we provide an extended definition, i.e., general backdoor path, for the backdoor criterion, which can expand the theoretical application boundary of the backdoor criterion in specific cases. Empirically, the experiments on multiple benchmark datasets demonstrate that D R can yield significant performance boosts in few-shot object detection.

READ FULL TEXT

page 5

page 12

research
09/16/2021

Label Assignment Distillation for Object Detection

Knowledge distillation methods are proved to be promising in improving t...
research
06/20/2019

GAN-Knowledge Distillation for one-stage Object Detection

Convolutional neural networks have a significant improvement in the accu...
research
03/20/2023

A closer look at the training dynamics of knowledge distillation

In this paper we revisit the efficacy of knowledge distillation as a fun...
research
11/30/2022

Random Copolymer inverse design system orienting on Accurate discovering of Antimicrobial peptide-mimetic copolymers

Antimicrobial resistance is one of the biggest health problem, especiall...
research
03/04/2019

TKD: Temporal Knowledge Distillation for Active Perception

Deep neural networks based methods have been proved to achieve outstandi...
research
07/25/2022

Few-Shot Object Detection by Knowledge Distillation Using Bag-of-Visual-Words Representations

While fine-tuning based methods for few-shot object detection have achie...
research
09/10/2023

Exploiting CLIP for Zero-shot HOI Detection Requires Knowledge Distillation at Multiple Levels

In this paper, we investigate the task of zero-shot human-object interac...

Please sign up or login with your details

Forgot password? Click here to reset