Distribution Distillation Loss: Generic Approach for Improving Face Recognition from Hard Samples

02/10/2020
by   Yuge Huang, et al.
2

Large facial variations are the main challenge in face recognition. To this end, previous variation-specific methods make full use of task-related prior to design special network losses, which are typically not general among different tasks and scenarios. In contrast, the existing generic methods focus on improving the feature discriminability to minimize the intra-class distance while maximizing the interclass distance, which perform well on easy samples but fail on hard samples. To improve the performance on those hard samples for general tasks, we propose a novel Distribution Distillation Loss to narrow the performance gap between easy and hard samples, which is a simple, effective and generic for various types of facial variations. Specifically, we first adopt state-of-the-art classifiers such as ArcFace to construct two similarity distributions: teacher distribution from easy samples and student distribution from hard samples. Then, we propose a novel distribution-driven loss to constrain the student distribution to approximate the teacher distribution, which thus leads to smaller overlap between the positive and negative pairs in the student distribution. We have conducted extensive experiments on both generic large-scale face benchmarks and benchmarks with diverse variations on race, resolution and pose. The quantitative results demonstrate the superiority of our method over strong baselines, e.g., Arcface and Cosface.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/06/2022

Evaluation-oriented Knowledge Distillation for Deep Face Recognition

Knowledge distillation (KD) is a widely-used technique that utilizes lar...
research
10/31/2020

ProxylessKD: Direct Knowledge Distillation with Inherited Classifier for Face Recognition

Knowledge Distillation (KD) refers to transferring knowledge from a larg...
research
07/20/2020

NPCFace: A Negative-Positive Cooperation Supervision for Training Large-scale Face Recognition

Deep face recognition has made remarkable advances in the last few years...
research
04/10/2023

Grouped Knowledge Distillation for Deep Face Recognition

Compared with the feature-based distillation methods, logits distillatio...
research
03/28/2022

OTFace: Hard Samples Guided Optimal Transport Loss for Deep Face Representation

Face representation in the wild is extremely hard due to the large scale...
research
06/26/2023

Cross Architecture Distillation for Face Recognition

Transformers have emerged as the superior choice for face recognition ta...
research
02/26/2020

Towards Universal Representation Learning for Deep Face Recognition

Recognizing wild faces is extremely hard as they appear with all kinds o...

Please sign up or login with your details

Forgot password? Click here to reset