Image classification has improved with the development of training
techn...
We need billion-scale images to achieve more generalizable and
ground-br...
Vision Transformer (ViT) extracts the final representation from either c...
In this paper, we aim to design a quantitative similarity function betwe...
The favorable performance of Vision Transformers (ViTs) is often attribu...
Ensembles of deep neural networks have demonstrated superior performance...
Trainable layers such as convolutional building blocks are the standard
...
The problem of class imbalanced data lies in that the generalization
per...
Vision Transformer (ViT) extends the application range of transformers f...
Knowledge distillation extracts general knowledge from a pre-trained tea...
ImageNet has been arguably the most popular image classification benchma...
State-of-the-art video action classifiers often suffer from overfitting....
This paper addresses representational bottleneck in a network and propos...
Normalization techniques, such as batch normalization (BN), have led to
...
We investigate the design aspects of feature distillation methods achiev...
In person re-identification (ReID) task, because of its shortage of trai...
An activation boundary for a neuron refers to a separating hyperplane th...
Many recent works on knowledge distillation have provided ways to transf...
Many recent works on knowledge distillation have provided ways to transf...