From Multimodal to Unimodal Attention in Transformers using Knowledge Distillation

10/15/2021
by   Dhruv Agarwal, et al.
0

Multimodal Deep Learning has garnered much interest, and transformers have triggered novel approaches, thanks to the cross-attention mechanism. Here we propose an approach to deal with two key existing challenges: the high computational resource demanded and the issue of missing modalities. We introduce for the first time the concept of knowledge distillation in transformers to use only one modality at inference time. We report a full study analyzing multiple student-teacher configurations, levels at which distillation is applied, and different methodologies. With the best configuration, we improved the state-of-the-art accuracy by 3 parameters by 2.5 times and the inference time by 22 performance-computation tradeoff can be exploited in many applications and we aim at opening a new research area where the deployment of complex models with limited resources is demanded.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset