Efficiency 360: Efficient Vision Transformers

02/16/2023
by   Badri N. Patro, et al.
5

Transformers are widely used for solving tasks in natural language processing, computer vision, speech, and music domains. In this paper, we talk about the efficiency of transformers in terms of memory (the number of parameters), computation cost (number of floating points operations), and performance of models, including accuracy, the robustness of the model, and fair & bias-free features. We mainly discuss the vision transformer for the image classification task. Our contribution is to introduce an efficient 360 framework, which includes various aspects of the vision transformer, to make it more efficient for industrial applications. By considering those applications, we categorize them into multiple dimensions such as privacy, robustness, transparency, fairness, inclusiveness, continual learning, probabilistic models, approximation, computational complexity, and spectral complexity. We compare various vision transformer models based on their performance, the number of parameters, and the number of floating point operations (FLOPs) on multiple datasets.

READ FULL TEXT
research
09/14/2020

Efficient Transformers: A Survey

Transformer model architectures have garnered immense interest lately du...
research
09/15/2022

Number of Attention Heads vs Number of Transformer-Encoders in Computer Vision

Determining an appropriate number of attention heads on one hand and the...
research
08/18/2023

Which Transformer to Favor: A Comparative Analysis of Efficiency in Vision Transformers

The growing popularity of Vision Transformers as the go-to models for im...
research
07/18/2023

Light-Weight Vision Transformer with Parallel Local and Global Self-Attention

While transformer architectures have dominated computer vision in recent...
research
06/15/2022

Efficient Adaptive Ensembling for Image Classification

In recent times, except for sporadic cases, the trend in Computer Vision...
research
12/03/2018

Structure Learning Using Forced Pruning

Markov networks are widely used in many Machine Learning applications in...
research
02/08/2023

Mitigating Bias in Visual Transformers via Targeted Alignment

As transformer architectures become increasingly prevalent in computer v...

Please sign up or login with your details

Forgot password? Click here to reset