Factorized Attention: Self-Attention with Linear Complexities

12/04/2018
by   Shen Zhuoran, et al.
0

Recent works have been applying self-attention to various fields in computer vision and natural language processing. However, the memory and computational demands of existing self-attention operations grow quadratically with the spatiotemporal size of the input. This prohibits the application of self-attention on large inputs, e.g., long sequences, high-definition images, or large videos. To remedy this, this paper proposes a novel factorized attention (FA) module, which achieves the same expressive power as previous approaches with substantially less memory and computational consumption. The resource-efficiency allows more widespread and flexible application of it. Empirical evaluations on object recognition demonstrate the effectiveness of these advantages. FA-augmented models achieved state-of-the-art performance for object detection and instance segmentation on MS-COCO. Further, the resource-efficiency of FA democratizes self-attention to fields where the prohibitively high costs currently prevent its application. The state-of-the-art result for stereo depth estimation on the Scene Flow dataset exemplifies this.

READ FULL TEXT

page 1

page 2

page 8

research
06/23/2022

YOLOSA: Object detection based on 2D local feature superimposed self-attention

We analyzed the network structure of real-time object detection models a...
research
07/01/2022

Rethinking Query-Key Pairwise Interactions in Vision Transformers

Vision Transformers have achieved state-of-the-art performance in many v...
research
07/12/2021

Locally Enhanced Self-Attention: Rethinking Self-Attention as Local and Context Terms

Self-Attention has become prevalent in computer vision models. Inspired ...
research
05/28/2021

Linear-Time Self Attention with Codeword Histogram for Efficient Recommendation

Self-attention has become increasingly popular in a variety of sequence ...
research
10/31/2021

DPNET: Dual-Path Network for Efficient Object Detectioj with Lightweight Self-Attention

Object detection often costs a considerable amount of computation to get...
research
06/22/2020

Limits to Depth Efficiencies of Self-Attention

Self-attention architectures, which are rapidly pushing the frontier in ...
research
08/08/2023

Exploring the Spatiotemporal Features of Online Food Recommendation Service

Online Food Recommendation Service (OFRS) has remarkable spatiotemporal ...

Please sign up or login with your details

Forgot password? Click here to reset