AiATrack: Attention in Attention for Transformer Visual Tracking

07/20/2022
by   Shenyuan Gao, et al.
0

Transformer trackers have achieved impressive advancements recently, where the attention mechanism plays an important role. However, the independent correlation computation in the attention mechanism could result in noisy and ambiguous attention weights, which inhibits further performance improvement. To address this issue, we propose an attention in attention (AiA) module, which enhances appropriate correlations and suppresses erroneous ones by seeking consensus among all correlation vectors. Our AiA module can be readily applied to both self-attention blocks and cross-attention blocks to facilitate feature aggregation and information propagation for visual tracking. Moreover, we propose a streamlined Transformer tracking framework, dubbed AiATrack, by introducing efficient feature reuse and target-background embeddings to make full use of temporal references. Experiments show that our tracker achieves state-of-the-art performance on six tracking benchmarks while running at a real-time speed.

READ FULL TEXT

page 2

page 7

page 9

page 21

page 24

research
01/26/2023

Compact Transformer Tracker with Correlative Masked Modeling

Transformer framework has been showing superior performances in visual o...
research
05/09/2021

TrTr: Visual Tracking with Transformer

Template-based discriminative trackers are currently the dominant tracki...
research
08/14/2021

PTT: Point-Track-Transformer Module for 3D Single Object Tracking in Point Clouds

3D single object tracking is a key issue for robotics. In this paper, we...
research
04/14/2020

Deformable Siamese Attention Networks for Visual Object Tracking

Siamese-based trackers have achieved excellent performance on visual obj...
research
09/15/2022

Beat Transformer: Demixed Beat and Downbeat Tracking with Dilated Self-Attention

We propose Beat Transformer, a novel Transformer encoder architecture fo...
research
09/29/2020

Attention that does not Explain Away

Models based on the Transformer architecture have achieved better accura...
research
06/07/2023

Cross-attention learning enables real-time nonuniform rotational distortion correction in OCT

Nonuniform rotational distortion (NURD) correction is vital for endoscop...

Please sign up or login with your details

Forgot password? Click here to reset