Efficient Linear Attention for Fast and Accurate Keypoint Matching

04/16/2022
by   Suwichaya Suwanwimolkul, et al.
0

Recently Transformers have provided state-of-the-art performance in sparse matching, crucial to realize high-performance 3D vision applications. Yet, these Transformers lack efficiency due to the quadratic computational complexity of their attention mechanism. To solve this problem, we employ an efficient linear attention for the linear computational complexity. Then, we propose a new attentional aggregation that achieves high accuracy by aggregating both the global and local information from sparse keypoints. To further improve the efficiency, we propose the joint learning of feature matching and description. Our learning enables simpler and faster matching than Sinkhorn, often used in matching the learned descriptors from Transformers. Our method achieves competitive performance with only 0.84M learnable parameters against the bigger SOTAs, SuperGlue (12M parameters) and SGMNet (30M parameters), on three benchmarks, HPatch, ETH, and Aachen Day-Night.

READ FULL TEXT

page 1

page 3

page 5

page 11

page 12

page 15

page 16

page 17

research
10/15/2022

Linear Video Transformer with Feature Fixation

Vision Transformers have achieved impressive performance in video classi...
research
03/02/2023

ParaFormer: Parallel Attention Transformer for Efficient Feature Matching

Heavy computation is a bottleneck limiting deep-learningbased feature ma...
research
01/08/2022

QuadTree Attention for Vision Transformers

Transformers have been successful in many vision tasks, thanks to their ...
research
01/19/2016

A Theory of Local Matching: SIFT and Beyond

Why has SIFT been so successful? Why its extension, DSP-SIFT, can furthe...
research
05/30/2021

Transformer-Based Deep Image Matching for Generalizable Person Re-identification

Transformers have recently gained increasing attention in computer visio...
research
10/08/2021

Token Pooling in Vision Transformers

Despite the recent success in many applications, the high computational ...
research
02/14/2022

CATs++: Boosting Cost Aggregation with Convolutions and Transformers

Cost aggregation is a highly important process in image matching tasks, ...

Please sign up or login with your details

Forgot password? Click here to reset