T-former: An Efficient Transformer for Image Inpainting

05/12/2023
by   Ye Deng, et al.
0

Benefiting from powerful convolutional neural networks (CNNs), learning-based image inpainting methods have made significant breakthroughs over the years. However, some nature of CNNs (e.g. local prior, spatially shared parameters) limit the performance in the face of broken images with diverse and complex forms. Recently, a class of attention-based network architectures, called transformer, has shown significant performance on natural language processing fields and high-level vision tasks. Compared with CNNs, attention operators are better at long-range modeling and have dynamic weights, but their computational complexity is quadratic in spatial resolution, and thus less suitable for applications involving higher resolution images, such as image inpainting. In this paper, we design a novel attention linearly related to the resolution according to Taylor expansion. And based on this attention, a network called $T$-former is designed for image inpainting. Experiments on several benchmark datasets demonstrate that our proposed method achieves state-of-the-art accuracy while maintaining a relatively low number of parameters and computational complexity. The code can be found at \href{https://github.com/dengyecode/T-former_image_inpainting}{github.com/dengyecode/T-former\_image\_inpainting}

READ FULL TEXT

page 1

page 5

page 7

page 8

research
03/29/2022

MAT: Mask-Aware Transformer for Large Hole Image Inpainting

Recent studies have shown the importance of modeling long-range interact...
research
01/13/2020

Natural Image Matting via Guided Contextual Attention

Over the last few years, deep learning based approaches have achieved ou...
research
03/02/2022

Incremental Transformer Structure Enhanced Image Inpainting with Masking Positional Encoding

Image inpainting has made significant advances in recent years. However,...
research
03/15/2021

UPANets: Learning from the Universal Pixel Attention Networks

Among image classification, skip and densely-connection-based networks h...
research
03/09/2021

Enhancing sensor resolution improves CNN accuracy given the same number of parameters or FLOPS

High image resolution is critical to obtain a good performance in many c...
research
10/09/2022

Strong Gravitational Lensing Parameter Estimation with Vision Transformer

Quantifying the parameters and corresponding uncertainties of hundreds o...
research
05/13/2022

Blind Image Inpainting with Sparse Directional Filter Dictionaries for Lightweight CNNs

Blind inpainting algorithms based on deep learning architectures have sh...

Please sign up or login with your details

Forgot password? Click here to reset