Efficient Transformer for Single Image Super-Resolution

08/25/2021
by   Zhisheng Lu, et al.
15

Single image super-resolution task has witnessed great strides with the development of deep learning. However, most existing studies focus on building a more complex neural network with a massive number of layers, bringing heavy computational cost and memory storage. Recently, as Transformer yields brilliant results in NLP tasks, more and more researchers start to explore the application of Transformer in computer vision tasks. But with the heavy computational cost and high GPU memory occupation of the vision Transformer, the network can not be designed too deep. To address this problem, we propose a novel Efficient Super-Resolution Transformer (ESRT) for fast and accurate image super-resolution. ESRT is a hybrid Transformer where a CNN-based SR network is first designed in the front to extract deep features. Specifically, there are two backbones for formatting the ESRT: lightweight CNN backbone (LCB) and lightweight Transformer backbone (LTB). Among them, LCB is a lightweight SR network to extract deep SR features at a low computational cost by dynamically adjusting the size of the feature map. LTB is made up of an efficient Transformer (ET) with a small GPU memory occupation, which benefited from the novel efficient multi-head attention (EMHA). In EMHA, a feature split module (FSM) is proposed to split the long sequence into sub-segments and then these sub-segments are applied by attention operation. This module can significantly decrease the GPU memory occupation. Extensive experiments show that our ESRT achieves competitive results. Compared with the original Transformer which occupies 16057M GPU memory, the proposed ET only occupies 4191M GPU memory with better performance.

READ FULL TEXT

page 1

page 3

page 4

page 6

page 9

page 10

research
04/28/2022

Lightweight Bimodal Network for Single-Image Super-Resolution via Symmetric CNN and Recursive Transformer

Single-image super-resolution (SISR) has achieved significant breakthrou...
research
07/06/2022

Cross-receptive Focused Inference Network for Lightweight Image Super-Resolution

With the development of deep learning, single image super-resolution (SI...
research
04/19/2022

Self-Calibrated Efficient Transformer for Lightweight Super-Resolution

Recently, deep learning has been successfully applied to the single-imag...
research
01/24/2023

Image Super-Resolution using Efficient Striped Window Transformer

Recently, transformer-based methods have made impressive progress in sin...
research
08/29/2020

Ultra Lightweight Image Super-Resolution with Multi-Attention Layers

Lightweight image super-resolution (SR) networks have the utmost signifi...
research
11/21/2022

N-Gram in Swin Transformers for Efficient Lightweight Image Super-Resolution

While some studies have proven that Swin Transformer (SwinT) with window...
research
12/29/2022

Efficient Image Super-Resolution with Feature Interaction Weighted Hybrid Network

Recently, great progress has been made in single-image super-resolution ...

Please sign up or login with your details

Forgot password? Click here to reset