DS-TransUNet:Dual Swin Transformer U-Net for Medical Image Segmentation

06/12/2021
by   Ailiang Lin, et al.
0

Automatic medical image segmentation has made great progress benefit from the development of deep learning. However, most existing methods are based on convolutional neural networks (CNNs), which fail to build long-range dependencies and global context connections due to the limitation of receptive field in convolution operation. Inspired by the success of Transformer in modeling the long-range contextual information, some researchers have expended considerable efforts in designing the robust variants of Transformer-based U-Net. Moreover, the patch division used in vision transformers usually ignores the pixel-level intrinsic structural features inside each patch. To alleviate these problems, we propose a novel deep medical image segmentation framework called Dual Swin Transformer U-Net (DS-TransUNet), which might be the first attempt to concurrently incorporate the advantages of hierarchical Swin Transformer into both encoder and decoder of the standard U-shaped architecture to enhance the semantic segmentation quality of varying medical images. Unlike many prior Transformer-based solutions, the proposed DS-TransUNet first adopts dual-scale encoder subnetworks based on Swin Transformer to extract the coarse and fine-grained feature representations of different semantic scales. As the core component for our DS-TransUNet, a well-designed Transformer Interactive Fusion (TIF) module is proposed to effectively establish global dependencies between features of different scales through the self-attention mechanism. Furthermore, we also introduce the Swin Transformer block into decoder to further explore the long-range contextual information during the up-sampling process. Extensive experiments across four typical tasks for medical image segmentation demonstrate the effectiveness of DS-TransUNet, and show that our approach significantly outperforms the state-of-the-art methods.

READ FULL TEXT

page 1

page 4

page 7

page 8

research
07/12/2021

TransAttUnet: Multi-level Attention-guided U-Net with Transformer for Medical Image Segmentation

With the development of deep encoder-decoder architectures and large-sca...
research
04/10/2023

HST-MRF: Heterogeneous Swin Transformer with Multi-Receptive Field for Medical Image Segmentation

The Transformer has been successfully used in medical image segmentation...
research
12/21/2022

DuAT: Dual-Aggregation Transformer Network for Medical Image Segmentation

Transformer-based models have been widely demonstrated to be successful ...
research
03/08/2023

UT-Net: Combining U-Net and Transformer for Joint Optic Disc and Cup Segmentation and Glaucoma Detection

Glaucoma is a chronic visual disease that may cause permanent irreversib...
research
03/02/2022

Contextual Attention Network: Transformer Meets U-Net

Currently, convolutional neural networks (CNN) (e.g., U-Net) have become...
research
04/03/2023

U-Netmer: U-Net meets Transformer for medical image segmentation

The combination of the U-Net based deep learning models and Transformer ...
research
07/11/2023

3D Medical Image Segmentation based on multi-scale MPU-Net

The high cure rate of cancer is inextricably linked to physicians' accur...

Please sign up or login with your details

Forgot password? Click here to reset