LAPFormer: A Light and Accurate Polyp Segmentation Transformer

10/10/2022
by   Mai Nguyen, et al.
0

Polyp segmentation is still known as a difficult problem due to the large variety of polyp shapes, scanning and labeling modalities. This prevents deep learning model to generalize well on unseen data. However, Transformer-based approach recently has achieved some remarkable results on performance with the ability of extracting global context better than CNN-based architecture and yet lead to better generalization. To leverage this strength of Transformer, we propose a new model with encoder-decoder architecture named LAPFormer, which uses a hierarchical Transformer encoder to better extract global feature and combine with our novel CNN (Convolutional Neural Network) decoder for capturing local appearance of the polyps. Our proposed decoder contains a progressive feature fusion module designed for fusing feature from upper scales and lower scales and enable multi-scale features to be more correlative. Besides, we also use feature refinement module and feature selection module for processing feature. We test our model on five popular benchmark datasets for polyp segmentation, including Kvasir, CVC-Clinic DB, CVC-ColonDB, CVC-T, and ETIS-Larib

READ FULL TEXT

page 3

page 5

research
03/26/2022

Feature Selective Transformer for Semantic Image Segmentation

Recently, it has attracted more and more attentions to fuse multi-scale ...
research
05/17/2022

ColonFormer: An Efficient Transformer based Method for Colon Polyp Segmentation

Identifying polyps is a challenging problem for automatic analysis of en...
research
03/07/2022

Stepwise Feature Fusion: Local Guides Global

Colonoscopy, currently the most efficient and recognized colon polyp det...
research
05/14/2022

Transformer Scale Gate for Semantic Segmentation

Effectively encoding multi-scale contextual information is crucial for a...
research
09/13/2022

DMTNet: Dynamic Multi-scale Network for Dual-pixel Images Defocus Deblurring with Transformer

Recent works achieve excellent results in defocus deblurring task based ...
research
05/02/2023

Exploring vision transformer layer choosing for semantic segmentation

Extensive work has demonstrated the effectiveness of Vision Transformers...
research
09/30/2022

An efficient encoder-decoder architecture with top-down attention for speech separation

Deep neural networks have shown excellent prospects in speech separation...

Please sign up or login with your details

Forgot password? Click here to reset