Bootstrap Representation Learning for Segmentation on Medical Volumes and Sequences

06/23/2021
by   Zejian Chen, et al.
0

In this work, we propose a novel straightforward method for medical volume and sequence segmentation with limited annotations. To avert laborious annotating, the recent success of self-supervised learning(SSL) motivates the pre-training on unlabeled data. Despite its success, it is still challenging to adapt typical SSL methods to volume/sequence segmentation, due to their lack of mining on local semantic discrimination and rare exploitation on volume and sequence structures. Based on the continuity between slices/frames and the common spatial layout of organs across volumes/sequences, we introduced a novel bootstrap self-supervised representation learning method by leveraging the predictable possibility of neighboring slices. At the core of our method is a simple and straightforward dense self-supervision on the predictions of local representations and a strategy of predicting locals based on global context, which enables stable and reliable supervision for both global and local representation mining among volumes. Specifically, we first proposed an asymmetric network with an attention-guided predictor to enforce distance-specific prediction and supervision on slices within and across volumes/sequences. Secondly, we introduced a novel prototype-based foreground-background calibration module to enhance representation consistency. The two parts are trained jointly on labeled and unlabeled data. When evaluated on three benchmark datasets of medical volumes and sequences, our model outperforms existing methods with a large margin of 4.5% DSC on ACDC, 1.7% on Prostate, and 2.3% on CAMUS. Intensive evaluations reveals the effectiveness and superiority of our method.

READ FULL TEXT

page 16

page 17

research
06/15/2023

Advancing Volumetric Medical Image Segmentation via Global-Local Masked Autoencoder

Masked autoencoder (MAE) has emerged as a promising self-supervised pret...
research
12/04/2022

Joint Self-Supervised Image-Volume Representation Learning with Intra-Inter Contrastive Clustering

Collecting large-scale medical datasets with fully annotated samples for...
research
06/16/2022

Volumetric Supervised Contrastive Learning for Seismic Semantic Segmentation

In seismic interpretation, pixel-level labels of various rock structures...
research
07/18/2021

A Positive/Unlabeled Approach for the Segmentation of Medical Sequences using Point-Wise Supervision

The ability to quickly annotate medical imaging data plays a critical ro...
research
11/09/2021

Metagenome2Vec: Building Contextualized Representations for Scalable Metagenome Analysis

Advances in next-generation metagenome sequencing have the potential to ...
research
10/27/2022

Federated Graph Representation Learning using Self-Supervision

Federated graph representation learning (FedGRL) brings the benefits of ...
research
07/13/2023

Explainable 2D Vision Models for 3D Medical Data

Training Artificial Intelligence (AI) models on three-dimensional image ...

Please sign up or login with your details

Forgot password? Click here to reset