Disentangle, align and fuse for multimodal and zero-shot image segmentation

11/11/2019
by   Agisilaos Chartsias, et al.
40

Magnetic resonance (MR) protocols rely on several sequences to properly assess pathology and organ status. Yet, despite advances in image analysis we tend to treat each sequence, here termed modality, in isolation. Taking advantage of the information shared between modalities (largely an organ's anatomy) is beneficial for multi-modality multi-input processing and learning. However, we must overcome inherent anatomical misregistrations and disparities in signal intensity across the modalities to claim this benefit. We present a method that offers improved segmentation accuracy of the modality of interest (over a single input model), by learning to leverage information present in other modalities, enabling semi-supervised and zero shot learning. Core to our method is learning a disentangled decomposition into anatomical and imaging factors. Shared anatomical factors from the different inputs are jointly processed and fused to extract more accurate segmentation masks. Image misregistrations are corrected with a Spatial Transformer Network, that non-linearly aligns the anatomical factors. The imaging factor captures signal intensity characteristics across different modality data, and is used for image reconstruction, enabling semi-supervised learning. Temporal and slice pairing between inputs are learned dynamically. We demonstrate applications in Late Gadolinium Enhanced (LGE) and Blood Oxygenation Level Dependent (BOLD) cardiac segmentation, as well as in T2 abdominal segmentation.

READ FULL TEXT

page 1

page 3

page 7

page 8

research
03/22/2019

Factorised Representation Learning in Cardiac Image Analysis

Typically, a medical image offers spatial information on the anatomy (an...
research
08/13/2020

Multi-Modality Pathology Segmentation Framework: Application to Cardiac Magnetic Resonance Images

Multi-sequence of cardiac magnetic resonance (CMR) images can provide co...
research
07/01/2022

SD-LayerNet: Semi-supervised retinal layer segmentation in OCT using disentangled representation with anatomical priors

Optical coherence tomography (OCT) is a non-invasive 3D modality widely ...
research
07/22/2020

Multi-modality imaging with structure-promoting regularisers

Imaging with multiple modalities or multiple channels is becoming increa...
research
03/19/2018

Factorised spatial representation learning: application in semi-supervised myocardial segmentation

The success and generalisation of deep learning algorithms heavily depen...
research
12/02/2020

Contour Transformer Network for One-shot Segmentation of Anatomical Structures

Accurate segmentation of anatomical structures is vital for medical imag...
research
07/04/2021

Controllable cardiac synthesis via disentangled anatomy arithmetic

Acquiring annotated data at scale with rare diseases or conditions remai...

Please sign up or login with your details

Forgot password? Click here to reset