Incomplete Multimodal Learning for Remote Sensing Data Fusion

04/22/2023
by   Yuxing Chen, et al.
0

The mechanism of connecting multimodal signals through self-attention operation is a key factor in the success of multimodal Transformer networks in remote sensing data fusion tasks. However, traditional approaches assume access to all modalities during both training and inference, which can lead to severe degradation when dealing with modal-incomplete inputs in downstream applications. To address this limitation, our proposed approach introduces a novel model for incomplete multimodal learning in the context of remote sensing data fusion. This approach can be used in both supervised and self-supervised pretraining paradigms and leverages the additional learned fusion tokens in combination with Bi-LSTM attention and masked self-attention mechanisms to collect multimodal signals. The proposed approach employs reconstruction and contrastive loss to facilitate fusion in pre-training while allowing for random modality combinations as inputs in network training. Our approach delivers state-of-the-art performance on two multimodal datasets for tasks such as building instance / semantic segmentation and land-cover mapping tasks when dealing with incomplete inputs during inference.

READ FULL TEXT

page 1

page 4

page 6

page 7

page 9

page 10

research
04/14/2021

Self-Supervised Learning of Remote Sensing Scene Representations Using Contrastive Multiview Coding

In recent years self-supervised learning has emerged as a promising cand...
research
01/29/2023

Supervised and Contrastive Self-Supervised In-Domain Representation Learning for Dense Prediction Problems in Remote Sensing

In recent years Convolutional neural networks (CNN) have made significan...
research
02/26/2022

Supervising Remote Sensing Change Detection Models with 3D Surface Semantics

Remote sensing change detection, identifying changes between scenes of t...
research
11/06/2021

Multi-modal land cover mapping of remote sensing images using pyramid attention and gated fusion networks

Multi-modality data is becoming readily available in remote sensing (RS)...
research
09/10/2020

Multimodal Noisy Segmentation based fragmented burn scars identification in Amazon Rainforest

Detection of burn marks due to wildfires in inaccessible rain forests is...
research
07/07/2023

General-Purpose Multimodal Transformer meets Remote Sensing Semantic Segmentation

The advent of high-resolution multispectral/hyperspectral sensors, LiDAR...
research
02/17/2023

Tensorized Optical Multimodal Fusion Network

We propose the first tensorized optical multimodal fusion network archit...

Please sign up or login with your details

Forgot password? Click here to reset