Juggling With Representations: On the Information Transfer Between Imagery, Point Clouds, and Meshes for Multi-Modal Semantics

03/12/2021
by   Dominik Laupheimer, et al.
22

The automatic semantic segmentation of the huge amount of acquired remote sensing data has become an important task in the last decade. Images and Point Clouds (PCs) are fundamental data representations, particularly in urban mapping applications. Textured 3D meshes integrate both data representations geometrically by wiring the PC and texturing the surface elements with available imagery. We present a mesh-centered holistic geometry-driven methodology that explicitly integrates entities of imagery, PC and mesh. Due to its integrative character, we choose the mesh as the core representation that also helps to solve the visibility problem for points in imagery. Utilizing the proposed multi-modal fusion as the backbone and considering the established entity relationships, we enable the sharing of information across the modalities imagery, PC and mesh in a two-fold manner: (i) feature transfer and (ii) label transfer. By these means, we achieve to enrich feature vectors to multi-modal feature vectors for each representation. Concurrently, we achieve to label all representations consistently while reducing the manual label effort to a single representation. Consequently, we facilitate to train machine learning algorithms and to semantically segment any of these data representations - both in a multi-modal and single-modal sense. The paper presents the association mechanism and the subsequent information transfer, which we believe are cornerstones for multi-modal scene analysis. Furthermore, we discuss the preconditions and limitations of the presented approach in detail. We demonstrate the effectiveness of our methodology on the ISPRS 3D semantic labeling contest (Vaihingen 3D) and a proprietary data set (Hessigheim 3D).

READ FULL TEXT

page 2

page 15

page 20

page 25

page 26

page 32

page 33

page 36

research
05/02/2018

Multi-Resolution Multi-Modal Sensor Fusion For Remote Sensing Data With Label Uncertainty

In remote sensing, each sensor can provide complementary or reinforcing ...
research
07/24/2018

Feature Fusion through Multitask CNN for Large-scale Remote Sensing Image Segmentation

In recent years, Fully Convolutional Networks (FCN) has been widely used...
research
03/28/2020

Inferring Semantic Information with 3D Neural Scene Representations

Biological vision infers multi-modal 3D representations that support rea...
research
04/09/2019

CMIR-NET : A Deep Learning Based Model For Cross-Modal Retrieval In Remote Sensing

We address the problem of cross-modal information retrieval in the domai...
research
05/11/2023

Multi-modal Multi-level Fusion for 3D Single Object Tracking

3D single object tracking plays a crucial role in computer vision. Mains...
research
02/10/2021

H3D: Benchmark on Semantic Segmentation of High-Resolution 3D Point Clouds and textured Meshes from UAV LiDAR and Multi-View-Stereo

Automated semantic segmentation and object detection are of great import...
research
06/19/2023

UniG3D: A Unified 3D Object Generation Dataset

The field of generative AI has a transformative impact on various areas,...

Please sign up or login with your details

Forgot password? Click here to reset