Bi-level Dynamic Learning for Jointly Multi-modality Image Fusion and Beyond

05/11/2023
by   Zhu Liu, et al.
0

Recently, multi-modality scene perception tasks, e.g., image fusion and scene understanding, have attracted widespread attention for intelligent vision systems. However, early efforts always consider boosting a single task unilaterally and neglecting others, seldom investigating their underlying connections for joint promotion. To overcome these limitations, we establish the hierarchical dual tasks-driven deep model to bridge these tasks. Concretely, we firstly construct an image fusion module to fuse complementary characteristics and cascade dual task-related modules, including a discriminator for visual effects and a semantic network for feature measurement. We provide a bi-level perspective to formulate image fusion and follow-up downstream tasks. To incorporate distinct task-related responses for image fusion, we consider image fusion as a primary goal and dual modules as learnable constraints. Furthermore, we develop an efficient first-order approximation to compute corresponding gradients and present dynamic weighted aggregation to balance the gradients for fusion learning. Extensive experiments demonstrate the superiority of our method, which not only produces visually pleasant fused results but also realizes significant promotion for detection and segmentation than the state-of-the-art approaches.

READ FULL TEXT

page 2

page 3

page 4

page 5

page 6

page 7

research
08/04/2023

Multi-interactive Feature Learning and a Full-time Multi-modality Benchmark for Image Fusion and Segmentation

Multi-modality image fusion and segmentation play a vital role in autono...
research
05/19/2023

Equivariant Multi-Modality Image Fusion

Multi-modality image fusion is a technique used to combine information f...
research
05/25/2023

A Task-guided, Implicitly-searched and Meta-initialized Deep Model for Image Fusion

Image fusion plays a key role in a variety of multi-sensor-based vision ...
research
03/30/2022

Target-aware Dual Adversarial Learning and a Multi-scenario Multi-Modality Benchmark to Fuse Infrared and Visible for Object Detection

This study addresses the issue of fusing infrared and visible images tha...
research
12/18/2019

A Cross-Modal Image Fusion Theory Guided by Human Visual Characteristics

The characteristics of feature selection, nonlinear combination and mult...
research
07/08/2021

Multi-Modality Task Cascade for 3D Object Detection

Point clouds and RGB images are naturally complementary modalities for 3...
research
07/17/2020

AE-Net: Autonomous Evolution Image Fusion Method Inspired by Human Cognitive Mechanism

In order to solve the robustness and generality problems of the image fu...

Please sign up or login with your details

Forgot password? Click here to reset