CDDFuse: Correlation-Driven Dual-Branch Feature Decomposition for Multi-Modality Image Fusion

11/26/2022
by   Zixiang Zhao, et al.
0

Multi-modality (MM) image fusion aims to render fused images that maintain the merits of different modalities, e.g., functional highlight and detailed textures. To tackle the challenge in modeling cross-modality features and decomposing desirable modality-specific and modality-shared features, we propose a novel Correlation-Driven feature Decomposition Fusion (CDDFuse) network for end-to-end MM feature decomposition and image fusion. In the first stage of the two-stage architectures, CDDFuse uses Restormer blocks to extract cross-modality shallow features. We then introduce a dual-branch Transformer-CNN feature extractor with Lite Transformer (LT) blocks leveraging long-range attention to handle low-frequency global features and Invertible Neural Networks (INN) blocks focusing on extracting high-frequency local information. Upon the embedded semantic information, the low-frequency features should be correlated while the high-frequency features should be uncorrelated. Thus, we propose a correlation-driven loss for better feature decomposition. In the second stage, the LT-based global fusion and INN-based local fusion layers output the fused image. Extensive experiments demonstrate that our CDDFuse achieves promising results in multiple fusion tasks, including infrared-visible image fusion and medical image fusion. We also show that CDDFuse can boost the performance in downstream infrared-visible semantic segmentation and object detection in a unified benchmark.

READ FULL TEXT

page 1

page 4

page 6

page 8

research
02/21/2021

A Deep Decomposition Network for Image Processing: A Case Study for Visible and Infrared Image Fusion

Image decomposition is a crucial subject in the field of image processin...
research
08/07/2023

Learning a Graph Neural Network with Cross Modality Interaction for Image Fusion

Infrared and visible image fusion has gradually proved to be a vital for...
research
05/19/2023

Equivariant Multi-Modality Image Fusion

Multi-modality image fusion is a technique used to combine information f...
research
03/23/2021

Generalizing Face Forgery Detection with High-frequency Features

Current face forgery detection methods achieve high accuracy under the w...
research
11/19/2021

DSPoint: Dual-scale Point Cloud Recognition with High-frequency Fusion

Point cloud processing is a challenging task due to its sparsity and irr...
research
11/21/2022

TFormer: A throughout fusion transformer for multi-modal skin lesion diagnosis

Multi-modal skin lesion diagnosis (MSLD) has achieved remarkable success...
research
11/20/2022

CoCoNet: Coupled Contrastive Learning Network with Multi-level Feature Ensemble for Multi-modality Image Fusion

Infrared and visible image fusion targets to provide an informative imag...

Please sign up or login with your details

Forgot password? Click here to reset