Next Article in Journal
Objective Priors for Invariant e-Values in the Presence of Nuisance Parameters
Previous Article in Journal
Daily Streamflow of Argentine Rivers Analysis Using Information Theory Quantifiers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SharDif: Sharing and Differential Learning for Image Fusion

1
College of Aerospace Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China
2
Low Speed Aerodynamics Institute, China Aerodynamics Research and Development Center, Mianyang 621000, China
3
School of Computer and Software Engineering, Xihua University, Chengdu 610039, China
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(1), 57; https://doi.org/10.3390/e26010057
Submission received: 4 December 2023 / Revised: 6 January 2024 / Accepted: 8 January 2024 / Published: 9 January 2024
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
Image fusion is the generation of an informative image that contains complementary information from the original sensor images, such as texture details and attentional targets. Existing methods have designed a variety of feature extraction algorithms and fusion strategies to achieve image fusion. However, these methods ignore the extraction of common features in the original multi-source images. The point of view proposed in this paper is that image fusion is to retain, as much as possible, the useful shared features and complementary differential features of the original multi-source images. Shared and differential learning methods for infrared and visible light image fusion are proposed. An encoder with shared weights is used to extract shared common features contained in infrared and visible light images, and the other two encoder blocks are used to extract differential features of infrared images and visible light images, respectively. Effective learning of shared and differential features is achieved through weight sharing and loss functions. Then, the fusion of shared features and differential features is achieved via a weighted fusion strategy based on an entropy-weighted attention mechanism. The experimental results demonstrate the effectiveness of the proposed model with its algorithm. Compared with the-state-of-the-art methods, the significant advantage of the proposed method is that it retains the structural information of the original image and has better fusion accuracy and visual perception effect.

1. Introduction

In many monitoring fields, it is difficult for a single sensor to capture enough information required to meet the monitoring tasks [1]. Sensors in different wavebands (for example, infrared and visible light) have obvious advantages in monitoring the same scene. However, on the one hand, multiple sensors bring data storage challenges, and on the other hand, the image information contained by a single sensor is flawed. Taking infrared and visible light images as an example, infrared sensors reflect the radiation characteristics of foreground targets via thermal radiation imaging, but infrared images often lack structural and texture information. The visible light sensor describes the background details of the scene via light reflection, but it is greatly affected by changes in lighting and weather conditions [2]. Therefore, image fusion has become a popular research field [3]. Image fusion is to fuse the input images into one image. At the same time, the fused image contains all the information of the input images and can even generate more significant information.
According to different application scenarios, image fusion is mainly divided into multi-focus [4], multi-spectral [5], and medical images [6]. The most studied multi-spectral image fusion is the fusion of infrared and visible light images, which also includes the fusion of hyperspectral images in the field of remote sensing. The fused multi-focus image can clearly image the background and foreground at the same time. The fused image of the multi-spectral image contains imaging information of multiple spectra. The fused image of magnetic resonance imaging (MRI) and computed tomography (CT) can clearly see the soft tissue and the bones at the same time.
The two core tasks of image fusion are feature extraction and feature fusion strategies. The original images are transformed into the feature domain, where the fusion rules are designed for features fusing, and then the fused features are reconstructed back to the original pixel space to obtain the fused image. For feature extraction tasks, existing pioneer image fusion works are divided into two major categories: methods of artificially designed transformations and methods based on feature representation learning. Feature fusion strategies are also divided into two types: manual design and global optimization learning.
Artificially constructed feature extraction methods and feature fusion rules at all levels are the most intensively studied areas of image fusion. Since such methods do not require training and are completely unsupervised methods, they have good versatility. The main feature transformation methods include discrete wavelet transform (DWT) [7], shearlet transform [8], nonsubsampled contourlet transform [9], low-rank representation (LRR) [10], and bilateral filter [11]. The manually designed fusion rules mainly include maximum value, average value, and nuclear norm. The usual approach is base parts adopt the average fusion rule, and detail parts adopt the maximum fusion rule [12]. In the representation learning domain, the typical methods are based on sparse representation (SP) [8,13,14]. SP learns a single-layer common over-complete dictionary from the input images, then performs sparse representation of the input images respectively; fuses the sparse coefficients; and reconstructs the fused image. Deep learning has better representation learning capabilities than SP, and it has become a popular research point in the field of image fusion [2,4,15,16,17,18]. Methods based on deep learning first train a common encoder and decoder using a large number of images; use the encoder to extract features of the input images respectively; use a fusion rule to fuse these feature maps; and finally, the decoder is used to reconstruct the fusion image [15,19].
For the fusion of representation features, manually designed fusion rules lack interpretability. Some research works try to learn fusion rules by defining loss functions. Global optimization methods such as particle swarm [20], Grasshopper [21], and membrane computing [22] are used for fusion rule learning. Another class of methods learns the fused decision map via deep learning [23,24].
Therefore, the core of image fusion is to transform the image from the original pixel space to a feature representation space that is easy to fuse. After fusion is achieved in the new feature space, the fused image is obtained via inverse transformation. Judging from the current research trends in academia, it is difficult to develop a new method in the traditional field of multi-scale transformation, and methods based on deep learning are the current research hotspots. The current general idea based on deep learning methods is to implement feature extraction via an encoder. After fusing the features, image fusion is achieved via a decoder. The core problem of these methods is weak interpretability and lack of criteria for judging the quality of extracted features. To address this problem, combined with the goal of merging complementary information in image fusion, we propose a new image fusion method based on shared and difference learning. The key motivation is to extract shared common features and differentiated individual features from input multi-source images. The fusion of complementary information is achieved via maximum operation. The fused image is then reconstructed via a densely connected encoder.
The major contributions of this paper include the following aspects:
(1)
A novel shared and differential feature learning image fusion method is proposed. In this model, a channel with shared weights realizes the extraction of shared features from multi-source images. Separate channels are used to extract personalized differential features of the input images.
(2)
In the fusion stage, a weighted fusion strategy based on an entropy-weighted attention mechanism is proposed and used to fuse the learned shared and differential features, which makes the fused image have richer texture information.
(3)
A light enhanced dense module (EDB) is proposed to extract low-level semantic features, and BiFormer is used to extract high-level semantic features.
(4)
Experiments were conducted on three typical datasets. Compared with the state-of-the-art fusion method, our proposed method has better performance, with obvious advantages in information fidelity and visual saliency. It shows that our method is obviously competitive.

2. Related Work

This paper takes infrared and visible light image fusion as an application case, focusing on complementary feature extraction of input images. At the same time, the motivation of this paper is to use deep learning to extract interpretable features from input multi-source images. Therefore, in this section, the research progress of feature extraction methods is introduced. Some methods are also used for comparative analysis in the experiments.

2.1. Model-Based Feature Extraction Method

Performing pixel-level transformation on the input source images and extracting multi-scale features of the original images are hot spots in early research. The image features extracted by such methods have very nice interpretability. They decompose the input images into low-frequency (base) parts and high-frequency (detail) parts. The high-frequency part reflects the basic semantic information of the scene, and the high-frequency part reflects the target information in the background. The nonsubsampled contourlet transform (NSCT) [9] is a pioneer work that is used for image fusion. To combine the advantages of multi-scale and deep learning, Wang et al. [25] proposed an image fusion method based on a convolutional neural network and NSCT. MDLatLRR [26] is a baseline method in this field, which first performs a multi-level low-rank sparse decomposition of the input images, and then fuses the base and detail parts separately. Li et al. [27] performed norm optimization on the fused images of MDLatLRR to obtain more significant fused images. Gaussian difference is used for image fusion, which is simple, efficient, and versatile [28].

2.2. Generative-Based Methods

GAN-based methods attempt to use generative neural network models to generate fused images conditioned on inputting multi-source images [29]. DDcGAN [30] drives the deep neural network to learn complementary features to reconstruct the fused image based on a defined loss function. GAN-FM [31] introduces a full-scale skip-connected generator and Markovian discriminators, and Fusion-UDCGAN [32] adopts a U-Type densely connected generation adversarial network. AT-GAN [33] proposes a generative adversarial network with intensity attention modules and semantic transition modules. This type of method can obtain additional image enhancement effects based on the definition of the loss function.

2.3. Task-Driven Approach

Fusion to improve target segmentation accuracy in low-light environments is one of the current research hotspots. This type of method hopes that the fused image will have higher brightness and more prominent target contours. SCFusion [24] achieves target saliency enhancement via a mask of the target area. SGFusion [34] achieves saliency guidance in the fusion process through multi-tasking of target segmentation. TIM [35] proposes a constrained strategy to incorporate information from downstream tasks to guide the unsupervised learning process of image fusion. SOSMaskFuse [36] also uses a target segmentation mask to achieve target enhancement. PIAFusion [37] realizes image fusion under low light conditions. This type of method generally has an enhanced effect on the original multi-source images, and the fused image has better gradient and visual saliency. But, the consistency with the original images is poor.

2.4. Autoencoder-Based Methods

The fusion method based on the autoencoder and decoder believes that neurons have a stronger amplitude response to salient areas. Fu et al. [38] proposed a dual branch network encoder to learn richer features. DeepFuse [39] performs feature extraction on multiple channels and is used for multi-exposure image fusion. DenseFuse [40] introduces a dense block in the encoder to extract multi-scale features. FusionDN [41] also uses a densely connected network and defines a multi-task loss function. NestFuse [42] introduces a nest connection architecture and also introduces a spatial attention mechanism to enhance the salient features. RFNNest [43] proposes a residual fusion network and can better retain detailed features. PSFusion [44] presents a practical infrared and visible image fusion network based on progressive semantic injection and scene fidelity constraints and the fusion images have good visual appeal. CDDFuse [45] is inspired by multi-scale decomposition and uses neural networks to decompose images into basic parts and detailed parts. The fused image is reconstructed after fusing the two parts respectively. This method requires two stages of training.

3. Proposed Method

3.1. The Neural Network Model

Our proposed shared and differential feature learning image fusion model consists of encoders, a fusion layer, and decoders, as shown in Figure 1. The encoder part contains four channels, and the share feature encoders (SHE) share weights and thus have exactly the same parameters. The other two encoders extract the differential features of infrared and visible light respectively. In the fusion part, the fusion of shared features and differential features is achieved, and then the fused image is reconstructed via the fusion encoder. This is a multi-task architecture that constrains the encoder to extract the correct features via the reconstruction of the input image. The top branch implements auto-encoding reconstruction of infrared images, expressed as
I R = I R D I R D I F E ( I R ) S H E ( I R ) ,
where I R represents the reconstructed image, I R represents the input original images, I R D ( · ) represents the decoder of the infrared image, I R D I F E ( · ) represents the difference feature encoder of the infrared image, S H E ( · ) is the shared feature encoder, and ⊕ is a point-wise addition operator, which achieves the fusion of shared features and differential features of a single modality image. The middle branch generates the fused image, expressed as
F S = F S D M A X I R D I F E ( I R ) , V I D I F E ( V I ) M A X S H E ( I R ) , S H E ( V I ) ,
where F S is the final fused image, F S D ( · ) is the encoder of the fused image, V I D I F F E ( · ) is the visible light encoder, and v i represents the input visible light image. M A X ( · , · ) represents point-wise maximum fusion rules, which intuitively reflect the fusion of complementary characteristics of the input multi-source images. The bottom branch implements the decomposition and reconstruction task of visible light images, expressed as
V I = V I D V I D I F E ( VI ) S H E ( VI ) ,
where V I D ( · ) is the visible light image decoder.

3.1.1. Branch Structure

In order to better extract the shared and differential features of multi-source input images, we carefully designed the autoencoder branch structure based on existing research work, as shown in Figure 2. The encoder part is a multi-scale feature extraction module containing five layers. The first layer is a 1 1 convolutional layer. Next are two lightweight external attention rest block encoders (TRB-E). The feature map is downsampled by 2× before the second TRB-E. The last two layers consist of two BiFormer [46] modules. Ordinary convolution is responsible for extracting the low-level features of the input images, mainly the details and texture information of the image. BiFormer is responsible for extracting high-level semantic features and making up for the lack of convolution’s ability to extract global features. The decoder consists of multiple TRB-D blocks, skip links, and a normal convolution. The decoder used can better fuse the low-level detail information and high-level semantic information of the images. The obtained fused image is semantically consistent with the input images and has better texture detail information.

3.1.2. TRB Block

In order to better extract the features contained in the image, we developed the external attention rest block (TRB) based on a lightweight resnet block and external attention, as shown in Figure 3. TRB blocks contain two convolutions ( 3 3 ) and an external attention module. Among them, the external attention module is used to enhance the global information acquisition capability of rest blocks. Given an input feature map F R N × d , N is the number of pixels in the image, and d is the feature dimension. M k and M v are two memory units. External attention can be expressed as
A = ( α ) i , j = N o r m ( F M k T ) ,
F o u t = A M v ,
where ( α ) i , j is the similarity between the i-th element and the j-th row of matrix M, and M is a learnable parameter independent of the input, which will be used as the memory of the entire training set. A is the attention map derived from the learned prior knowledge.

3.1.3. Biformer Block

The transformer has advantages in capturing long-distance contextual dependencies and high-level semantic features. Biformer demonstrates competitiveness in a variety of visual tasks. We introduced a simplified Biformer module in the encoder, as shown in Figure 4. The input feature maps are first subjected to patch merging (PM), and then the relative position information is implicitly encoded via a 3 × 3 convolution. Then, layer normalization (LN), Bi-Level Routing Attention (BiRA), and a Multi-Layer Perceptron (MLP) are applied in sequence.
BiRA is the core of Biformer. It proposes a novel dynamic sparse attention via bi-level routing to achieve more flexible computing allocation and content awareness, as shown in Figure 4. For the input feature map f, the calculation process of BiRA is expressed as [46]
Q = f W q ,
K g = g a t h e r ( K , I r ) , V g = g a t h e r ( V , I r ) ,
O = A t t e n t i o n ( Q , K g , V g ) + L C E ( V ) ,
where W q is a projection weight for the query, I r is the routing index matrix, and L C E ( · ) is the local context enhancement term [47].

3.2. Fusion Strategy Based on Entropy-Weighted Attention Mechanism

After the complementary features of the source image are extracted by the encoder, a strategy is needed to fuse these features. We propose a new fusion method, which is an improvement of the attention fusion [5], which is called fusion strategy based on an entropy-weighted attention mechanism, as show in Figure 5. We apply the objective metric entropy to measure the amount of information in each source feature map on the basis of information theory [4,41,48]. The entropy of a feature map is defined as
E N ( ϕ ) = l = 1 L 1 p l l o g 2 p l ,
where L is the number of gray levels, and p l is the probability of the corresponding level. In the fusion layer, entropy-weighted spatial attention and channel attention mechanisms are used, and l 1 -norm and softmax operations are adopted in the spatial attention mechanism. The entropy-weighted spatial attention mechanism can be expressed as
Φ ^ f m ( x , y ) = k = 1 K β k m ( x , y ) Φ k m ( x , y ) ,
Φ ^ k m ( x , y ) = β k m ( x , y ) Φ k m ( x , y ) ,
β k m ( x , y ) = λ 1 × Φ k m ( x , y ) 1 i = 1 K Φ i m ( x , y ) 1 + ( 1 λ 1 ) × E N ( Φ k m ) i = 1 K E N ( Φ i m ) ,
where · represents l 1 n o r m ; k { 1 , . . . , K } , K = 2, ( x , y ) represents the corresponding position in the multi-scale deep feature and weighted graph, respectively; and Φ ^ k m represents β k m -weighted deep features. Global pooling and softmax operations are used in the channel attention mechanism. Further, three types of global pooling are used: average pooling, max pooling, and kernel pooling. The entropy-weighted channel attention mechanism can be expressed as
Φ ˜ f m ( x , y ) = i = 1 K α i m ( n ) Φ i m ( n ) ,
α k m ( n ) = λ 1 × α ¯ k m ( n ) i = 1 K α ¯ i m ( n ) + ( 1 λ 1 ) × E N ( Φ k m ) i = 1 K E N ( Φ i m ) ,
α ¯ k m ( n ) = P ( Φ k m ( n ) ) ,
where n { 1 , 2 } , n is the channel number corresponding to the feature Φ k m , and P ( · ) represents the pooling operation.

3.3. Loss Function

SharDifF is a multi-task network model with a total of three outputs, namely the reconstructed infrared image I R , the reconstructed visible light image V I , and the fused image F S . The encoder is constrained to extract reliable features via the reconstruction errors of I R and V I , and the fused image can retain the structural information of the original input images via structural similarity. Through the maximum gradient similarity loss, the fused image can retain the edge texture features of the original images as much as possible. The loss function of the entire model is expressed as
L t o t a l = λ 1 L i r + λ 2 L v i + λ 3 L f s + λ 4 L d i f ,
where λ 1 λ 4 are the weighting factors to balance the four losses. The four sub losses are described in detail below.

3.3.1. Fidelity Loss

The purpose of reconstructing the input multi-source images is to ensure that the extracted features can reflect the essential information of the input images without causing mode collapse. We call it fidelity loss, which includes L i r and L v i .
L i r = 1 H W I R I R 1 + α 1 S S I M ( I R , I R ) ,
L v i = 1 H W V I V I 1 + α 1 S S I M ( V I , V I ) ,
where · 1 is a norm operator, α is a balancing factor, and S S I M ( · , · ) is the calculation of the structural similarity between two images, which is defined as
L S S I M ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ x 2 + C 2 ) ,
where μ x and μ y are the averages of all pixels in the two source images; σ x and σ y are the variances of the pixel values of the two source images; and C 1 and C 2 are constants to ensure the stability of the function.

3.3.2. Fusion Loss

The fusion loss L f s is used to obtain high-quality fusion images, which consists of three parts: content loss, gradient loss, and structure loss.
L f s = L c o n + β 1 L g r a d + β 2 1 S S I M ( I R , I R ) ,
where β 1 and β 2 are balance factors, and L c o n represents the content loss of the fused image, expressed as
L c o n = 1 H W f s max ( I R , V I ) 1 ,
where m a x ( · , · ) represents the pixel-wise maximum, and L g r a d represents the gradient loss, which is expressed as
L g r a d = 1 H W f s max ( I R , V I ) 1 ,
where ∇ represents the gradient operator and · is the modulo operation. The fused image is consistent with the maximum gradients of the input multi-source images and has richer texture information.

4. Experiment and Result

4.1. Dataset

In this paper, three widely used datasets are used for experimental verification. The first dataset is TNO, whose images come from [49]. TNO contains a total of 41 pairs of infrared and visible images, of which a sample of these infrared and visible images is shown in Figure 6. The second dataset used in the paper is M3FD [50], of which this dataset contains multiple outdoor scenes, and we randomly selected 43 pairs of images from its test data for testing, as shown in Figure 7. The third dataset is MSRS [37]. This dataset contains a total of 1083 training samples. Fifty-two samples are also randomly selected from the test set to evaluate the algorithm performance, as shown in Figure 8. Since the data scale of TNO is small, the experimental models that need to be trained are uniformly trained using the MSRS training set. The trained models were used for testing on M3FD, MSRS, and TNO, respectively.

4.2. Methods for Comparison

In order to verify the effectiveness of the image fusion model SharDif proposed in this paper, we compare it with eight state-of-the-art image fusion methods. Optimization-based generation methods include DDcGAN [30]. Methods based on deep neural network feature representation learning include DenseFuse [40], RFNNest [43], FlFuse [51], PIAFusion [37], and PSFusion [44]. Model-based methods include GDFusion [28], and DDFM [52]. It should be noted that (1) All methods participating in the comparison use the weights and parameters given in the original papers to achieve the fusion of test data. (2) The fusion rule in denseFuse is “add”, and the fusion rule in RFNNest is “max”.

4.3. Evaluation Index

In order to quantitatively compare the competitiveness of various image fusion methods, nine metrics are used to compare various methods on the test set. These metrics we selected include metrics that reflect the visual consistency of images, such as mutual information (MI) [53], Q A B / F (petrovic metric) [54], Q E (piella metric) [55], and feature mutual information (FMI) [56]. Indicators reflecting information fidelity, such as IFC [57], and other comprehensive indicators, such as structural similarity index measure (SSIM) [58], peak signal-to-noise ratio (PSNR), Q0 [58], and entropy (EN) [28], are included.

4.4. Ablation Experiment

In this section, we first verify the effectiveness of our proposed ShareDif for image fusion on the test set MSRS. ShareDif uses a backbone network similar to RFNNest [43], which is named Model 1. On this basis, we propose the idea of shared and differential feature learning, which is named Model 2. In the feature extraction stage, we proposed the EDB module, which was named Model 3. BiFormer was introduced at the high level to extract high-level semantic features to obtain Model 4. The fusion strategy-based entropy-weighted attention mechanism is added to Model 4, and finally, the fusion model ShareDif proposed in this paper was obtained.
The experimental results of these models on MSRS are shown in Table 1. It can be seen from the results that the shared difference learning method proposed in this paper has significantly improved the performance of image fusion, especially compared with the original model, it has been improved in almost all evaluation indicators, especially MI, Q a b f , PSNR, SSIM, IFC, etc., have greatly been improved. This shows that by learning shared features and differential features, and then designing an appropriate fusion strategy, it is beneficial for the fused image to contain complementary features of the input multi-source images. The EDB module is introduced into the backbone model, especially BiFormer, which is used to extract advanced semantic features, so that the fused image and the multi-source input image have better information consistency. Several indicators such as MI, Q a b f , PSNR, and IFC have been significantly improved, indicating that the fusion image obtained by our method has better information fidelity.

4.5. Comparative Analysis of Experimental Results

4.5.1. Visual Quantitative Comparison

TNO is a widely used dataset, its images contain many scenes, and the image resolutions also varies greatly. This dataset is small in size and has no training samples, which can reflect the generalization ability and universality of the algorithms. Six pairs of test samples were arbitrarily selected for subjective evaluation of visual effects. The fusion results of all methods are shown in Figure 9. It can be seen from the figure that the fusion results of the three methods, DenseFuse, RFN-Nest, and FIFuse are similar, and all have the disadvantages of weak image contrast and overall poor image quality. These three methods all use the COCO dataset to train the network model. They do not consider the differences in extraction features of different modal images. The advantage is that they do not require infrared and visible light image pairs for training. The disadvantage is that the fusion effect is not good. The generation-based method DDcGan can highlight the target, but the image is obviously distorted and the fused image is blurry and loses a lot of details. Both PIAFusion and PSFusion methods enhance the input multi-source images, which can improve the fusion effect under low illumination conditions. However, the fusion results of these two methods are less consistent with the source images. The advantage of the traditional method GDFusion is that it does not require training. From the fusion effect, the image consistency is poor and the target is not prominent. The target contrast of the fused image obtained using the DDFM method is low, and the overall visual saliency of the image is poor. Our method is trained on MSRS and the model can be well transferred to TNO data, indicating that ShareDif can extract the essential features of input multi-source images and complete the fusion. From the fusion effect point of view, the fusion result of ShareDif has a more obvious target. The overall image looks natural and has good consistency with the input original images.
The MSRS dataset is large in scale and is mainly taken in dark environments, which is more in line with the practical application scenarios of infrared and visible light fusion. It is also a widely used dataset in recent years. Six pairs of test samples were also selected for experiments, and the fusion results of all methods are shown in Figure 10. Through observation, the targets in the fused image of FLFuse are dim and the clarity of fused images is the worst. DDcGan has a significant enhancement effect on the image, and the overall brightness is very high, but the image is also seriously distorted. The results of DenseFuse, RFN-Nest, and DDFM are relatively similar, and the brightness of cars and pedestrians in the fused image is poor. Due to edge enhancement, GDFusion’s fused image has serious noise. The fusion results of PIAFusion and PSFusion are relatively close, and some areas in the fused image have brightness oversaturation. ShareDif’s fused image has moderate brightness, and the image quality is in line with human visual imaging characteristics.
M3FD has a larger number of images and contains a wider variety of scenes. Its visible light image has high brightness and serious noise. The test results of six samples are shown in Figure 11. Consistent with the previous conclusion, the image contrast and clarity of the fusion results of DenseFuse, RFN-Nest, FLFuse, and DDFM are insufficient. DDcGan, PIAFusion, and PSFusion significantly enhance images, and some areas are severely oversaturated, which affects the visual effects of the images. The results of ShareDif have the best visual effect and the target is more prominent, especially the last column of images. The fused images conform to human visual habits.

4.5.2. Quantitative Comparisons

Fusion image quality evaluation is still a problem worthy of study [18,36], and different measurement indicators reflect different characteristics of fusion images. For example, Mutual information (MI) indicates how many features are preserved in fused images. Q a b f reflects the quality of visual information obtained from the fusion of the input images. SSIM is a widely used metric which models the loss and distortion between two images according to their similarities in light, contrast, and structure information. The larger value of EN means that there is more information contained. On the other hand, EN can be affected by noise easily. The reason why multiple different metrics are needed to quantitatively compare the fusion results is that these metrics will not give consistent results. Therefore, in this paper, we selected nine indicators to compare the algorithms. In Table 2, Table 3 and Table 4, each specific metric is the average of all images in the test set.
The objective evaluation results of TNO are shown in Table 2. The fusion results of the convolutional neural network model DenseFuse and FLFuse trained using the COCO dataset have the best PSNR and SSIM. The results of DDcGan and DDFM reflect their weak information fidelity capabilities, and the fused image is greatly different from the multi-source input image. The indicators of PSFusion, GDFusion, FLFuse, and RFN-Nest are relatively close. PIA has good reviews and ranks second. For MI and IFC, our method ShareDif has very obvious advantages, indicating that the fused image of our method can greatly maintain the original information of multi-source input images.
Table 3 shows the test results of MSRS. As can be seen from the table, the SSIM indicators of DenseFuse, FLFuse, PIAFusion, and GDFusion are relatively close, indicating that they all pay more attention to the fusion of image texture edges. The fidelity capabilities of DDcGan and DDFM are still the worst. The results of PIAFusion and PSFusion are relatively close and have better Q a b f . ShareDif has a leading position in multiple indicators, especially MI and IFC, which is much higher than the second-ranked one, and its advantages are obvious.
The evaluation results of M3FD are shown in Table 4. Three methods, including DenseFuse, RFN-Nest, and FLFuse, have the best SSIM. PIAFusion and ShareDif have clear leading advantages. ShareDif significantly outperforms the remaining methods for PSNR. The results on three different datasets can reflect the generalization performance of the algorithm. Our method ShareDif has the strongest information fidelity and has obvious advantages on all three datasets.

5. Conclusions

In this paper, an intuitively motivated image fusion method is proposed. Since different multi-source images contain complementary information, the goal of image fusion is to retain the complementary information of the multi-source input images in the resulting image as much as possible. We propose an image fusion model of shared and differential learning, which can achieve information extraction from multi-source input images. The shared weight branch is used to extract common information from the input images, and the other channels are used to learn the differentiated features of the input images. Based on the semantic decomposition of the input images, the complementary information of the original images is fused. Our fusion model only requires one training session to optimize the model. Extensive experiments have been performed to demonstrate the proposed method’s superiority, including texture and edge preservation, illumination adaptation, and comprehensive performance in visual effects and metrics. In the future, similar ideas can be considered for multi-focus and medical image fusion, as well as primary image processing such as denoising and super resolution.

Author Contributions

Conceptualization, Z.G.; methodology, L.L.; software, L.L.; validation, Z.G.; formal analysis, Z.G.; investigation, L.L.; resources, L.L.; data curation, L.L.; writing—original draft preparation, L.L.; writing—review and editing, Z.G.; visualization, L.L.; supervision, Z.G.; project administration, L.L.; funding acquisition, Z.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been partially supported by the Sichuan Science and Technology Program (grant nos. 2023YFG0300, 2023YFQ0044, 2022YFG0095).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We thank anonymous reviewers and academic editors for their valuable comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, X. Benchmarking and comparing multi-exposure image fusion algorithms. Inf. Fusion 2021, 74, 111–131. [Google Scholar] [CrossRef]
  2. Wang, Z.; Wu, Y.; Wang, J.; Xu, J.; Shao, W. Res2Fusion: Infrared and visible image fusion based on dense Res2net and double nonlocal attention models. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  3. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  4. Zhang, X. Deep learning-based multi-focus image fusion: A survey and a comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4819–4838. [Google Scholar] [CrossRef] [PubMed]
  5. Yilmaz, C.S.; Yilmaz, V.; Gungor, O. A theoretical and practical survey of image fusion methods for multispectral pansharpening. Inf. Fusion 2022, 79, 1–43. [Google Scholar] [CrossRef]
  6. Hermessi, H.; Mourali, O.; Zagrouba, E. Multimodal medical image fusion review: Theoretical background and recent advances. Signal Process. 2021, 183, 108036. [Google Scholar] [CrossRef]
  7. Ben Hamza, A.; He, Y.; Krim, H.; Willsky, A. A multiscale approach to pixel-level image fusion. Integr. Comput. Aided Eng. 2005, 12, 135–146. [Google Scholar] [CrossRef]
  8. Gao, Z.; Yang, M.; Xie, C. Space target image fusion method based on image clarity criterion. Opt. Eng. 2017, 56, 053102. [Google Scholar] [CrossRef]
  9. Da Cunha, A.L.; Zhou, J.; Do, M.N. The nonsubsampled contourlet transform: Theory, design, and applications. IEEE Trans. Image Process. 2006, 15, 3089–3101. [Google Scholar] [CrossRef]
  10. Li, H.; Wu, X.J. Multi-focus image fusion using dictionary learning and low-rank representation. In Proceedings of the Image and Graphics: 9th International Conference, ICIG 2017, Shanghai, China, 13–15 September 2017; Revised Selected Papers, Part I 9. Springer: Cham, Switzerland, 2017; pp. 675–686. [Google Scholar]
  11. Li, X.; Zhou, F.; Tan, H.; Zhang, W.; Zhao, C. Multimodal medical image fusion based on joint bilateral filter and local gradient energy. Inf. Sci. 2021, 569, 302–325. [Google Scholar] [CrossRef]
  12. Liu, Y.; Wang, L.; Cheng, J.; Li, C.; Chen, X. Multi-focus image fusion: A survey of the state of the art. Inf. Fusion 2020, 64, 71–91. [Google Scholar] [CrossRef]
  13. Zhang, Q.; Liu, Y.; Blum, R.S.; Han, J.; Tao, D. Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: A review. Inf. Fusion 2018, 40, 57–75. [Google Scholar] [CrossRef]
  14. Yang, B.; Li, S. Multifocus image fusion and restoration with sparse representation. IEEE Trans. Instrum. Meas. 2009, 59, 884–892. [Google Scholar] [CrossRef]
  15. Zhang, H.; Xu, H.; Tian, X.; Jiang, J.; Ma, J. Image fusion meets deep learning: A survey and perspective. Inf. Fusion 2021, 76, 323–336. [Google Scholar] [CrossRef]
  16. Sun, H.; Xiao, W. Similarity Weight Learning: A New Spatial and Temporal Satellite Image Fusion Framework. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
  17. Xiao, Y.; Guo, Z.; Veelaert, P.; Philips, W. DMDN: Degradation model-based deep network for multi-focus image fusion. Signal Process. Image Commun. 2022, 101, 116554. [Google Scholar] [CrossRef]
  18. Wang, W.; Fu, X.; Zeng, W.; Sun, L.; Zhan, R.; Huang, Y.; Ding, X. Enhanced deep blind hyperspectral image fusion. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 1513–1523. [Google Scholar] [CrossRef]
  19. Zhang, Y.; Liu, Y.; Sun, P.; Yan, H.; Zhao, X.; Zhang, L. IFCNN: A general image fusion framework based on convolutional neural network. Inf. Fusion 2020, 54, 99–118. [Google Scholar] [CrossRef]
  20. Shehanaz, S.; Daniel, E.; Guntur, S.R.; Satrasupalli, S. Optimum weighted multimodal medical image fusion using particle swarm optimization. Optik 2021, 231, 166413. [Google Scholar] [CrossRef]
  21. Dinh, P.H. A novel approach based on grasshopper optimization algorithm for medical image fusion. Expert Syst. Appl. 2021, 171, 114576. [Google Scholar] [CrossRef]
  22. Li, B.; Peng, H.; Luo, X.; Wang, J.; Song, X.; Pérez-Jiménez, M.J.; Riscos-Núñez, A. Medical image fusion method based on coupled neural P systems in nonsubsampled shearlet transform domain. Int. J. Neural Syst. 2021, 31, 2050050. [Google Scholar] [CrossRef]
  23. Ma, B.; Yin, X.; Wu, D.; Shen, H.; Ban, X.; Wang, Y. End-to-end learning for simultaneously generating decision map and multi-focus image fusion result. Neurocomputing 2022, 470, 204–216. [Google Scholar] [CrossRef]
  24. Liu, H.; Ma, M.; Wang, M.; Chen, Z.; Zhao, Y. SCFusion: Infrared and visible fusion based on salient compensation. Entropy 2023, 25, 985. [Google Scholar] [CrossRef] [PubMed]
  25. Wang, Z.; Li, X.; Duan, H.; Su, Y.; Zhang, X.; Guan, X. Medical image fusion based on convolutional neural networks and non-subsampled contourlet transform. Expert Syst. Appl. 2021, 171, 114574. [Google Scholar] [CrossRef]
  26. Li, H.; Wu, X.J.; Kittler, J. MDLatLRR: A novel decomposition method for infrared and visible image fusion. IEEE Trans. Image Process. 2020, 29, 4733–4746. [Google Scholar] [CrossRef]
  27. Li, G.; Lin, Y.; Qu, X. An infrared and visible image fusion method based on multi-scale transformation and norm optimization. Inf. Fusion 2021, 71, 109–129. [Google Scholar] [CrossRef]
  28. Kurban, R. Gaussian of Differences: A Simple and Efficient General Image Fusion Method. Entropy 2023, 25, 1215. [Google Scholar] [CrossRef]
  29. Zhou, T.; Li, Q.; Lu, H.; Cheng, Q.; Zhang, X. GAN review: Models and medical image fusion applications. Inf. Fusion 2023, 91, 134–148. [Google Scholar] [CrossRef]
  30. Ma, J.; Xu, H.; Jiang, J.; Mei, X.; Zhang, X.P. DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. 2020, 29, 4980–4995. [Google Scholar] [CrossRef]
  31. Zhang, H.; Yuan, J.; Tian, X.; Ma, J. GAN-FM: Infrared and visible image fusion using GAN with full-scale skip connection and dual Markovian discriminators. IEEE Trans. Comput. Imaging 2021, 7, 1134–1147. [Google Scholar] [CrossRef]
  32. Gao, Y.; Ma, S.; Liu, J.; Xiu, X. Fusion-UDCGAN: Multifocus image fusion via a U-type densely connected generation adversarial network. IEEE Trans. Instrum. Meas. 2022, 71, 1–13. [Google Scholar] [CrossRef]
  33. Rao, Y.; Wu, D.; Han, M.; Wang, T.; Yang, Y.; Lei, T.; Zhou, C.; Bai, H.; Xing, L. AT-GAN: A generative adversarial network with attention and transition for infrared and visible image fusion. Inf. Fusion 2023, 92, 336–349. [Google Scholar] [CrossRef]
  34. Liu, J.; Dian, R.; Li, S.; Liu, H. SGFusion: A saliency guided deep-learning framework for pixel-level image fusion. Inf. Fusion 2023, 91, 205–214. [Google Scholar] [CrossRef]
  35. Liu, R.; Liu, Z.; Liu, J.; Fan, X.; Luo, Z. A Task-guided, Implicitly-searched and Meta-initialized Deep Model for Image Fusion. arXiv 2023, arXiv:2305.15862. [Google Scholar]
  36. Li, G.; Qian, X.; Qu, X. SOSMaskFuse: An Infrared and Visible Image Fusion Architecture Based on Salient Object Segmentation Mask. IEEE Trans. Intell. Transp. Syst. 2023, 24, 10118–10137. [Google Scholar] [CrossRef]
  37. Tang, L.; Yuan, J.; Zhang, H.; Jiang, X.; Ma, J. PIAFusion: A progressive infrared and visible image fusion network based on illumination aware. Inf. Fusion 2022, 83, 79–92. [Google Scholar] [CrossRef]
  38. Fu, Y.; Wu, X.J. A dual-branch network for infrared and visible image fusion. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Yichang, China, 16–18 September 2021; pp. 10675–10680. [Google Scholar]
  39. Ram Prabhakar, K.; Sai Srikar, V.; Venkatesh Babu, R. Deepfuse: A deep unsupervised approach for exposure fusion with extreme exposure image pairs. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4714–4722. [Google Scholar]
  40. Li, H.; Wu, X.J. DenseFuse: A fusion approach to infrared and visible images. IEEE Trans. Image Process. 2018, 28, 2614–2623. [Google Scholar] [CrossRef]
  41. Xu, H.; Ma, J.; Le, Z.; Jiang, J.; Guo, X. Fusiondn: A unified densely connected network for image fusion. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12484–12491. [Google Scholar]
  42. Li, H.; Wu, X.J.; Durrani, T. NestFuse: An infrared and visible image fusion architecture based on nest connection and spatial/channel attention models. IEEE Trans. Instrum. Meas. 2020, 69, 9645–9656. [Google Scholar] [CrossRef]
  43. Li, H.; Wu, X.J.; Kittler, J. RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf. Fusion 2021, 73, 72–86. [Google Scholar] [CrossRef]
  44. Tang, L.; Zhang, H.; Xu, H.; Ma, J. Rethinking the necessity of image fusion in high-level vision tasks: A practical infrared and visible image fusion network based on progressive semantic injection and scene fidelity. Inf. Fusion 2023, 101870. [Google Scholar] [CrossRef]
  45. Zhao, Z.; Bai, H.; Zhang, J.; Zhang, Y.; Xu, S.; Lin, Z.; Timofte, R.; Van Gool, L. Cddfuse: Correlation-driven dual-branch feature decomposition for multi-modality image fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 5906–5916. [Google Scholar]
  46. Zhu, L.; Wang, X.; Ke, Z.; Zhang, W.; Lau, R.W. BiFormer: Vision Transformer with Bi-Level Routing Attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 10323–10333. [Google Scholar]
  47. Ren, S.; Zhou, D.; He, S.; Feng, J.; Wang, X. Shunted self-attention via multi-scale token aggregation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10853–10862. [Google Scholar]
  48. Liu, Y.; Yang, X.; Zhang, R.; Albertini, M.K.; Celik, T.; Jeon, G. Entropy-Based Image Fusion with Joint Sparse Representation and Rolling Guidance Filter. Entropy 2020, 22, 118. [Google Scholar] [CrossRef] [PubMed]
  49. Toet, A. TNO Image Fusion Dataset. 2014. Available online: https://figshare.com/articles/TN_Image_Fusion_Dataset/1008029. (accessed on 31 September 2023).
  50. Liu, J.; Fan, X.; Huang, Z.; Wu, G.; Liu, R.; Zhong, W.; Luo, Z. Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5802–5811. [Google Scholar]
  51. Xue, W.; Wang, A.; Zhao, L. FLFuse-Net: A fast and lightweight infrared and visible image fusion network via feature flow and edge compensation for salient information. Infrared Phys. Technol. 2022, 127, 104383. [Google Scholar] [CrossRef]
  52. Zhao, Z.; Bai, H.; Zhu, Y.; Zhang, J.; Xu, S.; Zhang, Y.; Zhang, K.; Meng, D.; Timofte, R.; Van Gool, L. DDFM: Denoising diffusion model for multi-modality image fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 8082–8093. [Google Scholar]
  53. Qu, G.; Zhang, D.; Yan, P. Information measure for performance of image fusion. Electron. Lett. 2002, 38, 1. [Google Scholar] [CrossRef]
  54. Xydeas, C.S.; Petrovic, V. Objective image fusion performance measure. Electron. Lett. 2000, 36, 308–309. [Google Scholar] [CrossRef]
  55. Piella, G.; Heijmans, H. A new quality metric for image fusion. In Proceedings of the 2003 International Conference on Image Processing (Cat. No. 03CH37429), Barcelona, Spain, 14–17 September 2003; Volume 3, p. III–173. [Google Scholar]
  56. Haghighat, M.B.A.; Aghagolzadeh, A.; Seyedarabi, H. A non-reference image fusion metric based on mutual information of image features. Comput. Electr. Eng. 2011, 37, 744–756. [Google Scholar] [CrossRef]
  57. Sheikh, H.R.; Bovik, A.C.; De Veciana, G. An information fidelity criterion for image quality assessment using natural scene statistics. IEEE Trans. Image Process. 2005, 14, 2117–2128. [Google Scholar] [CrossRef]
  58. Wang, Z.; Bovik, A.C. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
Figure 1. The overall framework for ShareDif.
Figure 1. The overall framework for ShareDif.
Entropy 26 00057 g001
Figure 2. The structure of the encoder and decoder branches.
Figure 2. The structure of the encoder and decoder branches.
Entropy 26 00057 g002
Figure 3. The structure of the TRB block.
Figure 3. The structure of the TRB block.
Entropy 26 00057 g003
Figure 4. The structure of the Biformer block.
Figure 4. The structure of the Biformer block.
Entropy 26 00057 g004
Figure 5. The structure of the fusion block.
Figure 5. The structure of the fusion block.
Entropy 26 00057 g005
Figure 6. Some test samples from TNO. The top row contains infrared images, and the second row contains visible images.
Figure 6. Some test samples from TNO. The top row contains infrared images, and the second row contains visible images.
Entropy 26 00057 g006
Figure 7. Some test samples from M3FD. The top row contains infrared images, and the second row contains visible images.
Figure 7. Some test samples from M3FD. The top row contains infrared images, and the second row contains visible images.
Entropy 26 00057 g007
Figure 8. Some test samples from MSRS. The top row contains infrared images, and the second row contains visible images.
Figure 8. Some test samples from MSRS. The top row contains infrared images, and the second row contains visible images.
Entropy 26 00057 g008
Figure 9. Qualitative comparison of all methods on six typical image pairs from TNO. From top to bottom: fusion results of DenseFuse [40], RFNNest [43], DDcGan [30], FLFuse [51], PIAFusion [37], PSFusion [44], GDFusion [28], DDFM [52], and our ShareDif.
Figure 9. Qualitative comparison of all methods on six typical image pairs from TNO. From top to bottom: fusion results of DenseFuse [40], RFNNest [43], DDcGan [30], FLFuse [51], PIAFusion [37], PSFusion [44], GDFusion [28], DDFM [52], and our ShareDif.
Entropy 26 00057 g009
Figure 10. Qualitative comparison of all methods on six typical image pairs from MSRS. From top to bottom: fusion results of DenseFuse [40], RFNNest [43], DDcGan [30], FLFuse [51], PIAFusion [37], PSFusion [44], GDFusion [28], DDFM [52], and our ShareDif.
Figure 10. Qualitative comparison of all methods on six typical image pairs from MSRS. From top to bottom: fusion results of DenseFuse [40], RFNNest [43], DDcGan [30], FLFuse [51], PIAFusion [37], PSFusion [44], GDFusion [28], DDFM [52], and our ShareDif.
Entropy 26 00057 g010
Figure 11. Qualitative comparison of all methods on six typical image pairs from M3FD. From top to bottom: fusion results of DenseFuse [40], RFNNest [43], DDcGan [30], FLFuse [51], PIAFusion [37], PSFusion [44], GDFusion [28], DDFM [52], and our ShareDif.
Figure 11. Qualitative comparison of all methods on six typical image pairs from M3FD. From top to bottom: fusion results of DenseFuse [40], RFNNest [43], DDcGan [30], FLFuse [51], PIAFusion [37], PSFusion [44], GDFusion [28], DDFM [52], and our ShareDif.
Entropy 26 00057 g011
Table 1. Ablation experiment results on MSRS dataset.
Table 1. Ablation experiment results on MSRS dataset.
MetricENMI Q a b f Q 0 Q e PSNRFMISSIMIFC
Model 15.972.730.2940.3030.46413.750.420.5415.94
Model 26.354.340.6370.4610.46719.650.4610.7148.90
Model 36.614.550.6760.4770.47219.710.4710.7229.67
Model 46.634.730.6810.4820.47520.500.4720.72010.12
ShareDif6.654.710.6830.4910.48220.570.4750.72110.17
Table 2. Quantitative Comparisons of all methods on TNO (The bold is best).
Table 2. Quantitative Comparisons of all methods on TNO (The bold is best).
MetricENMI Q a b f Q 0 Q e PSNRFMISSIMIFC
DenseFuse (2018)6.352.210.360.470.2816.000.430.745.09
RFNNest (2021)6.422.670.320.290.2913.140.440.584.29
DDcGan (2020)7.371.840.360.360.1613.040.410.594.42
FLFuse (2022)6.362.170.410.490.2816.270.420.745.60
PIAFusion (2022)6.903.160.530.490.3514.600.440.697.70
PSFusion (2023)7.262.290.530.440.3214.260.420.645.76
GDFusion (2023)6.612.060.500.450.2515.730.430.725.28
DDFM (2023)6.841.740.260.230.1014.400.250.562.06
ShareDif7.083.650.540.480.3614.610.440.708.42
Table 3. Quantitative Comparisons of all methods on MSRS (The bold is best).
Table 3. Quantitative Comparisons of all methods on MSRS (The bold is best).
MetricENMI Q a b f Q 0 Q e PSNRFMISSIMIFC
DenseFuse (2018)5.962.670.370.450.3818.910.430.745.28
RFNNest (2021)5.972.730.290.300.4713.760.430.545.94
DDcGan (2020)7.241.880.380.240.209.770.370.393.57
FLFuse (2022)5.782.140.290.420.3119.250.430.725.08
PIAFusion (2022)6.643.990.670.480.4520.420.460.718.31
PSFusion (2023)6.872.920.680.460.4718.140.460.687.14
GDFusion (2023)6.362.310.620.440.3818.350.430.715.70
DDFM (2023)6.142.350.420.390.3218.770.320.694.09
ShareDif6.654.710.690.490.4820.570.480.7210.17
Table 4. Quantitative Comparisons of all methods on M3FD (The bold is best).
Table 4. Quantitative Comparisons of all methods on M3FD (The bold is best).
MetricENMI Q a b f Q 0 Q e PSNRFMISSIMIFC
DenseFuse (2018)6.432.890.380.480.4316.270.430.766.13
RFNNest (2021)6.442.450.400.460.3516.760.380.785.38
DDcGan (2020)7.412.540.490.370.3413.650.420.645.32
FLFuse (2022)6.513.200.270.460.3316.590.410.755.50
PIAFusion (2022)6.834.190.640.490.5116.410.460.719.60
PSFusion (2023)7.392.740.580.420.4613.390.420.656.51
GDFusion (2023)6.792.090.610.370.4215.910.390.705.93
DDFM (2023)6.682.770.440.420.4016.010.350.725.07
ShareDif6.764.560.640.490.5217.180.470.7310.21
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liang, L.; Gao, Z. SharDif: Sharing and Differential Learning for Image Fusion. Entropy 2024, 26, 57. https://doi.org/10.3390/e26010057

AMA Style

Liang L, Gao Z. SharDif: Sharing and Differential Learning for Image Fusion. Entropy. 2024; 26(1):57. https://doi.org/10.3390/e26010057

Chicago/Turabian Style

Liang, Lei, and Zhisheng Gao. 2024. "SharDif: Sharing and Differential Learning for Image Fusion" Entropy 26, no. 1: 57. https://doi.org/10.3390/e26010057

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop