Next Article in Journal
A Virtual Sensing Concept for Nitrogen and Phosphorus Monitoring Using Machine Learning Techniques
Next Article in Special Issue
Contribution of Geometric Feature Analysis for Deep Learning Classification Algorithms of Urban LiDAR Data
Previous Article in Journal
The Use of Soil Moisture and Pore-Water Pressure Sensors for the Interpretation of Landslide Behavior in Small-Scale Physical Models
Previous Article in Special Issue
A Novel Remote Sensing Image Registration Algorithm Based on Feature Using ProbNet-RANSAC
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Remote Sensing Image Fusion Based on Morphological Convolutional Neural Networks with Information Entropy for Optimal Scale

1
School of Computer and Control Engineering, Yantai University, Yantai 264005, China
2
School of Information Science and Technology, Hainan Normal University, Haikou 571158, China
3
School of Information Science and Engineering, University of Jinan, Jinan 250024, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(19), 7339; https://doi.org/10.3390/s22197339
Submission received: 17 August 2022 / Revised: 22 September 2022 / Accepted: 23 September 2022 / Published: 27 September 2022
(This article belongs to the Special Issue State-of-the-Art Multimodal Remote Sensing Technologies)

Abstract

:
Remote sensing image fusion is a fundamental issue in the field of remote sensing. In this paper, we propose a remote sensing image fusion method based on optimal scale morphological convolutional neural networks (CNN) using the principle of entropy from information theory. We use an attentional CNN to fuse the optimal cartoon and texture components of the original images to obtain a high-resolution multispectral image. We obtain the cartoon and texture components using sparse decomposition-morphological component analysis (MCA) with an optimal threshold value determined by calculating the information entropy of the fused image. In the sparse decomposition process, the local discrete cosine transform dictionary and the curvelet transform dictionary compose the MCA dictionary. We sparsely decompose the original remote sensing images into a texture component and a cartoon component at an optimal scale using the information entropy to control the dictionary parameter. Experimental results show that the remote sensing image fusion method proposed in this paper can effectively retain the information of the original image, improve the spatial resolution and spectral fidelity, and provide a new idea for image fusion from the perspective of multi-morphological deep learning.

1. Introduction

Due to the limitations of satellite technology, most remote sensing images can only be panchromatic (PAN) images and low-resolution multispectral (LRMS) images of the same area. The goal of remote sensing image fusion is to fuse the spectral information of LRMS images and the spatial information of PAN images to generate a remote sensing image with both high spatial resolution and high spectral resolution [1]. Classical component substitution (CS) [2] methods are the most widely used, but they often result in spectral distortion. Multiresolution analysis (MRA) [3] methods are also often utilized. Compared with the CS method, methods based on MRA retain the spectral information better, but the spatial details are seriously lost. Model-based [4] methods have also been applied to remote sensing image fusion. The aforementioned methods can effectively reduce spectral distortion, but usually lead to blurred results.
The popular convolutional neural networks (CNN) method can learn the correlation between PAN images and LRMS images because of its excellent nonlinear expression and achieves better fusion results than traditional remote sensing image fusion methods [5,6]. Therefore, many existing fusion methods choose to combine traditional methods with deep learning methods [7,8,9] and have achieved good results. However, one of the basic tasks of image analysis and computer vision is to extract different features of an image. Most of the existing deep learning fusion methods treat the source image as a single component without considering the diversity of image components, thus ignoring the different morphological details in the source image. Remote sensing image usually contain spectral information and spatial structure, among which the PAN image reflects the spatial distribution information and structure information of the image. The texture component of the PAN image contains the image surface information and its relationship with the surrounding environment, which can better reflect the spatial structure information of the PAN image. The boundary of the cartoon component of remote sensing image is smoother and the spectral information is retained, so the spectral information of the LRMS image can be completely characterized by its cartoon component, and the redundancy and noise can be filtered out.
Morphological component analysis (MCA), proposed by J. Starck et al. [10,11], has been used to solve problems such as image decomposition [12], image denoising [13], and image restoration [14]. The main idea of this algorithm is to associate each morphological component in the data with a dictionary of atoms. Each component of the image is assumed to correspond to a suitable dictionary enabling the sparsest representation vector. The sparse vector is reconstructed according to the corresponding dictionary to obtain the separated image components.
Therefore, in this paper, we propose a method combining the sparse decomposition-multi-scale MCA method and CNN for remote sensing image fusion, with optimal scale determined by information entropy. We use MCA to sparsely decompose the original images and acquire the texture components and cartoon components at multi-scale. Considering the variability of the different components of the image, we use information entropy to calculate the threshold of the decomposition parameters. This facilitates the extraction of the different components at the optimal scale and effectively acquires more detail from the image. We use the spectral and spatial information of the LRMS and PAN images, respectively, to input the cartoon component of the LRMS remote sensing image and the texture component of the PAN image into an attentional CNN for fusion. The remainder of this paper is organized as follows. Section 2 describes the multi-scale MCA method. Section 3 details the fusion network and displays multi-scale fusion results. Section 4 provides the overall experimental results and analysis. Finally, Section 5 concludes this research.

2. Multi-Scale MCA Algorithm

2.1. Image Decomposition via MCA

We represent an image as f = u + v , where u is the cartoon component of f , which is smooth and contains the geometric feature information of the image. v represents the texture component of the image and is the high-frequency part of the image. Decomposing an image into cartoon and texture components is essential for many applications. MCA joins two transform bases to sparsely decompose the image, and the joint local discrete cosine transform (LDCT) and curvelet transform (CT) are used as MCA decomposition dictionary: D = D 1 , D 2 . This enables the extraction of the texture components and cartoon components of the image, where D 1 represents the LDCT dictionary and D 2 represents the CT dictionary.
Assuming that the remote sensing image contains only the texture component X T , the LDCT dictionary D 1 can sparsely represent the texture image. The Equation for solving the texture sparse coefficient is as follows:
α P A N T = Arg min α T α T 0                               s u b j e c t   t o : X T = D 1 α T
where u 0 denotes the l 0   norm that effectively calculates the number of non-zero entries in the vector X T and α T is the coefficient for the dictionary representation. The LDCT dictionary D 1 represents the non-texture components in the image as zeros, maximizing the sparseness. The dictionary D 1 is sparse with respect to the texture components of the image but not sparse to the cartoon components of the image. Thus, the texture components of the remote sensing image are obtained using the above model.
Similarly, for a remote sensing image X C that contains only cartoon components, the image is represented by the CT dictionary D 2 , which is sparse only with respect to cartoon components. The equation is as follows:
α M S C = Arg   min α C α C 0                               s u b j e c t   t o : X C = D 2 α C
where α C is the coefficient for the dictionary. Using the CT dictionary D 2 , the non-cartoon elements in the image are represented as zeros. Because the CT dictionary only represents sparse cartoon components, this model extracts the cartoon components in a remote sensing image.
According to the above model, for any remote sensing image X containing both texture and cartoon components, it is necessary to decompose the components with the joint decomposition dictionary D containing both dictionary D 1 and dictionary D 2 , posing the following regularization problem:
α P A N T , α M S C = Arg   min α T , α C α T 0 + α C 0                     s u b j e c t   t o : X = D 1 α T + D 2 α C
To better retain fused image information, we analysis the morphological components of the PAN image with a single channel and the MS image with three channels, obtaining the texture components of the PAN image and cartoon components of the MS image. Equations (4) and (5) show the sparse decomposition of the PAN image and MS image, respectively:
X P A N = X P A N T + X P A N C = D 1 α P A N T + D 2 α P A N C = D α P A N T + α P A N C
X M S = X M S T + X M S C = D 1 α M S T + D 2 α M S C = D α M S T + α M S C
where α P A N T , α P A N C , α M S T , and α M S C represent the corresponding decomposition coefficients. X P A N T and X P A N C are texture and cartoon components of the PAN image, respectively. X M S T and X M S C are texture and cartoon components of the MS image, respectively.

2.2. Decomposition with Different Scales

The existing MCA method uses a single scale [15], while humans analyze remote sensing images with complex components at multi-scale. This inspires the analysis of the image at multi-scale for morphological components, and the decomposition of the remote sensing image into texture and cartoon components at multi-scale.
Different MCA decomposition parameters represent different scales, and different scales also represent different resolutions. As shown in Figure 1 and Figure 2, we decompose the MS and PAN images into cartoon and texture components at different scales, and we set five decomposition parameters with 16/512, 32/512, 64/512, 128/512, and 256/512. Figure 1 and Figure 2 show that the cartoon component of the MS image and the texture component of the PAN image are decomposed at different scales (resolutions) with different parameters.
As shown in Figure 1 and Figure 2, the image components at different scales are not the same. Figure 1 indicates that a small threshold value removes too many edge details from the MS image, resulting in side effects such as noise, ultimately causing spectral distortion of the fused image. Figure 2 indicates that a large threshold value removes too many texture details from the PAN image, resulting in insufficient component information, ultimately causing noise in the fused image. Our target is to preserve details, remove redundant information and noise, and effectively retain texture and cartoon components. Therefore, controlling the parameter thresholds to construct a multi-scale dictionary is essential to achieve sparse multi-scale component decomposition.

2.3. Information Entropy Metric

Information entropy reflects the amount of information contained in an image at a certain position [16,17]. The threshold value of the control parameter is calculated using information entropy to retain the rich amount of information contained in the image while eliminating irrelevant information. This facilitates morphological component decomposition at multi-scale and selects the fusion results at the optimal scale.
In our previous work [18], we assume that T and C are the two images to be fused, the joint information entropy of the fused images can be expressed as H T , C . The conditional information entropy can be expressed as H T / C and H C / T , and the mutual information entropy is M T ; C , representing the redundant information (repeated content) between T and C . Then, the relationship between them can be expressed as Equation (6) [19]. The relationship between the information entropy of the two input source images is also described in Figure 3.
H T , C   = H T C + H C T + M T ; C
The ideal fusion goal of image T and image C is that the information entropy of the fused image is H T , C . However, in the actual fusion process, in addition to the redundant information M T ; C , other noise and interference may also exist, affecting the fusion results. Figure 4 expresses the relationship between noisy image T and noisy image C . Thus, considering noise, the remote sensing image fusion process ideally maintains the maximum joint information entropy of the input source image.
Based on the above analysis, assuming that F R N × N represents the fused image of size N × N pixels, we first average the RGB values of the three channels in the same pixel position and convert the color image into a gray image. Then, the image is classified into L gray levels. f i denotes the gray value of the pixel with spatial index i in the image, where f i G L = 0 , 1 , , L 1 . Based on the theory of information entropy, f i ˜ is the mean gray value over the neighborhood of the fused image. The neighborhood mean gray value composes the spatial feature vector of the gray distribution and can form a feature binary group with the pixel gray values of the image ( f i , f i ˜ ). The comprehensive feature X f i , f i ˜ of the gray value and the gray distribution of surrounding pixels is expressed as:
X f i , f i ˜ = g f i , f i ˜ N 2
where g f i , f i ˜ represents the number of occurrences of a single pixel feature binary group at a certain position. Combined with the two-dimensional information entropy of the image, Equation (8) calculates the entropy value of the final fused image F .
H F = L 1 i = 0 X f i , f i ˜ log X f i , f i ˜
The information entropy H F of the image at different fusion scales is calculated by Equation (8) to gain the amount of the information of the fused image and utilize to determine the optimal fusion threshold.

2.4. Multi-Scale Spatial Attention Module

Selective visual attention enables humans to quickly locate salient objects in complex visual scenes, inspiring the development of algorithms based on human attention mechanisms [20]. In the field of deep learning, the attention mechanism can be seen as a weighted combination of input feature mappings, where the weights depend on the similarity between the input elements. Spatial attention is used to determine the location salient information in a target image. For the remote sensing image with complex structures, the lack of spatial structure leads to inaccurate positioning, with different weights between different regions of the same channel. Spatial attention is calculated using Equation (9).
M S F = σ f 5 × 5 A v e P o o l F ; M a x P o o l F = σ f 5 × 5 F a v g s ; F m a x s
where σ denotes the sigmoid activation function, F denotes the feature map, and A v e P o o l · and M a x P o o l · denote average pooling and maximum pooling, respectively. f 5 × 5 denotes a convolution operation with a 5 × 5 pixel kernel. In this paper, we add a spatial attention module under each scale to enhance the information interaction in space and to strengthen the focus on valid information along the spatial dimension. The structure of multi-scale spatial attention is denoted by the dotted box in Figure 5b.

3. Methods

The proposed method is mainly composed of three parts, including MCA, feature extraction and feature fusion respectively. Firstly, the PAN image and the MS image are decomposed by MCA, the multi-scale texture components of PAN image and the multi-scale cartoon components of LRMS image are obtained. The spectral and spatial information are preserved while the redundancy and noise are removed. As shown in Figure 5a, the feature extraction network module is composed of two branches cascade convolution layers, which extract spectral features and spatial features obtained by MCA, respectively. Then, feature fusion network is used to generate the MS image with high spatial resolution. Finally, the optimal fusion scale is judged by information entropy theory, so as to get the high-resolution multispectral (HRMS) image under the optimal scale.

3.1. Network Setup and Multi-Scale Fusion

3.1.1. Algorithm Flow

Figure 6 shows the flow chart of the proposed fusion method; the detailed steps are as follows:
  • A The joint LDCT dictionary D 1 and CT dictionary D 2 form the decomposition dictionary D = D 1 , D 2 , and MCA is performed on the input source images PAN and MS at multi-scale to extract the texture component and cartoon components, respectively.
  • The threshold values of the parameters are calculated using the information entropy of the fused image from Step 3 to select the best extraction scale for the cartoon component X M S C of the MS image and the texture component X P A N T of the PAN image.
  • The optimal-scale cartoon component X M S C and texture component X P A N T are fused by the attentional CNN to produce the final fused image.

3.1.2. Network Structure

P A N i ,   j and M S i ,   j are the corresponding pixels of the PAN image and MS image at position i ,   j , respectively. T i ,   j and C i , j are the pixels at the corresponding points of the texture component and the cartoon component, respectively. The fused image F is obtained by calculating the fused pixels F i ,   j . Let N T i ,   j and N C i ,   j be the neighboring pixel points of T i ,   j and C i ,   j , respectively. The texture component and cartoon component are through a 3 × 3 pixel convolution kernel to calculate N T and N C , respectively. Then, these neighboring pixels pass through a 1 × 1 pixel convolution kernel to obtain the fused image F .
Figure 5 shows the overall network model. The entire fusion network comprises 10 convolutional layers, where six convolutional layers X F u s i o n i T and X F u s i o n i C ( i = 1 ,   2 ,   3 ) have convolutional kernels of size 3 × 3 pixels and the remaining convolutional layers have convolutional kernels of size 1 × 1 pixel. After each linear convolution operation, we incorporate the Leaky ReLU (LReLU) activation function to further improve the fused image. The convolution operations are expressed in Equation (10).
F = LReLU X w
where X represents the input to the convolution. w is the convolution kernel and LReLU X = max 0 , x is the nonlinear activation function.
In the fusion network, X F u s i o n 1   T C represents the fused image of the cartoon component X F u s i o n C and the texture component X F u s i o n T after weighted averaging. The computation process involves integrating the cartoon component and the texture component to construct the new image X F u s i o n 1 T C and then applying the convolution operation. Unlike X F u s i o n 1 T C , the inputs of X F u s i o n 2 T C , X F u s i o n 3 T C , and X F u s i o n 4 T C all contain three feature maps. For example, we obtain X F u s i o n 4 T C by concatenating X F u s i o n 3 T , X F u s i o n 3 C , and X F u s i o n 3 T C and then convolving them, where X F u s i o n 1 T , X F u s i o n 2 T , and X F u s i o n 3 T have the same number of feature maps as X F u s i o n 1 C , X F u s i o n 2 C , and X F u s i o n 3 C (32, 64, and 128, respectively). Similarly, X F u s i o n 1 T C , X F u s i o n 2 T C , and X F u s i o n 3 T C have 32, 64, and 128 feature maps, respectively.
Let T k k = 1 , 2 , 3 and C k k = 1 , 2 , 3 denote the output of the kth convolutional layer in the two branches network of the texture component and the cartoon component, respectively. T k and C k are calculated using Equations (11) and (12), respectively.
T k i , j , c h = LReLU T k 1 i , j , c h w T c h k
C k i , j , c h = LReLU C k 1 i , j , c h w C c h k
where c h is the channel index. w T c h k and w C c h k denote the kth layer of convolution kernels for the texture component and the cartoon component, respectively.
In the fusion network, the fusion results of the previous layer are referred to in the convolution operation of each layer. For each pixel in the final fused image, we can choose to increase the size of its convolution kernel or use a deeper network model to expand the area of its corresponding pixel in the original image to improve the fusion ability of the network model.

3.1.3. Different Scale Fusion Results

Section 2.3 details the selection of the parameter 128/512 as the optimal fusion scale of a set of remote sensing data. To corroborate that the fusion scale is optimal, we use the proposed attentional CNN to fuse the cartoon component extracted from the MS image and the texture component extracted from the PAN image at different scales, and Figure 7 shows the corresponding fusion results. The Figure 7 confirms that using the 128/512 decomposition parameter yields the fewest artifacts and superior fusion results. Furthermore, the information entropy diagram in Figure 8 also proves that the optimal fusion result is obtained by using the parameter 128/512.
Figure 8 and Table 1 show the information entropy of the fused image F calculated by Equation (8). The details present in the fused results are different at different scales, and the calculation results show that the effective information contained in the image reaches saturation using the 128/512 scale parameters, and the spectral and spatial information of the source image is well preserved while removing part of the redundancy and noise. This is because small-scale decomposition parameters lose too much information from the original image, while an overly large scale does not increase the effective amount of information because of redundancy and noise.

4. Results and Discussion

4.1. Model Training

We use a regression model to train the fusion function: f u s i o n = F P A N , M S , using the l 2 paradigm as the loss function, as expressed in Equation (13).
L θ = 1 n n i = 1 I F u s i o n ( θ ; P A N , M S ) 2
where I is the original image from the training set, P A N represents a PAN image, and M S is a low-resolution multispectral image. F u s i o n θ ; P A N , M S is the fusion function of the model output and the number of training samples is denoted by n. To solve the fusion function F u s i o n , we need to minimize the I . The pixel values of the image range from 0 255 and are normalized to the interval 0 ,   1 before being input to the model.
Adam’s algorithm [21], an adaptive learning rate optimization algorithm of stochastic gradient descent, is used as the optimization algorithm of our model. The initial learning rate of the model was set to 0.001 and divided by 10 at 50% and 75% of the total number of training phases. The training took 50 min per cycle and we trained for eight cycles. The final training mean squared deviation of the model was 0.00017.

4.2. Experimental Data

To assess the effectiveness of the proposed method, we conducted experiments on four sets of remote sensing images with different topographical areas. The first set of experimental data (Figure 9a,b) is obtained by the SPOT-6 satellite, which captures PAN images with a spatial resolution of 1.5 m and MS images with a spatial resolution of 6 m. Figure 10 shows the histogram of the evaluation indexes of each experimental result of the first set of experimental data. The second set of experimental data (Figure 11a,b) is obtained by the WorldView-2 satellite, which captures PAN images with a spatial resolution of 0.5 m and MS images with a spatial resolution of 2 m. Figure 12 shows the histogram of the evaluation indexes of each experimental result of the second set of experimental data. The third set of experimental data (Figure 13a,b) are MS images with a resolution of 19.5 m from the China-Brazil Earth Resources Satellite (CBERS) image and PAN images with a resolution of 15 m from the Landsat ETM+ image. The test area is located in Doumen District, Zhuhai City, Guangdong Province, including agricultural land, water bodies and forest land. Figure 14 shows the histogram of the evaluation indexes of each experimental result of the third set of experimental data The last set of experimental data (Figure 15a,b) are MS images with 4 m resolution and PAN images with 1 m resolution from IKONOS images. The experimental area is located in Beijing Normal University, and includes a playground, vegetation, and buildings. Figure 16 shows the histogram of the evaluation indexes of each experimental result of the last set of experimental data

4.3. Evaluation Indexes

We use Figure 9a and Figure 11a as reference images to objectively verify the performance of different fusion methods in the first and second groups of experiments. We use four objective evaluation indexes to evaluate the experimental results: correlation coefficient (CC) [22], root mean square error (RMSE) [23], relative dimensionless global error synthesis (ERGAS) [22], and peak signal to noise ratio (PSNR) [24].
CC reflects the correlation between two images, and a larger correlation parameter indicates more similarity between two images.
CC I H , I W = i = 1 M j = 1 N ( I H i , j I H ¯ I W i , j I W ¯ i = 1 M j = 1 N ( I H i , j I H ¯ ) 2 × i = 1 M j = 1 N ( I W i , j I W ¯ ) 2
Among them, I H , I W represent the pixels of the fused image and the ideal reference image respectively. I H ¯ , I W ¯ represent the average of pixels. The ideal CC value is 1.
RMSE is the difference between the pixel values of the fused image and the reference image. The ideal value of RMSE is 0.
RMSE I H , I W = 1 M N i = 1 M j = 1 N ( I H i , j ( I W i , j ) 2
The spectral and spatial quality of the fused image is evaluated using the ERGAS algorithm.
ERGAS = 100 h l 1 L l = 1 L RMSE l u l 2
where h and l represent the resolution of PAN image and MS image respectively. L is the number of bands. u l is the mean value of the original MS band l . A smaller value indicates a higher quality fused image, and the ideal value is 0.
PSNR reflects the degree of noise and distortion level of the image.
PSNR = 10 × log 10 2 N 1 2 1 M N i = 1 M j = 1 N ( I H i , j ( I W i , j ) 2
The high value of PSNR indicates that the fused image is closer to the reference image and therefor of higher quality.
For the third and fourth groups of experiments, we use the following three common objective evaluation indexes to evaluate the experimental results: quality without reference (QNR) index [25], and two components D λ and D s to quantify the spectral distortion and spatial distortion, respectively [26].
D λ = 2 C C 1 i = 1 C j > 1 C Q I i , ^ I j ^ Q I i L M , I j L M
D S = 1 C i = 1 C Q I i , ^ P Q I i L M , P L M
where I L M represents the LRMS image and C represents the number of bands. I ^ indicates the HRMS image, and P indicates the PAN image. Q denotes the Q-index.
QNR = 1 D λ α 1 D S β
where α and β are usually set to 1. The ideal value of QNR is 1, and the ideal value of D λ and D s is 0.

4.4. Experimental Results

The experimental results compare our proposed approach with Brovey [27], GS [28], IHS [29], ATWT [30], PCA [31], DWT [32], PanNet [33], FCNN [34], and PNN [35]. For our method, we use the calculated optimal fusion threshold to obtain the final experimental results. Figure 9, Figure 11, Figure 13, and Figure 15, respectively, show the experimental results of different satellite data.
Figure 9. Fusion results for the first group of remote sensing image data.
Figure 9. Fusion results for the first group of remote sensing image data.
Sensors 22 07339 g009
Figure 9 shows the fusion results for the first set of data. As can be seen from Figure 9c–h, although the fusion images obtained by the traditional methods have high spatial resolution, the spectral color is too saturated and there is a large area of spectral distortion. Figure 9i,k show the fusion results of the two deep learning methods, with varying degrees of spectral distortion and low spatial resolution. The spectral distribution of landmarks and other parts in Figure 9j and the method in this paper (Figure 9l) are more uniform, and the color effect is closer to the spectral information of MS images. However, in comparison, our method better reflects the high-frequency detail features. In addition, in the wheat field and other large areas where the spectral information is relatively close, the effect of our method is optimal. Table 2 and Figure 10 display the evaluation indexes for the first set of data fusion results, where the bold numbers indicate the best score for each evaluation indexes. Compared with the other seven methods, our method achieves better results for all of the evaluation indexes. These quantitative results, in conjunction with the subjective visual results in Figure 9, show that our method outperforms existing fusion methods.
Figure 10. Histogram results of the first group of evaluation indexes.
Figure 10. Histogram results of the first group of evaluation indexes.
Sensors 22 07339 g010
Figure 11 shows the fusion results on the second set of data, which mainly contains mountains and vegetation. The traditional methods (Figure 11c,h) result in different degrees of distortion in vegetation color with over-brightness or darkness compared with the original MS image. Compared with the traditional methods, the deep learning methods used to obtain Figure 11i,k achieve better spectral quality but not high spatial quality. The spectral information of the vegetation part in Figure 11j does not reflect the obvious difference between light and dark, the edge of the mountain is not smooth enough, and the spatial resolution is not as good as that of the method in this paper (Figure 11l). These results combined with the evaluation indexes in Table 3 and Figure 12 show that our fusion results are superior.
Figure 11. Fusion results for the second group of remote sensing image data.
Figure 11. Fusion results for the second group of remote sensing image data.
Sensors 22 07339 g011
Figure 12. Histogram results of the second group of evaluation indexes.
Figure 12. Histogram results of the second group of evaluation indexes.
Sensors 22 07339 g012
Figure 13 shows the fusion results of different fusion methods on the third group of remote sensing images. Because of the relatively close resolution of images from this group of data sources, the optimal fusion scale is also different from the first two groups of experiments. All of the methods improve the quality of the fused images to some extent compared with the input PAN images and MS images. However, the fusion results of the traditional methods all show spectral distortion compared with the deep learning methods. It is clear from Figure 13c–h that both the mountainous part in the upper left corner and the vegetation part in the lower right corner exhibit more pronounced spectral distortion compared with the original multispectral image. Figure 13j and our method both retain the spectral and spatial information of the input image more completely, but our proposed fusion method still outperforms FCNN in terms of spatial information retention. Table 4 and Figure 14 list the third set of objective evaluation indexes. The bold numbers in Table 4 indicate the best value for each evaluation index. With the exception of the D λ metric, our method obtains the best results.
Figure 13. Fusion results for the third group of remote sensing image data.
Figure 13. Fusion results for the third group of remote sensing image data.
Sensors 22 07339 g013aSensors 22 07339 g013b
Figure 14. Histogram result of the third group of evaluation indexes.
Figure 14. Histogram result of the third group of evaluation indexes.
Sensors 22 07339 g014
Figure 15 shows the fusion results of different fusion methods on the last set of remote sensing images. Because one of the two football fields in this geographical location is a real turf and the other is artificial turf, there are some differences between the two football fields in the input source images. Figure 15c,h retain better spatial resolution in the building area, but have more severe spectral distortion, obtaining too dark and too bright spectra, respectively. Figure 15i,k present the same problems, The D λ index in Figure 15i also reached the best value, but its spatial resolution was very low, and the overall image appeared blurred. Figure 15j has a higher spatial resolution but still has some shortcomings in terms of spectral preservation compared with our method (Figure 15i). In terms of subjective visual effects, our method outperforms the other algorithms in terms of spectral preservation and texture detail. Table 5 and Figure 16 present the evaluation indexes for the fourth set of data, where the bold numbers indicate the best value for each evaluation index. Although our method does not obtain the best D λ metric, combined with the subjective visual results in Figure 15, our proposed algorithm outperforms the other fusion methods overall, especially in terms of spatial resolution.
Figure 15. Fusion results for the fourth group of remote sensing image data.
Figure 15. Fusion results for the fourth group of remote sensing image data.
Sensors 22 07339 g015
Figure 16. Histogram result of the fourth group of evaluation indexes.
Figure 16. Histogram result of the fourth group of evaluation indexes.
Sensors 22 07339 g016

5. Conclusions

In this paper, we propose a remote sensing image fusion method using morphological convolutional neural networks with information entropy for optimal scale. Our method extracts the texture and cartoon components of remote sensing images at multi-scale using MCA and selects the best scale using information entropy theory. The spectral and spatial information of the input image is fully utilized while avoiding information loss. In the network design stage, we obtain the final fusion result using an attentional convolutional neural network to retain source image information while enhancing the extraction of the input image details. We provide an experimental analysis on different types of data acquired from different satellites to demonstrate that our method better maintains the spectral information and obtains richer spatial details than existing fusion methods.
In future work, we will keep using the idea of MCA combined with deep learning to apply this work not only to MS image and PAN image fusion. Our scheme can be improved by continuing to refine the network structure to apply hyperspectral image and MS image fusion or hyperspectral image and PAN image fusion.

Author Contributions

Data curation, B.J. and J.X.; Formal analysis, B.J. and J.X.; Funding acquisition, J.X.; Methodology, B.J. and J.X.; Resources, J.X.; Supervision, J.X., H.X. and P.W.; Writing—original draft, B.J.; Writing—review & editing, J.X., H.X. and P.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of China, grant number 62072391 and 62066013; the Natural Science Foundation of Shandong, grant number ZR2019MF060; the Graduate Science and Technology Innovation Fund project of Yantai University, grant number KGIFYTU2226.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhou, H.Y.; Liu, Q.J.; Wang, Y.H. PGMAN: An unsupervised generative multiadversarial network for pansharpening. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2021, 14, 6316–6327. [Google Scholar]
  2. Choi, J.; Yu, K.; Kim, Y. A new adaptive component-substitution-based satellite image fusion by using partial replacement. IEEE Trans. Geosci. Remote Sens. 2011, 49, 295–309. [Google Scholar]
  3. Shah, V.P.; Younan, N.H.; King, R.L. An efficient pan-sharpening method via a combined adaptive PCA approach and contourlets. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1323–1335. [Google Scholar]
  4. Vivone, G.; Simoes, M.; Dalla Mura, M.; Restaino, R.; Bioucas-Dias, J.M.; Licciardi, G.A.; Chanussot, J. Pansharpening based on semiblind deconvolution. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1997–2010. [Google Scholar]
  5. Li, J.; Sun, W.X.; Jiang, M.H.; Yuan, Q.Q. Self-supervised pansharpening based on a cycle-consistent generative adversarial network. IEEE Trans. Image Process. 2022, 19, 1–5. [Google Scholar]
  6. Ozcelik, F.; Alganci, U.; Sertel, E.; Unal, G. Rethinking CNN-based pansharpening: Guided colorization of panchromatic images via GANs. IEEE Trans. Geosci. Remote Sens. 2021, 59, 3486–3501. [Google Scholar]
  7. Deng, L.J.; Vivone, G.; Jin, C.; Chanussot, J. Detail injection-based deep convolutional neural networks for pansharpening. IEEE Trans. Geosci. Remote Sens. 2021, 59, 6995–7010. [Google Scholar] [CrossRef]
  8. Wang, P.; Yao, H.Y.; Li, C.; Zhang, G.; Leung, H. Multiresolution analysis based on dual-scale regression for pansharpening. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–19. [Google Scholar] [CrossRef]
  9. Lu, H.Y.; Yang, Y.; Huang, S.; Tu, W.; Wan, W.G. A unified pansharpening model based on band-adaptive gradient and detail correction. IEEE Trans. Image Process. 2022, 31, 918–933. [Google Scholar]
  10. Starck, J.L.; Elad, M.; Donoho, D. Redundant multiscale transforms and their application for morphological component separation. Adv. Imageing Elect. Phys. 2004, 132, 287–348. [Google Scholar]
  11. Starck, J.L.; Elad, M.; Donoho, D.L. Image decomposition via the combination of sparse representations and a variational approach. IEEE Trans. Image Process. 2005, 14, 1570–1582. [Google Scholar] [CrossRef] [PubMed]
  12. Zhang, Z.; He, H. A customized low-rank prior model for structured cartoon-texture image decomposition. Signal Process. Image Commun. 2021, 96, 116308. [Google Scholar] [CrossRef]
  13. Deng, X.; Liu, Z. An improved image denoising method applied in resisting mixed noise based on MCA and median filter. In Proceedings of the 2015 11th International Conference on Computational Intelligence and Security (CIS), IEEE, Shenzhen, China, 19–20 December 2015; pp. 162–166. [Google Scholar]
  14. Elad, M.; Starck, J.L.; Querre, P.; Donoho, D.L. Simultaneous cartoon and texture image inpainting using morphological component analysis (MCA). Appl. Comput. Harmon. A 2005, 19, 340–3358. [Google Scholar] [CrossRef]
  15. Zhu, M.; Xu, J.; Liu, Z. Remote sensing image fusion based on multi-morphological convolutional neural network. In Machine Learning for Cyber Security; Springer: Cham, Switzerland, 2020; pp. 485–495. [Google Scholar]
  16. Abdi, A.; Rahmati, M.; Ebadzadeh, M.M. Entropy based dictionary learning for image classification. Pattern Recognit. 2021, 110, 107634. [Google Scholar] [CrossRef]
  17. Jeon, G. Information entropy algorithms for image, video, and signal processing. Entropy 2021, 23, 926. [Google Scholar] [CrossRef]
  18. Xu, J.; Ni, M.; Zhang, Y.; Tong, X.; Zheng, Q.; Liu, J. Remote sensing image fusion method based on multiscale morphological component analysis. J. Appl. Remote Sens. 2016, 10, 025018. [Google Scholar] [CrossRef]
  19. Cho, W.H.; Kim, S.W.; Lee, M.E.; Kim, S.H.; Park, S.Y.; Jeong, C.B. Multimodality image registration using spatial procrustes analysis and modified conditional entropy. J. Signal. Process Syst. 2009, 54, 101–114. [Google Scholar] [CrossRef]
  20. Qu, Y.; Baghbaderani, R.K.; Qi, H.; Kwan, C. Unsupervised pansharpening based on self-attention mechanism. IEEE Trans. Geosci. Remote Sens. 2021, 59, 3192–3208. [Google Scholar] [CrossRef]
  21. Diederik, K.; Jimmy, B. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  22. Yin, J.; Qu, J.; Chen, Q.; Ju, M.; Yu, J. Differential strategy-based multi-level dense network for pansharpening. Remote Sens. 2022, 14, 2347. [Google Scholar] [CrossRef]
  23. Zhang, H.; Xu, H.; Guo, X. SDPNet: A deep network for pan-sharpening with enhanced information representation. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4120–4134. [Google Scholar]
  24. Zhong, X.W.; Qian, Y.R.; Liu, H.; Chen, L. Attention_FPNet: Two-branch remote sensing image pansharpening network based on attention feature fusion. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2021, 14, 11879–11891. [Google Scholar] [CrossRef]
  25. Alparone, L.; Aiazzi, B.; Baronti, S.; Garzelli, A.; Nencini, F.; Selva, M. Multispectral and panchromatic data fusion assessment without reference. Photogramm. Eng. Remote Sens. 2008, 74, 193–200. [Google Scholar] [CrossRef]
  26. Li, C.; Zheng, Y.H.; Jeon, B. Pansharpening via subpixel convolutional residual network. IEEE Trans. Geosci. Remote Sens. 2021, 14, 10303–10313. [Google Scholar] [CrossRef]
  27. Zhang, N.; Wu, Q. Information influence on QuickBird images by brovey fusion and wavelet fusion. Remote Sens. Technol. Appl. 2011, 21, 67–70. [Google Scholar]
  28. Aiazzi, B.; Baronti, S.; Selva, M. Improving component substitution pansharpening through multivariate regression of MS+Pan data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  29. Carper, W.J.; Lillesand, T.M.; Kiefer, P.W. The use of intensity-hue-saturation transformations for merging SPOT panchromatic and multispectral image data. Photogramm. Eng. Remote Sens. 1990, 56, 459–467. [Google Scholar]
  30. Shensa, M.J. The discrete wavelet transform: Wedding the a trous and mallat algorithms. IEEE Trans. Signal Process. 1992, 40, 2464–2482. [Google Scholar] [CrossRef]
  31. Vivone, G.; Alparone, L.; Chanussot, J.; Dalla Mura, M.; Garzelli, A.; Licciardi, G.A.; Restaino, R.; Wald, L. A critical comparison among pansharpening algorithms. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2565–2586. [Google Scholar] [CrossRef]
  32. Mallat, S.G. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern. Anal. 1989, 11, 674–693. [Google Scholar] [CrossRef]
  33. Yang, J.; Fu, X.; Hu, Y.; Huang, Y.; Ding, X.; Paisley, J. PanNet: A deep network architecture for pan-sharpening. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), IEEE, Venice, Italy, 22–29 October 2017; pp. 1753–1761. [Google Scholar]
  34. Ye, F.; Li, X.; Zhang, X. FusionCNN: A remote sensing image fusion algorithm based on deep convolutional neural networks. Multimed. Tools Appl. 2019, 78, 14683–14703. [Google Scholar] [CrossRef]
  35. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. Pansharpening by convolutional neural networks. Remote Sens. 2016, 8, 594. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Cartoon components of the MS image at different scales.
Figure 1. Cartoon components of the MS image at different scales.
Sensors 22 07339 g001
Figure 2. Cartoon components of the PAN image at different scales.
Figure 2. Cartoon components of the PAN image at different scales.
Sensors 22 07339 g002
Figure 3. The relationship of information entropy between images T and C.
Figure 3. The relationship of information entropy between images T and C.
Sensors 22 07339 g003
Figure 4. The relationship between noised image T and noised image C.
Figure 4. The relationship between noised image T and noised image C.
Sensors 22 07339 g004
Figure 5. The overall frame diagram of the fusion network and structure details. (a) The overall frame diagram of the fusion network; (b) The structure details of the network.
Figure 5. The overall frame diagram of the fusion network and structure details. (a) The overall frame diagram of the fusion network; (b) The structure details of the network.
Sensors 22 07339 g005aSensors 22 07339 g005b
Figure 6. The flow chart of the fusion algorithm.
Figure 6. The flow chart of the fusion algorithm.
Sensors 22 07339 g006
Figure 7. Fusion results at different scales.
Figure 7. Fusion results at different scales.
Sensors 22 07339 g007
Figure 8. Information entropy line chart of fused images.
Figure 8. Information entropy line chart of fused images.
Sensors 22 07339 g008
Table 1. Information entropy of fused images at different scales.
Table 1. Information entropy of fused images at different scales.
Different decomposition coefficients (fusion scale)8/51216/51232/51264/512128/512256/512
The entropy of the fused image6.78986.86646.87016.87136.87316.8706
Table 2. The first group evaluation indexes of different fusion results.
Table 2. The first group evaluation indexes of different fusion results.
Fusion MethodPSNRERGASRMSECC
Brovey26.41006.151920.10440.9806
GS28.74205.114320.00540.9811
IHS28.73695.116020.01050.9810
ATWT28.80156.879319.91300.9740
PCA28.56466.095420.17770.9790
DWT27.41186.963619.15080.9726
PanNet28.54255.188018.20840.9908
FusionCNN27.51994.392517.73320.9916
PNN27.91944.415218.84550.9815
Ours28.81363.873416.27110.9934
Table 3. The second group evaluation indices of different fusion results.
Table 3. The second group evaluation indices of different fusion results.
Fusion MethodPSNRERGASRMSECC
Brovey31.99125.442527.38280.8618
GS32.51295.264626.67830.8663
IHS32.48235.299526.71860.8664
ATWT35.82582.556012.90340.9648
PCA31.93795.414527.45690.9657
DWT34.37092.602713.13780.9635
PanNet37.48643.130414.77220.9653
FusionCNN37.67862.554312.83650.9804
PNN37.52683.112114.82920.9671
Ours38.39222.540312.76770.9809
Table 4. The third group evaluation indexes of different fusion results.
Table 4. The third group evaluation indexes of different fusion results.
Fusion MethodQNR D λ D s
Brovey0.79790.33430.1022
GS0.88630.17170.0407
IHS0.73160.32450.0503
ATWT0.73750.08430.1465
PCA0.87670.24640.0562
DWT0.78860.11140.1302
PanNet0.92730.05220.0373
FusionCNN0.90410.06790.0356
PNN0.93680.05300.0396
Ours0.95280.06960.0348
Table 5. The fourth group evaluation indexes of different fusion results.
Table 5. The fourth group evaluation indexes of different fusion results.
Fusion MethodQNR D λ D s
Brovey0.72000.03230.1754
GS0.86730.02780.4198
IHS0.74810.03990.4409
ATWT0.87770.05310.2787
PCA0.78130.03650.3911
DWT0.89170.08960.1870
PanNet0.93880.02530.0424
FusionCNN0.90410.02730.0427
PNN0.93850.03430.0528
Ours0.95150.02620.0418
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jia, B.; Xu, J.; Xing, H.; Wu, P. Remote Sensing Image Fusion Based on Morphological Convolutional Neural Networks with Information Entropy for Optimal Scale. Sensors 2022, 22, 7339. https://doi.org/10.3390/s22197339

AMA Style

Jia B, Xu J, Xing H, Wu P. Remote Sensing Image Fusion Based on Morphological Convolutional Neural Networks with Information Entropy for Optimal Scale. Sensors. 2022; 22(19):7339. https://doi.org/10.3390/s22197339

Chicago/Turabian Style

Jia, Bairu, Jindong Xu, Haihua Xing, and Peng Wu. 2022. "Remote Sensing Image Fusion Based on Morphological Convolutional Neural Networks with Information Entropy for Optimal Scale" Sensors 22, no. 19: 7339. https://doi.org/10.3390/s22197339

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop