Next Article in Journal
Breast Cancer Histopathological Image Recognition Based on Pyramid Gray Level Co-Occurrence Matrix and Incremental Broad Learning
Next Article in Special Issue
MobileUNetV3—A Combined UNet and MobileNetV3 Architecture for Spinal Cord Gray Matter Segmentation
Previous Article in Journal
Ultra-Low-Cost Design of Ripple Carry Adder to Design Nanoelectronics in QCA Nanotechnology
Previous Article in Special Issue
Depth Estimation Method for Monocular Camera Defocus Images in Microscopic Scenes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Image Fusion Method Based on Sparse Decomposition

1
College of Translation Studies, Xi’an Fanyi University, Xi’an 710105, China
2
School of Automation, University of Electronic Science and Technology of China, Chengdu 610054, China
3
Department of Geography and Anthropology, Louisiana State University, Baton Rouge, LA 70803, USA
4
College of Computer Science and Cyber Security, Chengdu University of Technology, Chengdu 610059, China
*
Authors to whom correspondence should be addressed.
Electronics 2022, 11(15), 2321; https://doi.org/10.3390/electronics11152321
Submission received: 12 July 2022 / Revised: 23 July 2022 / Accepted: 23 July 2022 / Published: 26 July 2022
(This article belongs to the Special Issue Medical Image Processing Using AI)

Abstract

:
In the principle of lens imaging, when we project a three-dimensional object onto a photosensitive element through a convex lens, the point intersecting the focal plane can show a clear image of the photosensitive element, and the object point far away from the focal plane presents a fuzzy image point. The imaging position is considered to be clear within the limited size of the front and back of the focal plane. Otherwise, the image is considered to be fuzzy. In microscopic scenes, an electron microscope is usually used as the shooting equipment, which can basically eliminate the factors of defocus between the lens and the object. Most of the blur is caused by the shallow depth of field of the microscope, which makes the image defocused. Based on this, this paper analyzes the causes of defocusing in a video microscope and finds out that the shallow depth of field is the main reason, so we choose the corresponding deblurring method: the multi-focus image fusion method. We proposed a new multi-focus image fusion method based on sparse representation (DWT-SR). The operation burden is reduced by decomposing multiple frequency bands, and multi-channel operation is carried out by GPU parallel operation. The running time of the algorithm is further reduced. The results indicate that the DWT-SR algorithm introduced in this paper is higher in contrast and has much more details. It also solves the problem that dictionary training sparse approximation takes a long time.

1. Introduction

In a two-dimensional image, the amount of information expressed is generally extremely limited [1,2,3,4,5,6]. For example, infrared photography cannot show the real image well. The multi-focus image is vague, but for the area in the focal plane. The visible image details are obvious, while the contrast is weak. Thus, image fusion technology is used to obtain complementary information between images, which can make the image information a lot more aplenty all-around and make its content more accurate.
Image fusion [7] is the process of concentrating the imporTt information of two or more images into one image. The fused image is more complete and accurate than the source image. From a more intuitive point of view, the fused image will have a larger display range or higher definition than the source image. This kind of method makes it possible to preserve different information in multiple images of the same scene, such as hyperspectral images, infrared images, CT images, and ordinary images, which are complementary to each other, and the information combined together can reflect the information that could not be obtained originally [8,9], especially in the fields of medicine, astronomy, and meteorology [8,9,10,11]. In early April 2018, the Google “cell” journal published papers that use multi-frame optical microscope images for machine learning and can predict the characteristic areas that are clearly imaged under the fluorescence microscope. This paper shows that the information fusion theory of images has a high application value.
Image fusion is divided into pixel level, feature level, and decision level [12,13]. Pixel-level image fusion is the basis of the latter two, and its details such as edge, boundary, and texture can be more accurately retained so that the fused image has more information [3,14]. However, its disadvantage is just like its name. It is very expensive to calculate at pixel level 640 × 480. The surface fitting of points will cause the simulation software to die. Pixel level methods are commonly used in image fusion based on region segmentation, image fusion based on pyramid transform, image fusion based on Discrete Wavelet Transformation (DWT) transform, and fusion method based on the non-subsampled wave adaptive image fusion, etc. The theory of wavelet transform is widely used.
There are three categories of the existing pixel-level image fusion methods, namely, methods based on sparse representation, methods based on multiscale decomposition, and methods of directly fusing image pixels or other transform domains.
Multiscale transform is a popular image algorithm that has many applications in image fusion and other image processing [15,16,17]. Firstly, the multiscale representation of the input image is obtained by multiscale transformation, in which the image features are represented in the frequency domain. Then, according to the specific fusion rules, the fused multiscale representation is obtained, which considers the coefficients’ activity level and the correlation between adjacent pixels and different scale coefficients. Finally, by performing an inverse multiscale transformation of the fusion representation, the fusion image is obtained. There are two basic problems involved in the framework, namely, the fusion strategy for multiscale representation fusion and the selection of the multiscale decomposition method.
By simulating the sparse coding mechanism of the human visual system, sparse representation [18] is a novel image representation theory. It has been successfully applied to settle many image processing problems, such as denoising, interpolation, and recognition. Sparse representation can describe an image (or image block) by a atoms’ sparse linear combination, which is selected from a super complete dictionary. The obtained weighting coefficients are sparse, meaning that the significant information of the original image can effectively be represented by only a few non-zero elements in the sparse coefficients. The sparse representation theory can be applied to image fusion by using the characteristics of sparse coefficients. To be specific, for the cause of capturing local salient features and maintaining shift invariance, input images from multiple sources are first segmented into many overlapping blocks. Then, in order to obtain the corresponding sparse coefficients, we decompose these blocks into the same hyper-complete dictionary. Finally, the fusion coefficient and dictionary are used to reconstruct the image. There are two key problems in the model, which are the calculation or learning algorithm of sparse dictionaries and the optimization algorithm of the coefficient fitting.
Among them, the selection of a sparse dictionary is the most basic problem. Generally, it should be selected according to the characteristics of the target signal itself. Traditional signal transformation methods, such as the Conventional Fourier transformation method [19], DWT transformation method [20], Curvelet transformation method [21], and NSCT transformation method [14], belong to the pyramid transform class. This kind of method transforms the target signal through the specified basis function. Because of the orthogonality of the basis function, the construction is simple, and the computational complexity is low. It is a kind of method based on an analytic dictionary. However, in some cases, these pre-specified basis functions cannot fit the structural features of the target signal, such as the texture characteristics of the image and light and dark features. The second method combines the machine learning method, which has been popular in artificial intelligence for more than ten years. The training set is assigned for learning in advance, and then the sparse dictionary [22] suitable for the training set is obtained. Such a sparse dictionary can extract the inherent characteristics of the target signal more accurately, and its common algorithms are K-SVD dictionary learning [23], structure-based dictionary learning [24], and online dictionary learning [25].
In addition to sparse representation or fusion methods based on multiscale decomposition, transformations that are based on dimensionality reduction and color space have also been successfully used in image fusion, such as intensity hue saturation (IHS) transform [26,27]. These methods have been widely applied to the low-resolution multi-spectral and high-resolution panchromatic image fusion.
Although various image fusion methods that are based on color space and dimensionality reduction have been put forward, these methods are usually used for the fusion of color and gray images. Therefore, they are confined to some specific applications, such as the smoothing problem mentioned above. For the sake of combining the advantages of different transformations, a variety of fusion methods of different transformation combi-nations are combined by researchers to obtain a better fusion effect. Liu et al. [28] proposed a general image fusion framework that is based on multiscale transformation as well as sparse representation. Xu J. et al. [29] did not use the traditional multiscale transformation and proposed a fusion method that is based on morphological component analysis as well as sparse representation, in which researchers decompose the source images by using morphological component analysis and then fuse them by using the method based on sparse representation. Wang et al. [30] proposed a fusion method that is based on non-subsampled contourlet transform as well as sparse representation.
However, the existing image fusion technology has some problems. The image fusion method based on multiscale transformation can extract spatial structure information in multiscale space, but it ignores the sparse representation of low-frequency components, and the main energy of the image is concentrated in low-frequency components, so multiscale transformation has limitations on image feature representation, and also puts forward higher requirements for image registration and fusion rules. The fusion method based on sparse representation realizes the matching of sparse matrix and features through dictionary learning and obtains a more meaningful representation of the source image. However, the finite atoms in the dictionary cannot completely represent the small-scale details, and the algorithm has high computational complexity. When the image resolution is high, or the number of images is large, the training time will be greatly increased, so the calculation speed of the algorithm will be greatly affected. If we use the idea of parallel computing and improve the image fusion method based on sparse decomposition, the algorithm speed may be improved. Among them, discrete wavelet transform (DWT) is the most common method.
However, DWT has some fundamental defects, such as shift variance, aliasing, and lack of directivity. DWT is a widely used multi-resolution fusion technology because it can provide better spatial and spectral resolution than other traditional multi-resolution methods. In the DWT-based fusion method, the “maximum absolute” rule is used to select the fusion coefficient from the source image. However, this method is sensitive to noise and artifacts. In the transform-based fusion algorithm, a simple “average rule” is used to fuse the low-frequency coefficients. Because the low-frequency component contains detailed information of the source image, in order to gain better visual clarity, it is necessary to use the sharpness measure of selecting the focus pixel from the source image. Inspired by the above key problems, this paper proposes a new multi-focus image fusion method to fuse DWT coefficients.
In this paper, defocusing methods of defocused images, especially multi-focus image fusion methods, are studied. Firstly, the reason for image blurring in the microscopic scene is analyzed, which mainly lies in the shallow depth of field of the microscope. Aiming at this characteristic, complete and clear information is obtained by collecting a series of defocused images. Among them, transform-based image fusion algorithms, SR image fusion algorithms, and improved fusion algorithms are mainly used. The improved SR image fusion algorithm achieves the effect of parallel operation to a certain extent through the process of sparse representation of image blocks and image fusion and has a good effect on reducing the operation time. Secondly, the improved algorithm only uses the SR algorithm to fuse high-frequency coefficients, which can retain more focus areas and reduce artifacts. In general, by comparing several fusion algorithms, the effectiveness of the improved algorithm in this paper is verified.

2. Materials and Methods

In the past five years, in the field of signal and image processing, more and more researchers have been interested in heap sparse representation theory. This kind of method obtains, represents, and compresses high-dimensional signals so that the signal can be represented as complete as possible in the form of minimum data.
Generally, we express the sparse model as follows:
min x x 0 s . t y Dx
For signal y, we look forward to finding a dictionary D R m × p , so that x has the least representation of non-zero points.
For the sparse representation of the image, it needs to be partitioned first. Suppose the image Y is divided into blocks, and the blocks are arranged in dictionary order, where Y R m . The dictionary D R m × p also contains several atoms corresponding to the source image as its column vector. When p < m, it is called an over complete dictionary. After the dictionary is determined, an over-complete dictionary D can be used for any image, and the linear combination of atoms in D can be used for sparse approximation:
min x x 0 s . t y Dx 2 2 < ε
where ε is the allowable approximation error.
The dictionary learning process in sparse presentation also has problems in computational complexity, the same as other machine learning algorithms. With a high image resolution or a great number of images, the training time may be much greater than the sum of the other steps of image fusion, which can even reach dozens of times. This affects the algorithm’s calculation speed. We propose an improved image fusion method based on sparse decomposition by thinking of parallel computing to improve this problem. Discrete wavelet transform (DWT) is the most common multiscale fusion algorithm based on the transform in all of these methods.
The selection of fusion rules is the crux of this method. For different decomposition levels, we should select different fusion rules. For the reason of ensuring high visual clarity, the simplest average rule is often applied to low-frequency signals [3,11]. Inspired by this, the source image is decomposed into high-frequency and low-frequency signals by using the DWT algorithm. The high-frequency component indicates the images’ structural information. Here, the sparse representation (SR) fusion rule is applied to obtain the high-frequency fusion coefficient. At the same time, the low-frequency component is divided into fixed-size blocks, which are used to calculate the clarity index. Then, we select and fuse the block that has the largest sharpness index into the low-frequency coefficient. In the end, by using the method of inverse discrete wavelet transform (IDWT), we obtain the fused image. By using the same multithreading mechanism computer, the high-frequency signal is decomposed with a large amount of information at multiscale, and the sparse representation is fused with less information according to each scale by this method. As a result, the training dictionary time and image fusion time are reduced.
The fusion process of the improved method is shown in Figure 1.
In the process of image block processing, if the sliding window method is used, the sub-block itself has less information, but the adjacent sub-blocks have more overlapping information, which can cause overfitting in the training process, which is needed to avoid by machine learning algorithms. The method of the block is different from the common sparse representation area. Based on that, we used the non-overlapping block method. The image is divided into blocks according to the size of 4 × 4. Similarly, the K-SVD dictionary combined with DWT is used to train the corresponding sub dictionary for each scale sub-block. The purpose of decomposing the signal for parallel operation is achieved by segmenting the learning dictionary.
The schematic diagram of the improved sparse representation image fusion algorithm described in this section is shown in Figure 2. DWT transform is used to decompose two-dimensional images into high and low-frequency domain signals. The number of decomposition layers is set to three, and ten correlation components b A = b a ,   b 2 , ,   b 10 , b B = b b ,   b 2 , ,   b 10 are generated, respectively.
The fusion rules of low-frequency coefficients are as follows:
For LA and LB, the low-frequency bands are divided into 8 × 8 sub-blocks in order. Suppose the T block is divided, p LA i i = 1 T is the i-th sub-block in figure A, p LB i i = 1 T is the i-th sub-block in figure B.
Calculate the variance of LA and LB, var LA i and var LB i .
According to Equation (2), the focus block in the subband is selected based on the image variance as the definition evaluation index, and then the fusion coefficient is determined.
The low-frequency coefficients are calculated, and an inverse DWT transform is performed.
The fusion rules of high-frequency coefficients are as follows:
The high-frequency bands HA and HB are divided into 8 × 8 image blocks from the top left to the bottom right by the sliding window method, and the step size is one pixel. Assuming that it is divided into T subblock, HA and HB in the i-th subblock are denoted as p HA i i = 1 T and p HB i i = 1 T , respectively;
Rearrange image subblocks p HA i , p HB i into column vectors v HA i , v HB i ;
The sparse coefficient vector α HA i , α HB i corresponding to v HA i , v HB i is calculated by the OMP algorithm, and the calculation method is as follows:
α HA i = argmin α 0 s . t . v HA i D α 2 ε
α HB i = argmin α 0 s . t . v HB i D α 2 ε
where D is the corresponding dictionary.
According to Equation (3), sparse vectors are selected by using the maximum absolute value criterion:
α HF i = α HA i α HA i 1 > α HB i 1 α HB i α HA i 1 α HB i 1
v HF i = D α HF i was used to calculate the fusion coefficient.
The high-frequency coefficients are transformed by inverse DWT [31].
Finally, the column vectors obtained by the inverse DWT transform of high and low-frequency coefficients are combined according to the corresponding positions, and the matrix is the image fused by this method.
The experimental environment involved in this paper is summarized as follows:
  • One PC with I5-6500CPU/GTX1060ti/16G/256GSSD;
  • A ByslorPylon industrial camera, model: acA640-120uc;
  • A monocular microscope with its magnification of 0.5 × (0.7~4.5), lens radius of 35 mm as well an F-number of 4;
  • One USB 3.0 data bus;
  • The software involved in the experimental environment: all the experiments in this paper are completed under the windows 10 operating system. The PylonViewer of Bysler company is used to collect the image under the microscope, MATLAB 2016a is used for simulation, and a few of the preprocessing steps are completed using Microsoft Visual Studio 2013 and OpenCV;
  • The appearance of the experimental environment is shown in Figure 3.

3. Results

We first took two sets of multi-focus images using a microscope and Basler industrial camera acA640, as shown in Figure 4.
In the improved algorithm, the sparse representation process and sparse dictionary training are not only carried out in the frequency domain but also decomposed into nine frequency bands, and the parallel operation of computer GTX1060Ti is used to speed up the operation. In Figure 4, the paper sequence image is taken with the grid paper tilted at a 30° angle, and its ladder characteristic is conducive to the depth estimation affect detection. The coins’ sequence is an image of four coins with the original size of 640 × 480 through image registration and downsampling for image compression. After preprocessing, the DWT-SR depth estimation time evaluation with parallel operation and the DWT-SR time evaluation without parallel operation were carried out for the two groups of images in MATLAB 2016, respectively, and Table 1 was obtained.
At the same time, the DWT-based image fusion, sparse representation image fusion, and improved sparse representation fusion algorithm are compared. Among them, the wavelet transform method uses the rule that the low-frequency part takes the average, and the high-frequency part takes the large absolute value for fusion. The algorithm is based on sparse representation, and its improved algorithm uses the K-SVD training dictionary, sparse approximation completes the coefficient fitting process using OMP algorithm. The fusion results are shown in Figure 5.
Fusion quality assessment is very important for analyzing fusion results. The subjective evaluation method, which relies on the judgment of the human eyes and human brain to make a conclusion, is one of its evaluation methods. The subjective evaluation method is to rely on the human eye to make a subjective judgment on the effect of the fused image. For example, find some testers, let them recognize the specific target in the fused image obtained by different fusion algorithms, measure the recognition time, and count the recognition accuracy so as to judge the performance of the image fusion algorithm. The other is the objective evaluation method, and its fused image evaluation relies on some objective parameters from the fused image. We usually use standard deviation, spatial frequency, and entropy to evaluate the quality of the fused image.
Entropy (H) is regarded as one of the significant image quality indexes for evaluating the fused images’ information. It is defined as:
H = i = 0 255 p F i 1 log 2 p F i 1
where p F i 1 represents the probability of the intensity value i1 in image F. The higher the entropy value, the better the information content of the fused image is.
Standard deviation δ is regarded as one of the best indicators for the contrast value measurement of fused images. If the standard deviation is high, the image contrast will be better. It is defined as:
δ = 1 MN i = 1 M j = 1 N F i , j μ 2
where the mean µ is gray level F.
The spatial frequency (SF) reflects the edge information which is stored in the fused image. It is defined as:
SF = RF 2 + CF 2
RF = 1 MN i = 1 M j = 2 N F i , j F i , j 1 2
CF = 1 MN i = 1 M j = 2 N F i , j F i 1 , j 2
where RF and CF represent the line frequency and column frequency of the fused image F; F(i,j) is the gray value at the pixel (i,j); M and N are the width and height of the image, respectively.
Q AB / F is gradient-based fusion performance, and the higher this value is, the higher the image quality is, and the more information about the original image will be retained.
Q AB / F = m = 1 M n = 1 N Q AF n , m w A n , m + Q BF n , m w B n , m i = 1 M j = 1 N w A i , j + w B i , j
QAF (n,m), QBF (n,m) is the edge strength of A,B, respectively; wA (n,m), wB (n,m) is the corresponding weight.
Q_w is a weighted quality evaluation index. Because the importance of each image block is different, the proportion of image block quality in quality evaluation is set according to the importance:
Q AF Q w = C w λ w Q 0 A , F / w + 1 + λ w Q 0 B , F / w
w is an area in the image; C w is the standardization coefficient; λ w is the importance of the image block in the whole image.
In the experiment of the image fusion algorithm in this paper, researchers respectively tested and verified two groups of defocused image sequences coins and paper. H, SF, and QAB/F performance indexes were used to evaluate the algorithm’s fusion effect, and DWT, SR, and DW-SR algorithms were compared. The objective evaluation index results in Table 2 were obtained.

4. Discussion

Through the comparison of the operation time, it can be seen that the multiscale decomposition method of wavelet transform can effectively reduce the operation time. The fusion process of the spatial domain is put into operation after frequency domain decomposition. The decomposition operation process is divided into several threads, which reduces the operating pressure of each thread. The data show that the improvement effect is remarkable.
According to the results, the DWT-SR algorithm provides better contrast. That is, the image is clearer. To sum up, DWT-SR is more suitable for multi-focus image fusion [29,30,31] in the microscopic scene than others. With regard to CPU response time, DWT-SR’s operation speed is faster than the SR algorithm, which is increased by one order of magnitude and improves the existing problems. The DWT algorithm does not require sparse approximation for each image block, so the calculation speed of the DWT method is higher than that of the DWT-SR method. However, the effect of the DWT algorithm is worse than the DWT-SR method for the same reason.
This paper focuses on the image fusion algorithm based on DWT and SR image fusion algorithm optimized by DWT. The DWT-SR algorithm obtains the fusion image, which processes higher contrast and more details. It also improves the problem that dictionary training and sparse approximation take a long training time. In the specific practical application, we can have better access to precise texture information.

Author Contributions

Conceptualization, W.Z. and S.L.; methodology, B.Y. and L.Y.; software, P.W. and Y.B.; validation, Y.B. and S.L.; formal analysis, P.W. and L.Y.; investigation, B.Y.; resources, X.Q. and S.L.; data curation, X.Q. and P.W.; writing—original draft preparation, X.Q., Y.B. and L.Y.; writing—review and editing, X.Q., M.L., S.L. and W.Z.; visualization, P.W. and L.Y.; supervision, B.Y.; project administration, W.Z.; funding acquisition, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

Support by the Sichuan Science and Technology Program, 2021YFQ0003.

Data Availability Statement

The data used are not publicly available. Please contact the corresponding author for access to the dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, B.; Liu, C.; Huang, K.; Zheng, W. A triangular radial cubic spline deformation model for efficient 3D beating heart tracking. Signal Image Video Process. 2017, 11, 1329–1336. [Google Scholar] [CrossRef] [Green Version]
  2. Zhou, Y.; Zheng, W.; Shen, Z. A new algorithm for distributed control problem with shortest-distance constraints. Math. Probl. Eng. 2016, 2016, 1604824. [Google Scholar] [CrossRef]
  3. Tang, Y.; Liu, S.; Deng, Y.; Zhang, Y.; Yin, L.; Zheng, W. Construction of force haptic reappearance system based on Geomagic Touch haptic device. Comput. Methods Programs Biomed. 2020, 190, 105344. [Google Scholar] [CrossRef] [PubMed]
  4. Tan, L.; Chen, Y.; Zhang, W. Multi-focus Image Fusion Method based on Wavelet Transform. J. Phys. Conf. Ser. 2019, 1284, 012068. [Google Scholar] [CrossRef] [Green Version]
  5. Li, S.; Yang, B. Multifocus image fusion using region segmentation and spatial frequency. Image Vis. Comput. 2008, 26, 971–979. [Google Scholar] [CrossRef]
  6. Helck, A.; D’Anastasi, M.; Notohamiprodjo, M.; Thieme, S.; Sommer, W.; Reiser, M.; Clevert, D. Multimodality imaging using ultrasound image fusion in renal lesions. Clin. Hemorheol. Microcirc. 2012, 50, 79–89. [Google Scholar] [CrossRef]
  7. Li, H.; Manjunath, B.; Mitra, S.K. Multisensor image fusion using the wavelet transform. Graph. Models Image Process. 1995, 57, 235–245. [Google Scholar] [CrossRef]
  8. Wang, Y.; Tian, J.; Liu, Y.; Yang, B.; Liu, S.; Yin, L.; Zheng, W. Adaptive neural network control of time delay teleoperation system based on model approximation. Sensors 2021, 21, 7443. [Google Scholar] [CrossRef]
  9. Zhang, Z.; Liu, Y.; Tian, J.; Liu, S.; Yang, B.; Xiang, L.; Yin, L.; Zheng, W. Study on reconstruction and feature tracking of silicone heart 3D surface. Sensors 2021, 21, 7570. [Google Scholar] [CrossRef]
  10. Guo, F.; Yang, B.; Zheng, W.; Liu, S. Power frequency estimation using sine filtering of optimal initial phase. Measurement 2021, 186, 110165. [Google Scholar] [CrossRef]
  11. Li, Y.; Zheng, W.; Liu, X.; Mou, Y.; Yin, L.; Yang, B. Research and improvement of feature detection algorithm based on FAST. Rend. Lincei Sci. Fis. E Nat. 2021, 32, 775–789. [Google Scholar] [CrossRef]
  12. Yang, B.; Liu, C.; Zheng, W.; Liu, S. Motion prediction via online instantaneous frequency estimation for vision-based beating heart tracking. Inf. Fusion 2017, 35, 58–67. [Google Scholar] [CrossRef]
  13. Ni, X.; Yin, L.; Chen, X.; Liu, S.; Yang, B.; Zheng, W. Semantic representation for visual reasoning. MATEC Web Conf. 2019, 277, 02006. [Google Scholar] [CrossRef]
  14. Li, Y.; Sun, Y.; Huang, X.; Qi, G.; Zheng, M.; Zhu, Z. An image fusion method based on sparse representation and sum modified-Laplacian in NSCT domain. Entropy 2018, 20, 522. [Google Scholar] [CrossRef] [Green Version]
  15. Xu, C.; Yang, B.; Guo, F.; Zheng, W.; Poignet, P. Sparse-view CBCT reconstruction via weighted Schatten p-norm minimization. Opt. Express 2020, 28, 35469–35482. [Google Scholar] [CrossRef]
  16. Chen, X.; Yin, L.; Fan, Y.; Song, L.; Ji, T.; Liu, Y.; Tian, J.; Zheng, W. Temporal evolution characteristics of PM2. 5 concentration based on continuous wavelet transform. Sci. Total Environ. 2020, 699, 134244. [Google Scholar] [CrossRef]
  17. Ding, Y.; Tian, X.; Yin, L.; Chen, X.; Liu, S.; Yang, B.; Zheng, W. Multi-scale relation network for few-shot learning based on meta-learning. In Proceedings of the International Conference on Computer Vision Systems, Thessaloniki, Greece, 23–25 September 2019; pp. 343–352. [Google Scholar]
  18. Zhang, L.; Yang, M.; Feng, X. Sparse representation or collaborative representation: Which helps face recognition? In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 471–478. [Google Scholar]
  19. Goda, K.; Jalali, B. Dispersive Fourier transformation for fast continuous single-shot measurements. Nat. Photonics 2013, 7, 102–112. [Google Scholar] [CrossRef]
  20. Kashyap, N.; Sinha, G. Image watermarking using 3-level discrete wavelet transform (DWT). Int. J. Mod. Educ. Comput. Sci. 2012, 4, 50. [Google Scholar] [CrossRef]
  21. Starck, J.-L.; Candès, E.J.; Donoho, D.L. The curvelet transform for image denoising. IEEE Trans. Image Process. 2002, 11, 670–684. [Google Scholar] [CrossRef] [Green Version]
  22. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G. Online dictionary learning for sparse coding. In Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–18 June 2009; pp. 689–696. [Google Scholar]
  23. Rubinstein, R.; Faktor, T.; Elad, M. K-SVD dictionary-learning for the analysis sparse model. In Proceedings of the 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Kyoto, Japan, 25–30 March 2012; pp. 5405–5408. [Google Scholar]
  24. Shen, L.; Wang, S.; Sun, G.; Jiang, S.; Huang, Q. Multi-level discriminative dictionary learning towards hierarchical visual categorization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 383–390. [Google Scholar]
  25. Wang, H.; Wang, P.; Song, L.; Ren, B.; Cui, L. A novel feature enhancement method based on improved constraint model of online dictionary learning. IEEE Access 2019, 7, 17599–17607. [Google Scholar] [CrossRef]
  26. Xydeas, C.S.; Petrovic, V. Objective image fusion performance measure. Electron. Lett. 2000, 36, 308–309. [Google Scholar] [CrossRef] [Green Version]
  27. Huang, W.; Zheng, W.; Mo, L. Distributed robust H∞ composite-rotating consensus of second-order multi-agent systems. Int. J. Distrib. Sens. Netw. 2017, 13, 1550147717722513. [Google Scholar] [CrossRef]
  28. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  29. Xu, J.; Ni, M.; Zhang, Y.; Tong, X.; Zheng, Q.; Liu, J. Remote sensing image fusion method based on multiscale morphological component analysis. J. Appl. Remote Sens. 2016, 10, 025018. [Google Scholar] [CrossRef]
  30. Wang, J.; Peng, J.; Feng, X.; He, G.; Wu, J.; Yan, K. Image fusion with nonsubsampled contourlet transform and sparse representation. J. Electron. Imaging 2013, 22, 043019. [Google Scholar] [CrossRef] [Green Version]
  31. Liu, S.; Wang, L.; Liu, H.; Su, H.; Li, X.; Zheng, W. Deriving bathymetry from optical images with a localized neural network algorithm. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5334–5342. [Google Scholar] [CrossRef]
Figure 1. K-SVD Schematic diagram of multiscale dictionary training.
Figure 1. K-SVD Schematic diagram of multiscale dictionary training.
Electronics 11 02321 g001
Figure 2. Flow chart of image fusion based on multiscale sparse representation.
Figure 2. Flow chart of image fusion based on multiscale sparse representation.
Electronics 11 02321 g002
Figure 3. The experimental environment of this paper.
Figure 3. The experimental environment of this paper.
Electronics 11 02321 g003
Figure 4. Registration of two groups of partial defocusing.
Figure 4. Registration of two groups of partial defocusing.
Electronics 11 02321 g004
Figure 5. Fusion results of each algorithm.
Figure 5. Fusion results of each algorithm.
Electronics 11 02321 g005
Table 1. Comparison of computing speed of improved algorithm with or without GPU parallel operation.
Table 1. Comparison of computing speed of improved algorithm with or without GPU parallel operation.
ImageEvaluating IndicatorFusion Method
DWT-SR Method with
Parallel Operation
DWT-SR Method without
Parallel Operation
coinsCalculation time/S7.314950.2311
paperCalculation time/S8.937248.3188
Table 2. Objective evaluation index of different fusion methods.
Table 2. Objective evaluation index of different fusion methods.
ImageEvaluating IndicatorFusion Method
DWTSRDWT-SR
coinentropy7.31497.32197.3512
spatial frequency25.031425.192225.9344
QAB/F7.66398.27868.4933
paperentropy7.36887.39197.3188
spatial frequency25.038125.313725.4399
QAB/F7.63998.17298.2729
CPU response time1.15932s313.49317s8.93533s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Qin, X.; Ban, Y.; Wu, P.; Yang, B.; Liu, S.; Yin, L.; Liu, M.; Zheng, W. Improved Image Fusion Method Based on Sparse Decomposition. Electronics 2022, 11, 2321. https://doi.org/10.3390/electronics11152321

AMA Style

Qin X, Ban Y, Wu P, Yang B, Liu S, Yin L, Liu M, Zheng W. Improved Image Fusion Method Based on Sparse Decomposition. Electronics. 2022; 11(15):2321. https://doi.org/10.3390/electronics11152321

Chicago/Turabian Style

Qin, Xiaomei, Yuxi Ban, Peng Wu, Bo Yang, Shan Liu, Lirong Yin, Mingzhe Liu, and Wenfeng Zheng. 2022. "Improved Image Fusion Method Based on Sparse Decomposition" Electronics 11, no. 15: 2321. https://doi.org/10.3390/electronics11152321

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop