Next Article in Journal
Comparative Performance Studies of Up- and Down-Milling on AA7075 Plate
Previous Article in Journal
Studies on In Situ Alloy Formation Using Mild Steel–Inconel 625 Twin Filler Wire Gas Tungsten Arc Weld Deposition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Image Fusion Techniques Based on Optimization Algorithms: A Review †

1
Faculty of Engineering, Department of Computer Science & Engineering-AIML, Academy of Business and Engineering Science, ABESEC, AKTU, Ghaziabad 201001, U.P., India
2
Department of Computer Engineering & Application, FET, Mangalayatan University, Beswan 201016, U.P., India
3
Department of Computer Science & Engineering, IMS Engineering College, AKTU, Ghaziabad 201001, U.P., India
*
Author to whom correspondence should be addressed.
Presented at the International Conference on Recent Advances in Science and Engineering, Dubai, United Arab Emirates, 4–5 October 2023.
Eng. Proc. 2023, 59(1), 225; https://doi.org/10.3390/engproc2023059225
Published: 26 January 2024
(This article belongs to the Proceedings of Eng. Proc., 2023, RAiSE-2023)

Abstract

:
In image processing applications, image fusion techniques gain popularity because they combine the most appropriate features of different source images in order to generate a single image that contains more information and is more beneficial. In this paper, initially, we have analysed the conventional spatial and transform domain image fusion techniques. These techniques face numerous challenges, such as low contrast, noise, and redundancy. To overcome these challenges, adaptive image fusion methods using nature-inspired optimization algorithms (YSGA) are deployed. These algorithms search for the optimal solution for the image fusion technique based on the objective function. Therefore, the main focus of this paper is to study and analyse the optimization algorithms based on various factors.

1. Introduction

Image fusion is a technique that amalgamates multiple data sets from related observations to produce a composite output featuring key aspects of each component. This process involves combining various sensory, temporal, and view images into a single, high-quality image, resulting in enhanced spatial, temporal, and spectral information compared to individual images [1]. The core objective of image fusion is to retain significant aspects of the input images while eliminating unwanted data. This method accommodates images with diverse spectral and spatial properties obtained at different times and using various sensors. Two domains, spatial and frequency, are utilized for image fusion. The spatial domain processes image pixels directly, leading to spatial distortion in the fused images [2]. Hence, frequency domain fusion is preferred, even though it presents shift invariance challenges [3]. Various optimization strategies, including genetic algorithms (GA) [4,5], grasshopper optimization (GO) [6], grey wolf optimization, particle swarm optimization [7], and hybrid combinations [8] are explored to address these issues. The effectiveness of these methods depends on their exploration and exploitation rates, with hybridization increasing computational complexity. Alternative optimization techniques with improved exploration and exploitation rates are sought. One possible solution among these is the Yellow Saddle Goatfish (YSGA) algorithm, which draws inspiration from the hunting behaviour of Yellow Saddle Goatfish.

2. Review of the Literature

In this section, we have reviewed and analysed existing image fusion methods from the literature.
In order to improve fusion efficacy, Shaik Shehanaz, Ebenezer Daniel, and Sivaji Satrasupalli [7] describe an optimal weighted average fusion (OWAF) technique that makes use of particle swarm optimization (PSO). They use the DWT method to split the input from different modalities into subgroups, then PSO is used to reweight the energy bands. MRI-PET, MRI-SPECT, and MRI-CT images are used to assess the approach, which shows robustness against noise and a reduction in calculation time. Using an optimal homomorphic wavelet fusion (OHWF) is recommended by Ebenezer Daniel [8] for the integration of multimodal medical pictures. With the use of the MR-PET, MR-SPECT, MR T1-T2, and MR-CT modalities, this method improves fusion quality by combining homomorphic filtering and wavelet transform. In comparison to other fusion approaches, grey wolf optimization (HG-GWO) produces better outcomes by choosing the optimal scale values based on metrics such as MI, QAB/F STD, and entropy. Tsagaris and Anastassopoulos [9] propose an information-based global measure (IFPM) using mutual information for quantitative evaluation of image fusion methods. This metric ensures that shared information in input photos is considered only once in producing the fused image. Huang and Yang [10] investigate a number of techniques, such as deep learning, transform domain, and spatial domain methods, for merging multimodal medical pictures. Every approach has benefits and drawbacks, with deep learning exhibiting potential despite its limits in terms of data and processing power.
Huang and Yang [10] explore various approaches to fusing multimodal medical images, including deep learning, transform domain, and spatial domain methods. Each method has its advantages and disadvantages, with deep learning showing promise but having limitations in terms of data and computational resources. With an emphasis on high- and low-frequency signal analysis, Kairui Cheng et al. [11] compare wavelet transform-based fusion with Laplace pyramid image fusion. For these signals, they employ various fusion criteria, and they assess performance with respect to quality metrics such as spatial frequency, mean gradient, information entropy, and standard deviation. In this work, wavelet-based fusion seems to have more advantages.
An optimal search approach can be used to describe the hunting strategy for Yellow Saddle Goatfish that Zaldívar, Morales, Rodríguez, Valdivia-G, Cuevas, and Pérez-Cisneros [12] introduce. The fitness value of the two search agent groups—chasers and blockers—used in the method is used to determine how effective the hunting is. This technique has been tried in numerous engineering optimization jobs, demonstrating its effectiveness, precision, and durability. It also demonstrates enhanced optimization results in precision and convergence.

3. Methodology

The medical industry has advanced to the point that a variety of devices that create digital images of the human body are available for purchase, including CT scan machines, MRI machines, and ultrasound equipment. Doctors analyse these pictures to determine different illnesses [13]. Consequently, crisp, high-quality images are needed. In the medical field, image fusion techniques are needed to accomplish this goal. Several fusion techniques are examined and analysed, and difficulties are discovered in this study. Adaptive image fusion is developed utilizing nature-inspired algorithms in response to these issues. Both colour and grayscale image can be processed with the suggested technique. The suggested solution uses a nature-inspired algorithm to tune parameters based on the objective function. The program, which draws inspiration from nature, searches the solution space to find the best solution. As a result, the suggested method varies nature-inspired algorithm parameters like population and iterations and analyses the effect on the final images. The flowchart of the suggested method is shown in Figure 1. input image1 and input image2 are read at the start. The coefficients from both images are then extracted using the first processing model. The optimal weights for the coefficients are then determined by running them through the Nature Inspired (NIA) algorithm. Once the optimal weight values are determined, the coefficients are updated and combined. The coefficients of the two images are fused using the average fusion method. An inverse transform is employed in the last stage to produce an output that is a fused image. Once image fusion is complete, assess the combined images’ performance using several metrics, such as mutual information (MI), peak signal to noise ratio (PSNR), and root mean square error (RMSE).

Yellow Saddle Goatfish Algorithm

The YSG algorithm, a meta-heuristic software that imitates the Yellow Saddle Goatfish’s pack hunting approach, was proposed by Zaldvar et al. in 2018. As shown in Figure 2. There are five steps in the YSG algorithm, which are covered below:
Initial stage: As P ( ), a population of n goatfish is formed at random and uniformly distributed inside the dimensional search space’s perimeter. This is how the initialization is written:
p j i = r a n d · b i h i g h b i l o w + b i l o w
where j = 1 ,   2 ,   3 . n (population); i = 1 ,   2 ,   3 . d (dimension), when the random distribution between is (population); (dimension). The element wise multiplication is represented by a dot. The elementwise multiplication is represented by a dot.
Chaser fish: The YSGA are led in their hunts by a single chaser fish. Which fish will act as the chaser is determined by fitness levels. A fish that is closer to the answer is chosen to be the chaser fish in each group. This chaser fish uses a random stroll to try to locate its target.
Blocker fish: After selecting the chaser fish, every other fish turns into the blocker fish. Blocker fish movement is modelled as an algorithmic spiral. One fish that is closest to its victim is a chaser fish. Blocker fish follow the chaser fish in a spiral pattern, and each blocker fish takes a different path after each iteration.
Exchange of role: Once the blocker and chaser fish are decided, the fish that pursues its prey is situated closer to it. Blocker fish try to stop their victim from escaping. As the prey goes through the hunting area, roles are switched. The blocker fish transforms into the chaser fish if the prey is nearby, and vice versa. An exchange of roles is what is meant by this phenomenon. The element in the algorithm that has the highest fitness is selected as the chaser fish.
Change of zone: The fish group moves to the adjacent sector to look for new prey after fully utilizing the area by hunting every prey. It is shown as:
p g t + 1 = b e s t + p g t 2
where p g t + 1 is the fish’s new location. b e s t is the cluster with the most effective response, p g t is the present state of the fish (chaser or blocker). If the value of a chaser fish has increased within a cluster, it gets swapped out for the best worldwide. The local optima are avoided by this method.

4. Results and Discussion

For the diagnosis and evaluation of medical issues, digital image processing is becoming more and more prevalent. A high-quality digital image is needed for the diagnosis. The qualities of the images vary from image to image in reality, though. Consequently, an adaptive-based image fusion model is suggested in this study. Adaptive pre-processing techniques were needed for the image based on its characteristics since they begin to process the image depending on how much noise, enhancement, or fusion is needed. The effectiveness of an image fusion can be assessed using essentially two different sorts of techniques [14]:
1. Qualitative approach: This method compares the original and fused images visually to determine the spectral content. But, since these measurements are subjective, they are typically not preferred.
2. Quantitative approach: YSGO algorithm. A collection of defined image metrics will be used to measure the spectral and spatial similarity between the fused image and the input images. For assessing the visual information, these criteria are typically preferred. The metrics used in this method include the standard deviation, peak signal-to-noise ratio (PSNR), mean squared error (MSE), root mean square error (RMSE), cross correlation (CC), mutual information (MI), and spectral distortion index (SDI). The various performance metrics parameters used to evaluate the efficiency of the Image Fusion technique are displayed in Table 1.

5. Conclusions

An examination of the literature revealed that the high complexity and inconsistent nature of all image fusion techniques are problems. Each image has a different set of qualities [19,20]. As a result, the typical fusion process has drawbacks with fused images such as unclear images, blurry effects, and level of fusion. The results of multimodal medical image fusion research are becoming more and more important, yet issues with colour distortion and feature extraction in the fusion effect are still not entirely resolved. Weights were selected as the coefficients in order to overcome the issues that were previously highlighted. Primary weighted average fusion may not produce a meaningful fused image because of the weight value that was assigned to the fusion process. Therefore, choosing the right weights for multi-level decomposition components through optimal weight selection may improve performance.

Author Contributions

Conceptualization, A.G. and M.S., Methodology, K.M. and A.G., Testing P.K.S., Validation A.G. and J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article.

Acknowledgments

We sincerely thank everyone who assisted in making this study endeavour successful. We want to start by expressing our sincere thanks to our supervisor, Javed Wasim, and Prabhat Kumar Srivastava for their direction, and assistance, and encouragement during the entire process. Their guidance and knowledge were extremely beneficial to us in guiding the course of our research and realizing our goals. We also like to thank Kanika Malik, Monica Singh, and members of our research team for their contributions, advice, and help during the entire process. We are incredibly appreciative of their efforts and dedication, as they were crucial to the successful completion of our research.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ardeshir Goshtasby, A.; Nikolov, S. Guest editorial: Image fusion: Advances in the state of the art. Inf. Fusion 2007, 8, 114–118. [Google Scholar] [CrossRef]
  2. Tirupal, T.; Mohan, B.C.; Kumar, S.S. Multimodal medical image fusion techniques—A review. Curr. Signal Transduct. Ther. 2021, 16, 142–163. [Google Scholar] [CrossRef]
  3. Qayyum, H.; Majid, M.; Anwar, S.M.; Khan, B. Facial expression recognition using stationary wavelet transform features. Math. Probl. Eng. 2017, 2017, 9854050. [Google Scholar] [CrossRef]
  4. Kavitha, S.; Thyagarajan, K.K. Efficient DWT-based fusion techniques using genetic algorithm for optimal parameter estimation. Soft Comput. 2017, 21, 3307–3316. [Google Scholar] [CrossRef]
  5. Bhardwaj, J.; Nayak, A.; Gambhir, D. Multimodal medical image fusion based on discrete wavelet transform and genetic algorithm. In Proceedings of the International Conference on Innovative Computing and Communications, Delhi, India, 21–23 February 2020; Springer: Singapore, 2021; pp. 1047–1057. [Google Scholar]
  6. Dinh, P.H. A novel approach based on grasshopper optimization algorithm for medical image fusion. Expert Syst. Appl. 2012, 171, 114576. [Google Scholar] [CrossRef]
  7. Shehanaz, S.; Daniel, E.; Guntur, S.R.; Satrasupalli, S. Optimum weighted multimodal medical image fusion using particle swarm optimization. Optik 2021, 231, 166413. [Google Scholar] [CrossRef]
  8. Daniel, E. Optimum wavelet-based homomorphic medical image fusion using hybrid genetic–grey wolf optimization algorithm. IEEE Sens. J. 2018, 18, 6804–6811. [Google Scholar] [CrossRef]
  9. Tsagaris, V.; Anastassopoulos, V. Global measure for assessing image fusion methods. Opt. Eng. 2006, 45, 026201. [Google Scholar] [CrossRef]
  10. Huang, B.; Yang, F.; Yin, M.; Mo, X.; Zhong, C. A review of multimodal medical image fusion techniques. Comput. Math. Methods Med. 2020, 2020, 8279342. [Google Scholar] [CrossRef] [PubMed]
  11. Cheng, K.; Yi, L.; Li, L.; Zhutian, M.; Sun, G. Study of Wavelet Transformbased Image Fusion Methods. J. Phys. Conf. Ser. 2021, 1757, 012049. [Google Scholar] [CrossRef]
  12. Zaldívar, D.; Morales, B.; Rodríguez, A.; Valdivia-G, A.; Cuevas, E.; Pérez-Cisneros, M. A novel bio-inspired optimization model based on Yellow Saddle Goatfish behaviour. Biosystems 2018, 174, 1–21. [Google Scholar] [CrossRef] [PubMed]
  13. Rajagopal, M.; Venkatesan, A.M. Image fusion and navigation platforms for percutaneous image-guided interventions. Abdom. Radiol. 2016, 41, 620–628. [Google Scholar] [CrossRef] [PubMed]
  14. Zhang, Y. Methods for image fusion quality assessment—A review, comparison and analysis. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2008, 37, 1101–1109. [Google Scholar]
  15. Bharati, S.; Khan, T.Z.; Podder, P.; Hung, N.Q. A Comparative Analysis of Image Denoising Problem: Noise Models, Denoising Filters and Applications. In Cognitive Internet of Medical Things for Smart Healthcare; Springer: Cham, Switzerland, 2020; pp. 49–66. [Google Scholar]
  16. Choi, H.-H.; Jeong, J. Speckle noise reduction in ultrasound images using SRAD and guided filter. In Proceedings of the International Workshop on Advanced Image Technology (IWAIT), Chiang Mai, Thailand, 7–9 January 2018. [Google Scholar]
  17. Dabass, J.; Vig, R. Biomedical Image Enhancement Using Different Techniques—A Comparative Study. In Data Science and Analytics; Springer: Singapore, 2018; pp. 260–286. [Google Scholar]
  18. An, F.P.; Ma, X.M.; Bai, L. Image fusion algorithm based on unsupervised deep learning-optimized sparse representation. Biomed. Signal Process. Control. 2022, 71, 103140. [Google Scholar] [CrossRef]
  19. Hermessi, H.; Mourali, O.; Zagrouba, E. Multimodal medical image fusion review: Theoretical background and recent advances. Signal Process. 2012, 183, 108036. [Google Scholar] [CrossRef]
  20. Du, J.; Li, W.; Lu, K.; Xiao, B. An overview of multi-modal medical image fusion. Neurocomputing 2016, 215, 3–20. [Google Scholar] [CrossRef]
Figure 1. The proposed image fusion model’s flowchart.
Figure 1. The proposed image fusion model’s flowchart.
Engproc 59 00225 g001
Figure 2. YSGO Algorithm Flowchart [12].
Figure 2. YSGO Algorithm Flowchart [12].
Engproc 59 00225 g002
Table 1. Performance metrics.
Table 1. Performance metrics.
S. NoParameterEquation
1.Root Mean Square Error (RMS): The spectral quality can be seen in this parameter. The difference between the reference image and the fused image is used to calculate it. R M S E = i = 1 M j = 1 N ( I R i , j I F i , j ) 2 M N [15]
where   M N   denotes   the   size   of   the   image .   I R   and   I F is the referenced and fused image.
2.Peak-Signal-to-Noise-Ratio (PSNR): to determine the overall quality of the fused image, this parameter is computed to its maximum value in image fusion methods. P S N R = 10 log 10 L 2 M S E [16]
where M S E denotes the mean square error and L denotes the number of grey levels in the image. In a grey scale image, its value is 255 .
3.Mutual Information (MI) is a parameter used to compare reference and fused image similarity. The more details and textual information there are, the better the value of MI. M I = i = 1 M j = 1 N h I R I F ( i , j ) × log 2 h I R I F ( i , j ) h I R i , j h I F ( i , j ) [17]
where h I R I F denotes the joint grey level histogram of I R and I F .
4.Structural Similarity Index Measure (SSIM): This metric assesses how comparable the reference and fused images’ structures are. Its value fluctuates between −1 and 1. Value 1 demonstrates that structural information is preserved and that both images are similar. S S I M = ( 2 μ I R μ I F + C 1 ) ( 2 σ I R I F + C 2 ) ( μ I R 2 + μ I F 2 + C 1 ) ( μ I R 2 + μ I F 2 + C 2 ) [17]
where μ I R μ I F means that the referenced and fused image’s mean intensity values are shown. C 1 C 2 are constant.
5.Entropy: the average quantity of information contained in the fusion image is indicated by the size of information entropy. E = i = 1 L p i log 2 p i
p i   is   the   ratio   of   the   number   of   pixels   whose   Gray   value   is   i   to   the   total   number   of   pixels   in   the   image ,   and   satisfies   i = 1 L p i = 1.
The higher value of entropy represents the maximum information contained by image [18].
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Goel, A.; Wasim, J.; Srivastava, P.K.; Malik, K.; Singh, M. Image Fusion Techniques Based on Optimization Algorithms: A Review. Eng. Proc. 2023, 59, 225. https://doi.org/10.3390/engproc2023059225

AMA Style

Goel A, Wasim J, Srivastava PK, Malik K, Singh M. Image Fusion Techniques Based on Optimization Algorithms: A Review. Engineering Proceedings. 2023; 59(1):225. https://doi.org/10.3390/engproc2023059225

Chicago/Turabian Style

Goel, Anamika, Javed Wasim, Prabhat Kumar Srivastava, Kanika Malik, and Monika Singh. 2023. "Image Fusion Techniques Based on Optimization Algorithms: A Review" Engineering Proceedings 59, no. 1: 225. https://doi.org/10.3390/engproc2023059225

Article Metrics

Back to TopTop