Next Article in Journal
An Analytical Model for Understanding Active Directivity Noise Control with Near-Field Error Sensing
Next Article in Special Issue
Fast and Accurate Visual Tracking with Group Convolution and Pixel-Level Correlation
Previous Article in Journal
Fundamental Investigations of the Deformation Behavior of Single-Crystal Ni-Mn-Ga Alloys and Their Polymer Composites via the Introduction of Various Fields
Previous Article in Special Issue
A Novel Constant Damping and High Stiffness Control Method for Flexible Space Manipulators Using Luenberger State Observer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Image Denoising Method for a Visible Light Camera in a Complex Sky-Based Background

Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2023, 13(14), 8484; https://doi.org/10.3390/app13148484
Submission received: 30 May 2023 / Revised: 18 July 2023 / Accepted: 20 July 2023 / Published: 22 July 2023
(This article belongs to the Special Issue Recent Advances in Robotics and Intelligent Robots Applications)

Abstract

:
For space target images captured by a sky-based visible light camera, various conditions are influenced by the imaging system itself and the observation environment; these conditions include uneven image background intensity, complex noise, stray light composition, and diverse target forms. A mean wavelet transform algorithm is proposed. This algorithm initially performs mean filtering and wavelet transform filtering on the noise-containing space target images and then performs a wavelet inverse transform on the filtered results to generate the final image. The experimental results show that our algorithm has good denoising performance and can effectively maintain the image details.

1. Introduction

Studies on sky-based image denoising techniques have focused on extracting space targets by processing the image prior to object detection. The objective is to suppress or eliminate inhomogeneous background noise.
Methods for suppressing or eliminating image noise can be categorized into two groups: transform domain methods [1,2,3,4] and spatial domain methods [5,6,7,8,9,10,11,12,13]. The transform domain approach involves removing the noise in the transformed domain of the image and then inverting the transform to achieve image denoising. Spatial domain methods, on the other hand, involve directly operating on the original image, taking advantage of the grayscale values in the neighborhood of the pixels to perform relevant operations on the pixels.
Unlike conventional images, in space target images, the captured images not only contain Gaussian noise, salt-and-pepper noise, Poisson noise, speckle noise, etc., but also the imaging size, shape, and gray value of space targets are highly similar to noise, and the background strength is unevenly distributed due to the scattered light. These factors cause difficulty for traditional image denoising algorithms [14,15,16,17,18,19] to be applied to the denoising of star maps, easily causing loss of target information or false targets. New requirements for image processing technology are needed to suppress or eliminate the complex background interference in star maps while preserving space target information.
This paper proposes a mean curvelet transform method. Curvelet transform is a multiscale and directionally sensitive transformation method that extracts detailed features from an image. On the other hand, mean processing is a basic image smoothing method that reduces noise by calculating the average value of pixel neighborhoods. By integrating these two methods, one can first utilize the curvelet transform to extract detailed features from the image and then use mean processing to smooth the extracted details, thereby reducing noise. The effectiveness of the proposed algorithm is verified through simulation experiments.

2. Materials and Methods

2.1. Wavelet Transform and Curvelet Transform

2.1.1. Wavelet Transform

In the 1980s, J. Morlet first proposed the concept of wavelets [20]. In the same period, mathematicians Y. Meyer and S. Mallat further studied the Mallat algorithm based on previous work, and the wavelet transform was officially created [21,22]. From a theoretical perspective, wavelet transform theory evolved from Fourier analysis. Fourier analysis can be used to analyze the characteristics of data separately using the time axis and frequency axis transformations, but it cannot reflect the characteristics of signals changing simultaneously along both the time and frequency axes. It can only consider one of the two axes, allowing a global understanding of the signal’s characteristics, but is unable to describe the local characteristics of a specific time-frequency region of the signal. In reality, signals are often nonstationary, and it is difficult to maintain stability. Most signals are nonstationary, and the local time-frequency characteristics precisely describe the distinctive features of nonstationary signals. While scaling the signal by shrinking or expanding the variable multiplier and transforming the signal through smooth distance shifts, the wavelet transform can decompose the signal into multiple levels, surpassing the limitations of the Fourier transform in this aspect.
The traditional approach of the wavelet transform is to separate the high-frequency and low-frequency signals for processing. The low-frequency signal, which undergoes minimal changes, is expanded to analyze the fine details of the rapidly changing parts in the high-frequency signal. In the time dimension, the signal is further segmented, while in the low-frequency signal, the frequency is divided more finely, adjusting the time-frequency characteristics analysis based on the specific high and low frequencies of the signal.
The continuous wavelet transform involves extensive data operations, so it needs to be discretized during implementation to simplify the form. In the expression of continuous wavelets, the binary number scale parameter a and the translation parameter b remain continuous, and only the parameters in the semi-discrete continuous wavelet formula are discretized while the variable t remains unchanged. The discrete form of the wavelet transform is:
W Ψ f 1 2 j , b = f ( t ) 2 j / 2 Ψ 2 j ( x b ) d x
This transformation is also known as the binary wavelet transform, which meets the condition:
A j z 2 j ω 2 B
Here, 0 < A < B < , where A and B are constants. For a certain level of decomposition, the binary wavelet transform can be considered a function of the variable b in Equation (2), and b is not discretized. Equation (2) can be transformed into a convolution operation expression, which is shown below:
W s f ( x ) = f * Ψ s ( x )
By scaling the wavelet mother function Ψ ( x ) with a factor of s (where s = 2 j ), we obtain the wavelet function as follows:
Ψ s ( x ) = 1 s Ψ s x s
Let us assume f τ ( x ) = f ( x τ ) . Then,
W 2 j f τ ( x ) = W 2 j [ f ( x ) ]
If we first shift the function f along the scale and then perform the binary wavelet transform, it is equivalent to first applying the binary wavelet transform and then shifting it along the scale. In other words, W 2 j f possesses the properties of f, which is the meaning of translation invariance.
Although the binary wavelet transform is in a discrete form, the variable b is not discretized along the scale, and the transformation of the wavelet transform at different scales is not discrete. This translation invariance exhibits a zooming characteristic. By adjusting b to enlarge or reduce the scaling factor, the wavelet transform acts like a microscope, enabling the study of finer or coarser details in the signal.

2.1.2. Limitations of the Wavelet Transform and Introduction of the Ridgelet Transform

Although the wavelet transform, known as the “mathematical microscope”, can express the details of image waveforms and has significant application value and advantages, it has a limitation due to the lack of anisotropy in wavelet bases. The wavelet transform can only reflect discontinuities or points where the derivative is zero in 1D images, and it cannot capture the edge features of 2D spatial images. Therefore, at the end of the 20th century, Candes developed the ridgelet transform algorithm [23], which expresses line singularities and can provide a good approximation of line singularities in multivariable functions. The expression of the ridgelet transform is given by:
Ψ γ ( x ) = a 1 2 Ψ [ u , x ] b a
where Ψ γ is called a ridgelet generated from Ψ under specific admissibility conditions. a represents the scale parameter, "u" represents the direction parameter, and b represents the position parameter. Ψ γ has rapid decay properties that satisfy:
Ψ ( t ) d t = 0
For functions with several variables, the ridgelet transform is defined as:
R ( f ) ( r ) = f , Ψ r
The ridgelet transform represents the direction by adding an additional parameter to the wavelet basis function. In the cross-section, the basis of the ridge is the wavelet. The continuous ridgelet transform in 2D space is expressed in Equation (9).
C R T f a , b x , b y = f ( x , y ) Ψ a , b x , b y ( x , y ) d x d y
When comparing the expressions of the 2D wavelet transform, as shown in Equation (10), with the ridgelet transform, we can observe the following contrast: in the wavelet transform formula, the parameter is a point, whereas in the ridgelet transform, the parameter is a line.
w f a , b x , b y = f ( x , y ) Ψ a , b x , b y ( x , y ) d x d y
The introduction of the ridgelet transform solves the problem of the wavelet transform’s inability to describe 2D edge singularities. However, the ridgelet transform still has limitations when dealing with functions with curve singularities. Moreover, in practical images, edges are rarely straight lines, which limits the widespread application of the ridgelet transform. Additionally, the ridgelet transform suffers from high computational redundancy. Therefore, in the last year of the 20th century, Candes and Donoho developed the curvelet transform and formulated its theory. The curvelet transform effectively captures the singular points and characteristics of curves in images. The basis of the curvelet transform exhibits directionality and anisotropy, allowing for the optimal approximation and description of nonlinear parts of 2D images. Thus, curvelet analysis combines the properties of wavelet multiresolution and locality with directionality, providing an optimal approach for representing images with these three features.

2.1.3. Curvelet Transform

The curvelet transform, introduced by E.J. Cande and David L. Donoho at the end of the 20th century, is a method aimed at addressing the limitations of the wavelet transform in expressing the anisotropy of image boundaries and line features [24]. While the wavelet transform is suitable for representing isotropic objects, it falls short in capturing the directional variations in image boundaries and line features. Therefore, the curvelet transform was proposed as a first-generation extension of the wavelet transform, preserving the advantages of multiresolution and time-frequency localization while introducing the ability to perform multiscale, translation, and multidirectional transformations. This enables the curvelet transform to better describe the singular changes in images, which cannot be effectively expressed by the wavelet transform.
Numbered lists can be added as follows:
  • First-generation curvelet. The first-generation curvelet transform primarily involves subband decomposition and a multiscale ridgelet transform, which is a nonadaptive representation method. As shown in Figure 1, the first-generation curvelet includes subband decomposition, smoothing partitioning, normalization, and ridgelet analysis. The decomposition process introduces significant data computation, making it quite complex in digital implementation. Based on this, E.J. Candes and Donoho proposed a second-generation transform three years later, and this transform is an improved curvelet algorithm that is easier to understand and simpler to implement. Furthermore, the two scholars introduced a fast curvelet algorithm that directly divides frequencies, eliminating the need for the ridgelet transform. This construction significantly differs from the first-generation curvelet algorithm, resulting in a reduction in data and computational complexity.
  • Second-generation curvelet transform. The construction of the second-generation curvelet transform differs from the construction of the first generation transform, but the second-generation transform is an improvement of the first-generation transform. The idea behind the first-generation curvelet transform is to partition the target function space into infinitely small blocks to approximate curves as straight lines; then, local ridgelet analysis is utilized. In contrast, the second-generation curvelet transform only retains the abstract principles of the ridgelet transform, such as the framework and tight frame, while discarding its specific computational methods. The continuous curvelet transform serves as an example.
    The formula for the continuous curvelet transform in the time domain is as follows:
    C ( i , j , k ) = f , φ i , j , k = R 2 f ( x ) φ i , j , k ( x ) d x ¯
    where
    φ i , j , k ( x ) = φ j R θ x x k j
    indicates the position at
    x k ( j ) = R 0 1 k 1 × 2 j , k 2 × 2 j / 2
    When the angle position of all points at 2( j ) is considered, the curvelet transform can be obtained by rotating the curvelet basis φ j by a certain angle θ and shifting it by K.
    θ = 2 π × 2 [ j / 2 ] × l , l = 0 , 1 , , 0 θ 1 2 π
    The parameter k = k 1 , k 2 z 2 represents the displacement sequence in the 2-dimensional space.

2.1.4. Curvelet Transform Algorithm

The algorithm for the curvelet transform is described as follows:
(1)
Perform a J-level decomposition of the original image I to obtain the subband sequences CJ and DJ;
(2)
Set the initial block size (i.e., the finest scale block) to BMIN, and let B1 = BMIN;
(3)
For j = 1 to J:
-
Divide Dj into subblocks of size BJ with overlapping regions;
-
Apply the discrete ridgelet transform to each subblock;
-
If (j mod 2 = 1), set Bj + 1 = 2Bj; otherwise, set Bj + 1 = Bj.
The curvelet transform itself is a redundant transformation. If the number of wavelet decompositions in the first step is J, then the overall redundancy factor of the entire transform is 16J + 1. Since each step of the above decomposition is reversible, it is possible to construct the corresponding inverse transform algorithm to completely reconstruct the original image.

2.1.5. Analysis of Wavelet Transform Denoising

The wavelet transform method solves the correlation values of the spatial image in a set of established anisotropic bases. The method of wavelet transform denoising is as follows: first, the noisy image is decomposed into subbands of different scales, and the wavelet transform coefficients corresponding to each subband are obtained through the Plancherel operation; then, the obtained wavelet transform coefficients are processed by the hard threshold method (abandoning the smaller transform coefficients and retaining the larger transform coefficients) to filter out noise in the image and retain the edge characteristics of the image. The expression of the hard threshold in the wavelet transform denoising is as follows:
C 0 ( i , h ) = C ( i , h ) , | C ( i , h ) | T i 0 , | C ( i , h ) | < T i
where C ( i , h ) is the wavelet coefficient at scale i and direction h, T i is the threshold corresponding to different scales, and C 0 ( i , h ) is the wavelet transform coefficient after hard threshold denoising; the expression for selecting the threshold T i is as follows:
T i = k i , σ · σ i
where σ is the noise standard deviation, σ i is the noise standard deviation corresponding to the subband after transformation, and k i is the adaptive constant corresponding to each subband. The Monte Carlo algorithm with strong adaptability is used to estimate the noise standard deviation in this method.

2.2. Analysis of Mean Filtering Algorithm

The principle of mean filtering is to calculate the mean of the pixels in the current point neighborhood to replace the gray value of the current point, and this is a linear filtering algorithm. It is equivalent to convolving the image with a filtering template of a certain size, filtering out the objectives with poor correlation, and obtaining a smooth star map background. The formula for mean filtering is as follows:
g ( x , y ) = W ( i , j ) f ( x , y )
where f ( x , y ) is the original image data, g ( x , y ) is the filtered image data, and W ( i , j ) is the filtering template, which is represented as follows:
W ( i , j ) = 1 M × N 1 1 1 1 1 1 1 1 1 M × N

2.3. Mean Wavelet Transform Denoising Process

In a space target image, the background information is primarily represented by pixels with low grayscale values, while the edges and details of the target are mainly represented by pixels with higher grayscale values, assuming there is no contamination from stray light. The curvelet transform is known for its multi-scale and multi-directional sensitivity, which enables effective capture of detailed features in images. However, in practical scenarios, space target images usually contain significant amounts of non-uniform stray light, which can vary in intensity. When the grayscale value of the background information is similar to that of the target edges, the curvelet transform may struggle to effectively capture these subtle changes. As a result, it becomes challenging to distinguish the target from the background, leading to blurry outcomes.
To address this issue, we introduce the mean filtering algorithm to enhance contrast and aid in distinguishing the target from the background. Subsequently, we apply the curvelet transform for filtering. Nevertheless, mean filtering has its limitations, such as insufficient smoothing of the grayscale values in the background information and potential blurring of the target edges. These limitations restrict the denoising capability of the current transform. Furthermore, the non-uniformity of stray light in the space target image can result in regions with no stray light or severely weak stray light, where pixel values already contain valid target information. However, the mean filtering averages over these regions, potentially leading to blurring of object detail.
Hence, in this paper, we propose an image processing approach based on weighted averaging. We denote the two source images as X and Y, where the X image is denoised directly using the curvelet transform and the Y image is filtered by the mean prior to the current transform. The final composite image, denoted as Z, is obtained by taking the weighted average of the two images. The mean curvelet denoising formula can be expressed as follows:
C ( Z , p ) = K C ( X , p ) + ( 1 k ) C ( Y , p )
The weighting coefficient, denoted as k, its expression is as follows:
k = a × x = M x N x y = M y N y I ( x , y ) 2 N N x M x + 1 × N y M y + 1
M x , M y , N x , N y represent the coordinates of a pixel in the image, I ( x , y ) represents the value of the pixel at coordinates (x, y), N represents the number of gray levels in the image, a represents an empirical value within the range of a [ 0 , 1 ] .
In regions of the background where strong stray light is present, we amplify the k value to enhance the proportion of the denoised image Y, which is obtained after mean filtering, in the composite image Z. Conversely, in regions with weak stray light in the background, we reduce the k value to enhance the proportion of the denoised image X, obtained through direct curvelet filtering, in the composite image Z. The flowchart of the mean wavelet transform denoising algorithm is shown in Figure 2.

3. Results

This paper verifies the effectiveness of the mean wavelet transform denoising algorithm through two comparison experiments.

3.1. Lena Image of a Hat Part in the Reverse Color Image Experiment

The resolution of the inverse color image of the hat part of the Lena image is a gray image of 512 × 512 with an 8-bit grayscale. We use the mean filtering algorithm, wavelet filtering algorithm and mean wavelet filtering algorithm to remove the white noise in the original image, and the experimental results are shown in Figure 3 and Table 1.
From Figure 3, the mean wavelet transform algorithm restores the slanted lines on the hat more realistically than the other methods.

3.2. Real Star Image Experiment

The real star image resolution is a 2048 × 2048 gray image with a gray level of 12 bits. We use the mean filtering algorithm, wavelet filtering algorithm and mean wavelet filtering algorithm to filter out the white noise in the original image. The experimental results are shown in Figure 4 and Table 2.
As shown in Figure 4, in the real star image, the mean wavelet transform can not only remove nonuniform noise but also significantly improve the signal-to-noise ratio of the image.

4. Discussion

This paper proposes the mean-curvelet transform denoising algorithm to address the issues of uneven background intensity, complex noise and interference, and diverse target shapes in images of space targets captured by visible light cameras under the constraints of the imaging system and observation environment. The algorithm applies mean filtering and curvelet transform filtering to the noisy space target images separately and then applies the inverse curvelet transform to the filtered results to generate the final image. The experimental results show that this algorithm effectively improves the signal-to-noise ratio of images.
Furthermore, since the fixed threshold is used in the curvelet transform in this paper, we will study the application of the variable threshold curvelet transform to the algorithm in future experiments.
However, since the curvelet transform is mostly implemented on hardware platforms such as X86 and X64, which cannot tolerate long-term operation in space radiation environments, the algorithm proposed in this paper currently does not have the capability of real-time processing in orbit. In future studies, we will select some hardware platforms with radiation resistance indicators, such as DSP or FPGA, and apply the algorithm on these hardware platforms to achieve real-time processing in orbit.

5. Conclusions

Based on our study, and compared with existing methods, our proposed mean curvelet transform method not only has better suppression performance for images of space targets with uneven background intensity, complex noise and clutter but also maintains good target shape diversity and preserves image details and other information, significantly improving the image quality. Therefore, our proposed algorithm has great application value for denoising space target images. In future research, we will further investigate the performance of the algorithm.

Author Contributions

Project administration, Z.M.; data curation, Q.Z., X.Q., W.L. and S.W.; X.C. wrote the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

Key Technology R&D Program of Jilin Province, China (No. 20220203053SF); National Natural Science Foundation of China (No. 12103053).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are unavailable due to privacy or ethical restrictions.

Acknowledgments

Thanks to all the Project team members.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mallat, S.G. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 11, 674–693. [Google Scholar] [CrossRef] [Green Version]
  2. Donoho, D.L.; Johnstone, I.M. Ideal spatial adaptation by wavelet shrinkage. Biometrika 1994, 81, 425–455. [Google Scholar] [CrossRef]
  3. Gu, G.Q.; Wang, K.F.; Xu, X. Denoising in digital speckle pattern interferometry using fast discrete curvelet transform. Imaging Sci. J. 2014, 62, 106–110. [Google Scholar] [CrossRef]
  4. Chen, Z.A.; Hu, Z. Remote sensing image denoising based on improved wavelet threshold algorithm. Bull. Surv. Mapp. 2018, 28–31. [Google Scholar] [CrossRef]
  5. Narendra, P.M. A separable median filter for image noise smoothing. IEEE Trans. Pattern Anal. Mach. Intell. 1981, 3, 20–29. [Google Scholar] [CrossRef] [PubMed]
  6. Eng, H.L.; Ma, K.K. Noise adaptive soft-switching median filter. IEEE Trans. Image Process. Publ. IEEE Signal Process. Soc. 2001, 10, 242–251. [Google Scholar] [CrossRef] [Green Version]
  7. Guariglia, E. Harmonic Sierpinski Gasket and Applications. Entropy 2018, 20, 714. [Google Scholar] [CrossRef]
  8. Zheng, X.; Tang, Y.Y.; Zhou, J. A Framework of Adaptive Multiscale Wavelet Decomposition for Signals on Undirected Graphs. IEEE Trans. Signal Process. 2019, 67, 1696–1711. [Google Scholar] [CrossRef]
  9. Guido, R.C.; Pedroso, F.; Contreras, R.C.; Rodrigues, L.C.; Guariglia, E.; Neto, J.S. Introducing the Discrete Path Transform (DPT) and its applications in signal analysis, artefact removal, and spoken word recognition. Digit. Signal Process 2021, 117, 103158. [Google Scholar] [CrossRef]
  10. Guariglia, E.; Silvestrov, S. Fractional-Wavelet Analysis of Positive Definite Distributions and Wavelets on D’(C) D’(C)[C]//Engineering Mathematics II: Algebraic, Stochastic and Analysis Structures for Networks, Data Classification and Optimization; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 337–353. [Google Scholar]
  11. Yang, L.; Su, H.; Zhong, C.; Meng, Z.; Luo, H.; Li, X.; Tang, Y.Y.; Lu, Y. Hyperspectral image classification using wavelet transform-based smooth ordering. Int. J. Wavelets Multiresolution Inf. Process. 2019, 17, 1950050. [Google Scholar] [CrossRef]
  12. Guariglia, E.; Guido, R.C. Chebyshev Wavelet Analysis. J. Funct. Spaces 2022, 2022, 5542054. [Google Scholar] [CrossRef]
  13. Berry, M.V.; Lewis, Z.V.; Nye, J.F. On the Weierstrass-Mandelbrot fractal function. Proc. R. Soc. Lond. Math. Phys. Sci. 1980, 370, 459–484. [Google Scholar]
  14. Beck, A.; Teboulle, M. Fast gradient-based algorithms for constrained total variation image denoising and deblurring problems. IEEE Trans. Image Process 2009, 18, 2419–2434. [Google Scholar] [CrossRef] [Green Version]
  15. Fu, B.; Li, W.W.; Fu, Y.P.; Song, C.M. An image topic model for image denoising. Neurocomputing 2015, 169, 119–123. [Google Scholar] [CrossRef]
  16. Yuan, J.; He, J. Blocking sparse method for image denoising. Pattern Anal. Appl. 2021, 24, 1125–1133. [Google Scholar] [CrossRef]
  17. Liao, X.R. An improved ROF denoising model based on time-fractional derivative. Front. Inf. Technol. Electron. Eng. 2020, 21, 856–865. [Google Scholar] [CrossRef]
  18. Boyat, A.; Joshi, B.K. Image denoising using wavelet transform and median filtering. In Proceedings of the 2013 Nirma University International Conference on Engineering (NUiCONE), Ahmedabad, India, 28–30 November 2013; pp. 1–6. [Google Scholar] [CrossRef]
  19. Song, Q.; Ma, L.; Cao, J.; Han, X. Image Denoising Based on Mean Filter and Wavelet Transform. In Proceedings of the 2015 4th International Conference on Advanced Information Technology and Sensor Application (AITS), Harbin, China, 21–23 August 2015; pp. 39–42. [Google Scholar] [CrossRef]
  20. Morlet, J. Wave Propagation and Sampling Theory. Geophysics 1982, 47, 222–236. [Google Scholar] [CrossRef] [Green Version]
  21. Meyer, Y. Wavelets: Algorithms & Applications; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 1993. [Google Scholar]
  22. Mallat, S. A Wavelet Tour of Signal Processing: The Sparse Way; Academic Press: Cambridge, MA, USA, 2009. [Google Scholar]
  23. Candes, E.J. Ridgelets: Theory and Applications. Ph.D. Thesis, Stanford University, Stanford, CA, USA, 1998. [Google Scholar]
  24. Candès, E.J.; Donoho, D.L. Ridgelets: A key to higher-dimensional intermittency? Philos. Trans. R. Soc. Lond. Ser. Math. Phys. Eng. Sci. 1999, 357, 2495–2509. [Google Scholar] [CrossRef]
Figure 1. Flowchart of The first-generation curvelet transform algorithm.
Figure 1. Flowchart of The first-generation curvelet transform algorithm.
Applsci 13 08484 g001
Figure 2. Flowchart of the mean wavelet transform denoising algorithm.
Figure 2. Flowchart of the mean wavelet transform denoising algorithm.
Applsci 13 08484 g002
Figure 3. Lena image filtering results: (a) original image with noise; (b) mean filtered image; (c) wavelet transformed filtered image; (d) mean wavelet filtered image.
Figure 3. Lena image filtering results: (a) original image with noise; (b) mean filtered image; (c) wavelet transformed filtered image; (d) mean wavelet filtered image.
Applsci 13 08484 g003
Figure 4. Real star image filtering results: (a) original image with noise; (b) mean filtering image; (c) wavelet filtering image; (d) mean wavelet filtering image.
Figure 4. Real star image filtering results: (a) original image with noise; (b) mean filtering image; (c) wavelet filtering image; (d) mean wavelet filtering image.
Applsci 13 08484 g004
Table 1. Quality of the lena image of a hat part reverse color image and experiment images obtained via different methods.
Table 1. Quality of the lena image of a hat part reverse color image and experiment images obtained via different methods.
MethodPeak Signal to Noise RatioNote
Original noisy image21.13-
Mean filtering24.45large amount of distortion
Wavelet filtering28.95small amount of distortion
Mean wavelet filtering29.55very little distortion
Table 2. Quality of real star image and experimental images obtained via different methods.
Table 2. Quality of real star image and experimental images obtained via different methods.
MethodPeak Signal to Noise RatioNote
Original noisy image6.20-
Mean filtering10.23small amount of distortion
Wavelet filtering6.95large amount of distortion
Mean wavelet filtering12.94very little distortion
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ma, Z.; Zhao, Q.; Che, X.; Qi, X.; Li, W.; Wang, S. An Image Denoising Method for a Visible Light Camera in a Complex Sky-Based Background. Appl. Sci. 2023, 13, 8484. https://doi.org/10.3390/app13148484

AMA Style

Ma Z, Zhao Q, Che X, Qi X, Li W, Wang S. An Image Denoising Method for a Visible Light Camera in a Complex Sky-Based Background. Applied Sciences. 2023; 13(14):8484. https://doi.org/10.3390/app13148484

Chicago/Turabian Style

Ma, Zelong, Qinglei Zhao, Xin Che, Xinda Qi, Wenxian Li, and Shuxin Wang. 2023. "An Image Denoising Method for a Visible Light Camera in a Complex Sky-Based Background" Applied Sciences 13, no. 14: 8484. https://doi.org/10.3390/app13148484

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop