Next Article in Journal
Machine-Learned Free Energy Surfaces for Capillary Condensation and Evaporation in Mesopores
Next Article in Special Issue
Research on Color Image Encryption Algorithm Based on Bit-Plane and Chen Chaotic System
Previous Article in Journal
Improved Dividend Estimation from Intraday Quotes
Previous Article in Special Issue
S2A: Scale-Attention-Aware Networks for Video Super-Resolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Synthetic Aperture Radar Image Despeckling Based on Multi-Weighted Sparse Coding

School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(1), 96; https://doi.org/10.3390/e24010096
Submission received: 9 November 2021 / Revised: 4 January 2022 / Accepted: 4 January 2022 / Published: 7 January 2022
(This article belongs to the Special Issue Advances in Image Fusion)

Abstract

:
Synthetic aperture radar (SAR) images are inherently degraded by speckle noise caused by coherent imaging, which may affect the performance of the subsequent image analysis task. To resolve this problem, this article proposes an integrated SAR image despeckling model based on dictionary learning and multi-weighted sparse coding. First, the dictionary is trained by groups composed of similar image patches, which have the same structural features. An effective orthogonal dictionary with high sparse representation ability is realized by introducing a properly tight frame. Furthermore, the data-fidelity term and regularization terms are constrained by weighting factors. The weighted sparse representation model not only fully utilizes the interblock relevance but also reflects the importance of various structural groups in despeckling processing. The proposed model is implemented with fast and effective solving steps that simultaneously perform orthogonal dictionary learning, weight parameter updating, sparse coding, and image reconstruction. The solving steps are designed using the alternative minimization method. Finally, the speckles are further suppressed by iterative regularization methods. In a comparison study with existing methods, our method demonstrated state-of-the-art performance in suppressing speckle noise and protecting the image texture details.

1. Introduction

Synthetic aperture radar (SAR) systems provide fine-resolution images regardless of the weather conditions. Therefore, they are widely used in environmental monitoring, military reconnaissance, and other fields [1,2,3]. Nevertheless, coherent imaging commonly causes multiplicative speckle noise in SAR images, which limits the analysis and interpretation of the scene and reduces the visualization of the images [4,5,6]. To avoid these problems, the speckle noise should be reduced as far as possible while retaining the structural features and texture information in SAR images [7,8,9,10,11].
In recent work, much attention was devoted to spatial-domain adaptive filtering based on Bayesian statistical estimation. Techniques such as linear minimum mean-square error (LMMSE)-based filtering [12] and maximum posterior probability (MAP)-based filtering [13] were introduced at this time. LMMSE filtering is performed by Lee [14], Frost [15], and Kuan filters [16], whereas MAP filtering [17] employs Gaussian and gamma filters. Although filters are easily implemented and have low computational complexity, they lose much of the texture information [18]. Despeckling methods in the wavelet domain have also been proposed. Initially, multiplicative noise was converted to additive noise through homomorphic filtering. Later, the normal inverse Gaussian model was introduced [19], which removes speckle noise without the shortcomings of spatial filter wave technology and has set a new direction for SAR image processing. Transform-domain-based methods better preserve the edges than spatial-domain filters, but tend to generate artifacts. Another important property of SAR images is nonlocal self-similarity, which is easily incorporated in SAR image despeckling. For example, nonlocal means (NLM) filtering [20] exploits self-similarity for image denoising. In an NLM filter, the resemblance between image patches is measured by the Euclidian distance and weighted averaging is performed in nonlocal regions. Inspired by the success of the NLM denoising filter, the authors of [21] developed the probabilistic patch-based (PPB) algorithm, in which the similarity between image patches in the SAR image is determined by the noise distribution rather than the traditional Euclidian distance.
In recent years, the continuous development of sparse theory has sparked interest in image denoising methods based on sparse representation [22,23,24,25,26]. In these methods, the most critical steps are selecting the appropriate dictionary, constructing the sparse model, and estimating the sparse coefficient. The earliest sparse representation methods, such as the discrete cosine transform and wavelet transform, relied on fixed, over-complete dictionaries. However, fixed dictionaries lack sufficient flexibility and self-adaptability for disparate image structures and the exact expression of signals. A more flexible dictionary would better characterize the prior information of the images. Recently, image patches have been identified as a flexible sparse target. Unlike the whole image, a patch is low-dimensional and overlapped patches provide opportunities for adaptively learning a dictionary or transforming the image into a sparse representation. Sparse dictionaries of training image patches extracted from images are known to ensure sparse representation coefficients for specific image structures [27,28]. The famous K singular value decomposition (K-SVD) algorithm [29] assumes that each patch has a sparse representation in a global over-complete dictionary. Each atom in the dictionary is then determined by minimizing the fitting error of the represented patches under the sparsity constraint. Considering the similarity and geometric direction of image patches, the authors of [30] classified image patches and learned an orthogonal dictionary for each type of image patch. As this method sparingly represents the types of image patches, it achieves faster dictionary learning, sparser coding, and better image reconstruction quality than the K-SVD algorithm. Unlike fixed dictionaries, a learned dictionary can adapt to the structural features of images and effectively obtains the sparse representations of images.
In recent works, the sparsity and self-similarity of images are often combined to improve the representation performance. The BM3D algorithm [31] combines similar image patches with three-dimensional (3D) transformation to enhance the sparsity level of 3D groups. For this purpose, BM3D performs a 3D wavelet transform on the groups and estimates the real-image wavelet coefficients by wiener filtering, which effectively removes Gaussian white noise from the images. The SAR-BM3D algorithm [32] extends the BM3D algorithm to SAR images. This algorithm replaces the wiener filter with a local LMMSE estimator that more accurately estimates the wavelet coefficients of the image, and hence reduces the speckle effect on SAR images. The fast adaptive nonlocal SAR despeckling (FANS) algorithm [33] is a faster and spatially-adaptive version of SAR-BM3D, which ensures a better speckle rejection in homogeneous areas. Dong et al. [34] proposed a nonlocal centralized sparse representation model based on the traditional sparse representation model, which estimates the sparse coefficients by nonlocal averaging of the image patches. As the image patches are used for constructing the coefficient correction terms in the sparse representation, the sparse coefficient is close to the real-image coefficient. To estimate the real image, some researchers exploit the low-rank property of similar image patch sets [35] and apply a soft thresholding that handles the singular values of similar noisy image patch sets.
As is well known, image regions differ in structure, so the sparsities of various image regions differ in the transformation domain. For this reason, imposing constant constraints on the coefficients is unlikely to fully utilize the prior information. To characterize the structural differences in the image, weighted processing of the image coefficients is required. The authors of [36] proposed a weighted nuclear norm minimization (WNNM) algorithm. The standard nuclear norm minimization (NNM) algorithm treats all singular values equally and shrinks them with the same threshold. However, the equalization process of NNM ignores the differences between image singular values in collections of similar image patches with different image prior information. The WNNM algorithm treats different singular values by assigning them with different weights. The weighting fully utilizes the prior information in different singular values, thus improving the flexibility of the kernel parameter constraint. Inspired by the Schatten p-norm and WNNM, the authors of [37] developed the weighted Schatten p-norm minimization model, which replaces the nuclear norm in the standard NNM model with the Schatten p-norm. They also recommended a weight parameter for the Schatten p-norm that obtains a new low-rank regular term. The problem is then solved by a generalized threshold contraction method. The trilateral weighted sparse coding model [38] has achieved good results in practical image denoising. This model accurately locates different noise intensities in different image regions, introduces three weight matrices into the model, and characterizes the statistics of realistic noise and image priors with data fidelity term and regularization term of the sparse coding framework. The high significance of the weights ensures that the estimated sparse coding coefficients approach the real-image coefficients.
Inspired by the above weighting idea, we propose a novel SAR image despeckling model based on dictionary learning and multi-weighted sparse coding. In the proposed multi-weighted sparse coding (MWSC), the efficient representation capacity of the orthogonal dictionary is integrated with the strong sparsity constraint of the multi-weighting model. This work makes three main contributions to the existing literature. First, we establish our general analysis dictionary learning and MWSC-based SAR image despeckling model. The proposed model continuously and efficiently trains the orthogonal dictionaries through SVD. Second, we introduce a weight matrix for nonuniform regularization, which explicitly and reasonably characterizes the diverse nonlocal characteristic of a group rather than performing uniform nonlocal regularizations. Another weight matrix is introduced to the data-fidelity term for characterizing the noise property. Third, we design an alternative minimization solution step for simultaneous orthogonal dictionary learning, weight parameter update, nonconvex sparse coding, and image reconstruction. As demonstrated in subsequent experimental results, the orthogonal dictionary learning and weight constraints both contribute to the final performance improvements. In terms of speckle noise removal and preservation of details, the MWSC recovered higher-quality images than other SAR image despeckling methods.

2. Traditional Sparse Coding Model

Speckles in SAR images can be characterized by the following multiplicative noise model [39]:
y = x u ,
where y is the observed SAR image, x is the real value of the SAR image, and u is the fully developed speckle noise. First, the multiplicative noise model must be transformed into an additive noise model. To this end, the SAR image is preprocessed as
y = x u = x + x ( u 1 ) = x + n ,
where n is the observed additive noise, which depends on the signal. In sparse representation theory, the sparse coding problem of each patch y k over dictionary D is formulated as
c k = arg min c k 1 2 y k D c k 2 2 + λ c k P ,
where y k D c k 2 2 is the data-fidelity term, y k = R k y is the image patch, and R k is a defined operator that extracts patch y k from y . c k P is the regularization term denoting the image prior, and λ is the regularization parameter used for balancing the relationship between the sparse approximation error and the sparse constraint. The sparse coding process of each patch y k attempts to find a sparse coefficient c k over a given dictionary D . Therefore, most of the entries in the vector c k are zero or close to zero. Once c k is computed, the latent clean patch matrix x ^ k is estimated as x ^ k = D c k . The clean image x to be reconstructed is then sparsely represented by a set of sparse codes { c k } formulated as follows:
x = ( k = 1 n R k T R k ) 1 k = 1 n ( R k T D c k )

3. The Proposed SAR Despeckling Algorithm

3.1. Weighted Sparse Representation Model

As the observed image y is seriously damaged by speckle noise, the image patch-based sparse coding method exploits only the local sparsity, which is insufficient for estimating the sparse coefficients c k from y . The correlation among image patches is another important prior information of the image. To exploit this prior, we construct and utilize groups of patches (rather than a single patch) as the basic unit of sparse representation. To retain the correlation in the group, regularizing its group sparsity is an obvious solution; however, the group sparsity may not be sufficiently guaranteed for all groups, because the levels of similarity widely vary in different image regions. This variability is demonstrated on the “Lena” image in Figure 1. Two groups, labeled “A” and “B” in Figure 1, are found in the homogeneous region and a heterogeneous region, respectively. The enlarged patches in Figure 1b clarify the similarity differences between the groups “A” and “B.” In the current group-based sparse coding approaches, nonlocal similarity is enforced by uniform nonlocal regularizations with the same penalty parameter, such as the norm. If the same constraints are imposed on each image patch, the structural features of the image cannot be fully captured and utilized, and some details and texture information are inevitably lost from the image. In addition, the data-fidelity term in the sparse coding model ensures the authenticity of the sparse coefficient in the sparse representation process and conforms to the image despeckling process. In reality, the additive noise n depends on the signals in the SAR image and its level varies in different local patches. Therefore, in the actual SAR image despeckling process, the important work of the data-fidelity term y k D c k 2 2 is varying the local noise intensity.
To rectify the problems in traditional patch-based sparse representation, we build groups of similar patches and simultaneously exploit the local sparsity and the nonlocal self-similarity of SAR images in a unified framework. We first divide the image y into n overlapped patches, denoting each patch by a vector y i , i.e., i = 1 , 2 , , n . For each target image patch y i in the S × S training window, we search for N 1 best matched patches. Finally, all patches (the target image patch and all similar image patches) are stacked into a matrix Y i representing the group. This step requires a suitable metric for measuring the similarity between noisy patches. Here, the Euclidean distance is selected as the similarity metric between the target image patch y i and the candidate patch y i , j in the training window. The Euclidean metric is given by
d ( y i , y i , j ) = y i y i , j 2 2 .
The group Y i containing the searched similar patches is then represented as
Y i = [ R i , 1 y , R i , 2 y , , R i , N y ] = [ y i , 1 , y i , 2 , y i , N ] ,
where R i , j is an operator that extracts the j-th similar patch y i , j = R i , j y from y , and N is the number of similar patches in group Y i . As all image patches in Y i share a similar underlying structure, the column vectors of Y i are strongly correlated. The sparse coefficients of the internal patches of group Y i are collected into a coefficient matrix denoted by C = [c1, c2,…,cN].
The current group-based SAR image despeckling approaches impose the same constraint on each group. Because they ignore the varying local noise intensity, they cannot fully capture and utilize the nonlocal characteristics of the image. To improve the accuracy of C , we analyze its probability statistics. More specifically, we apply the maximum a posterior (MAP) estimation derived from Gaussian Naïve Bayesian formula to the coefficient matrix C :
C = arg max C   P ( C |   Y i ) = arg max C   P ( Y i |   C ) P ( C ) P ( Y i ) = arg max C   P ( Y i |   C ) P ( C ) = arg max C {   ln   P ( Y i |   C ) + ln P ( C ) }
In Equation (7), ln   P ( Y i |   C ) is the data-fidelity term. Its form depends on the statistical characteristics of the image noise that penalizes the differences between group Y i and the coefficient matrix C . As the noise intensity varies in different regions of the SAR image, we assume that the additive noise n follows a Gaussian distribution of noise standard deviations. Hence, the data-fidelity term P ( Y i |   C ) is defined as
P ( Y i |   C ) = j = 1 N 1 2 π σ j exp ( 1 2 σ j 2 y i , j D c j 2 2 ) ,
where σ j is the noise standard deviation of the additive noise in image patch y i , j .
By the group property, all image patches in Y i share a similar underlying structure. Therefore, their coefficient vectors under one dictionary should share the same sparse profile. Specifically, the elements in one row of C have the same magnitude. We thus assume that each row of C is independent of all other rows and that its scale parameters follow a Laplace distribution. Therefore, the prior term in Equation (7) is defined as
P ( C ) = j = 1 N l = 1 M 1 2 δ l exp ( | c j l | δ l ) = l = 1 M 1 ( 2 δ l ) N exp ( c l 1 δ l ) ,
where c l is denotes the l-th row vector of C and δ l are the scale parameters of c l . Substituting Equations (8) and (9) into Equation (7), the MAP estimation model of coefficient matrix C is obtained as
C = arg min C j = 1 N 2 ( y i , j D c j ) σ i , j 1 F 2 + l = 1 M δ l 1 c l 1 = arg min C ( Y i D C ) Q 1 F 2 + Q 2 1 C 1
where Q 1 = d i a g ( 2 σ 1 1 , , 2 σ N 1 ) and Q 2 = d i a g ( δ 1 , , δ M ) are two diagonal matrices. The design of Q 1 and Q 2 to characterize the varying statistics of realistic noise and the sparsity prior of SAR images is described in Section 3.2.
Unlike the traditional sparse representation model, we modify the data-fidelity term and regularization term with the weight matrices Q 1 and Q 2 , respectively. The weight matrix Q 1 characterizes the residual Y i D C in the data-fidelity term ( Y i D C ) Q 1 F 2 . When the diagonal element 2 σ j 1 is large, the speckle noise intensity is high and a stronger constraint is imposed on the weighted residual term. Conversely, when the speckle noise intensity is low, the constraint on the residual term is weakened. By introducing the weight matrix Q 1 into the data-fidelity term, we capture and utilize the realistic noise properties in different patches. Meanwhile, Q 2 in the regularization term reflects the importance of each element in C . Specifically, one weight in Q 2 is assigned to the elements in one row of C to enforce the same sparse profile among similar patches, whereas the elements of C in different rows are assigned different weights that change the strengths of the sparsity constraints. Overall, the weighted sparse constraint model not only reflects the varying local noise intensity, but also simultaneously utilizes the patch-level sparsity and relevance among patches, which effectively improves the adaptability of the model to real SAR images and the accuracy of the coefficient matrix C .

3.2. Adaptive Matrix Parameter Learning

Because the noise intensity varies in different image regions, the weight matrix Q 1 must learn to adaptively describe the statistical properties of the noise. In the actual speckle-reduction process, the weight matrix Q 1 is updated through iterative regularization. During the k-th iteration, the parameter σ j in weight Q 1 is updated as
σ j = γ a b s ( σ 2 - y j - y j k 2 2 / N ) ,
where σ is the noise benchmark parameter, γ < 1 is a control factor, which is tuned according to the intensity of noise. In general, a smaller γ is set for low noise level and a larger γ is set for high noise level. y j is the j-th image patch extracted from the original observation image y , and y j k is the j-th image patch extracted from the input image y k during the k-th iteration of iterative regularization (see Section 3.4. for details).
An effective regularization term that describes the image priors is required for a high-quality recovered image. In the proposed model given by Equation (10), the combinational regularization term Q 2 1 C 1 is precisely controlled by a weight matrix Q 2 describing the importance of the different elements of the coefficient matrix. In our method, the weight matrix Q 2 of the regularization term is learned from Y i for adaptively adjusting the weights of the coefficients. Clearly, Q 2 is individual to group Y i and effectively captures the specific structural information of that group.
Fixed dictionaries are unlikely to effectively express the signals in a sparse representation, but a trained dictionary is expected to enhance the sparsity of coefficients. The dictionary learning method in the traditional sparse coding model (Equation (3)) jointly optimizes the dictionary D and the coefficient vector c k . The dictionary D learned by this method adapts to a given whole image, not merely a group. When the same dictionary is shared by all image patches, it may not adapt to specific local structures in the image. Instead of learning a redundant dictionary for the whole image, we directly learn an individual dictionary D for each group Y i , which covers all patches in that group. The model also imposes a tight frame constraint D T D = I , which not only restrains the correlations among dictionary atoms, but also simplifies the complexity of the sparse decomposition operations.
Finally, the complete group-based dictionary learning and MWSC model is defined as follows:
( C , Q 2 , D ) = arg min C , Q 2 , D ( Y i D C ) Q 1 F 2 + Q 2 1 C 1   s . t .   D T D = I ,
where C , Q 2 and D are matrices to be solved.

3.3. Model Optimization

To facilitate the solution of Equation (12), we define an auxiliary variable A = Q 2 1 C . Equation (12) is then rewritten as
( A , Q 2 , D ) = arg min A , Q 2 , D ( Y i D Q 2 A ) Q 1 F 2 + A 1   s . t .   D T D = I .
Equation (13) includes three unknown variables. Because Equation (13) is nonconvex, it is solved by an alternative minimization method. In each iteration, Equation (13) is decomposed into three subproblems: A , Q 2 and D for learning the orthogonal dictionary D , updating the weight Q 2 , and sparse coding of A , respectively.
Given D and A , the subproblem of Equation (13) with respect to Q 2 is
Q 2 = arg min Q 2 ( Y i D Q 2 A ) Q 1 F 2   .
As D is an orthogonal dictionary, the subproblem of Q 2 can be rewritten as
Q 2 = arg min Q 2 D T Y i Q 1 Q 2 A Q 1 F 2 .
The objective function Equation (15) is separable into the diagonal elements of Q 2 and its minimization problem can be consequently decoupled into multiple independent scalar optimizations of the form
δ l = arg min δ l ( D T Y i Q 1 ) l δ l ( A Q 1 ) l F 2 ,
where ( D T Y i Q 1 ) l and ( A Q 1 ) l represent the l-th rows of the matrices D T Y i Q 1 and A Q 1 , respectively. Equation (16) is a quadratic minimization problem, so its diagonal element δ l is solved as
δ l = ( D T Y i Q 1 ) l ( A Q 1 ) l T ( A Q 1 ) l 2 2 .
Estimating each l-th diagonal element δ l of Q 2 by Equation (17), the solution of Q 2 is obtained.
Given Q 2 and A , the subproblem of Equation (13) with respect to D is
D = arg min D Y i Q 1 D Q 2 A Q 1 F 2   s . t .   D T D = I .
Equation (18) is a Frobenius norm minimization problem with an orthogonal constraint, which can be solved by SVD of Y i Q 1 ( Q 2 A Q 1 ) T as
U Δ V T = Y i Q 1 ( Q 2 A Q 1 ) T ,
where U and V are the matrices of the left and right singular vectors, respectively, and Δ is the singular value matrix. By SVD, Equation (19) is solved as
D = U V T .
Clearly, the dictionary D is self-adaptive to each group Y i and requires only one SVD for each group, which is computationally efficient.
Given D and Q 2 , the subproblem of Equation (13) with respect to A is
A = arg min A ( Y i D Q 2 A ) Q 1 F 2 + A 1 .
As D is orthogonal, the first Frobenius norm term in Equation (21) is rewritten as D T Y i Q 1 Q 2 A Q 1 F 2 . The minimization of Equation (21) is separable into the elements of the matrices A and Y i as
a j = arg   min a j 2 σ j 2 D T y i , j Q 2 a j F 2 + a j 1 .
where a j is the j-th column vector of matrix A . The minimization of Equation (22) can be further decoupled into multiple independent scalar optimizations of the form
a j , l = arg min a j , l 1 2 ( a j , l d l T y i , j δ l ) 2 + σ j 2 4 δ l 2 | a j , l | ,
where a j , l is the l-th element of a j and d l is the l-th column of dictionary D . Equation (23) can be efficiently and accurately solved by soft thresholding, that is
a j , l = sign ( d l T y i , j δ l ) max ( | d l T y i , j δ l | σ j 2 4 δ l 2 , 0 ) ,
where sign ( · ) is a symbolic function. The soft-thresholding approach provides an explicit solution to the original Equation (21).

3.4. SAR Image Despeckling

Equations (14), (18) and (21) are solved until the stopping criterion is satisfied. The clean group X i is then estimated using the solved D , A and Q 2 as X i = D Q 2 A . Based on the estimated clean group X i , the SAR image reconstruction model is constructed as
x = arg min x x y 2 2 + η i j R i , j x ( D Q 2 A ) j 2 2 ,
where η is a regularization parameter and ( D Q 2 A ) j is the j-th column of the estimated clean group D Q 2 A . Equation (25) is a least squares problem that is minimized as follows:
x = ( I + η i j R i , j T R i , j ) 1 (   y + η i j R i , j T ( D Q 2 A ) j ) ,
where i j R i , j T R i , j is a diagonal matrix whose diagonal elements are the numbers of the corresponding pixels extracted from all similar patches. The inverse of the large-scale matrix I + η i j R i , j T R i , j is easily obtained. Solving Equation (26) yields the reconstructed SAR image.
To further improve the speckle-reduction performance of the proposed model, we repeatedly remove speckles using an iterative regularization technique that filters the noise back to the denoised image as the new input image:
y k + 1 = x k + ξ ( y x k ) ,
where k is the number of iterations and ξ is the relaxation parameter. The above despeckling process is repeated on the input image y k + 1 to reconstruct the SAR image x k + 1 . This procedure is repeated several times until the number of iterations reaches the threshold M, and the final despeckled SAR image is output from the model. Figure 2 is a flowchart of the MWSC algorithm for suppressing speckle noise in SAR images.

4. Experimental Results and Analysis

In this section, we experimentally verify the performance of the proposed MWSC approach for SAR image despeckling. The despeckling performance was compared with those of Gamma-MAP filters, the Log-KSVD method, the PPB method, the POTDF method, the FANS method, and the SAR-BM3D method. PPB and SAR-BM3D are well-known classical algorithms for processing SAR images. The executable codes of the competing methods were downloaded from the authors’ website (http://www.grip.unina.it/research/80-sar-despeckling.html) (16 October 2021).
To investigate the impacts of patch size and overlapping factor, despeckling experiments on three test images are separately conducted under various patch sizes and overlapping factors. Figure 3 shows the curves of PSNR versus patch size and overlapping factor. From Figure 3a, we have the following observations. As the patch size increases, the PSNR first increases and then drops. This is because using a large patch size allows more robust discrimination between noisy patches. However, using a too large patch will prevent the algorithm from finding enough similar patches. From Figure 3b, it is concluded that the performance of our proposed algorithm is not quite sensitive to overlapping factor since all the curves are almost flat. The highest performance for each test image is achieved with the overlapping factor in the range [2,5]. Based on the above analysis, in the proposed MWSC algorithm, we set the size of image patch as 8 × 8 and overlapping factor is empirically set as 4. Besides, the range S × S of the training window for searching similar patches is set as 30 × 30. The number of image patches in one group is set as 32 and the size of the learning dictionary is set as 64 × 64.
As noise-free signals are lacking in real SAR images, we simulated a SAR image for the experiment based on an optical image. In our simulated test, we chose the “House” and “Cameraman” images of size 256 × 256 and the “Lena” image of size 512 × 512. The despeckling performance was quantitatively evaluated by two performance measures: the peak-signal-to-noise ratio (PSNR) and the structural similarity index (SSIM), which evaluate the speckle-suppression capability of the algorithm and the structural similarity of the recovered image to the original image, respectively.
On real SAR images, the despeckling performance was evaluated by the equivalent number of looks (ENL) and the overall edge preservation index (EPI). The ENL is a widely used index that measures the degree of speckle suppression in a homogeneous region. A large ENL denotes a strong speckle-suppression ability. Meanwhile, the EPI reflects the degree of image-detail preservation. A large EPI indicates a strong detail-preservation ability.

4.1. Despeckling Results of Simulated SAR Images

We first validated the proposed method on simulated SAR images generated by adding different levels of gamma speckle noise (L = 4 or 16) to the natural images “House,” “Cameraman”, and “Lena”. The number of looks L represents the speckle noise level, where smaller L denotes a greater noise intensity. The original images are shown in Figure 4.
Figure 5, Figure 6, Figure 7 and Figure 8 show the images obtained by the different despeckling methods on the simulated optical images contaminated by speckle. All methods achieved satisfactory despeckling performance, although Gamma-MAP introduced some obvious artifacts that severely degraded the image quality of Lena (Figure 5c, Figure 6c, Figure 7c and Figure 8c). The Log-KSVD method left significant noise residues in the smoothed areas (Figure 5b and Figure 6b). Although PPB provided a much smoother result than Log-KSVD and Gamma-MAP, it blurred some important details, such as the hat texture in Figure 6d and the window edge in Figure 7d. POTDF keeps working very well on the heterogeneous region while unsatisfactory on the homogeneous region (Figure 7e). The SAR-BM3D method well preserved the texture information of the image and reasonably suppressed the speckle noise, but some residual noise appeared in the despeckled image of Lena (Figure 6f). Likewise, FANS produces a smoother output, but many wavelet-related artifacts appear. We observe that the proposed MWSC model adequately suppressed speckle noise and better restored the edges and textures of the original images than the competing methods.
To demonstrate the processing results of each algorithm on image details and point targets, we enlarged some local details in Figure 5 and Figure 7. The enlargements are shown in Figure 9 and Figure 10, respectively. Here we compared only the results of PPB, POTDF, SAR-BM3D, FANS and MWSC, the results of the Log-KSVD and Gamma-MAP algorithms were omitted because their visual effects were poor. Visible speckle noise persisted in the House image despeckled by PPB (Figure 10). The POTDF tended to produce an unfavorable oversmoothing effect. On the other hand, SAR-BM3D, FANS, and the proposed MWSC method best preserved the image details, but MWSC better suppressed the speckle noise than the SAR-BM3D and FANS. For example, significant speckle noise appeared in the enlarged local details of House obtained by SAR-BM3D and FANS (Figure 10). Besides eliminating the speckle noise, the proposed MWSC algorithm better preserved the sharper edges and finer details than the other algorithms, and thereby obtained much clearer and better visual results. The corresponding PSNR and SSIM evaluations of the different algorithms on the three synthetic images are displayed in Table 1. The MWSC algorithm achieved the highest PSNR and SSIM results in most cases (highlighted in bold font), consistent with the visual inspections.

4.2. Despeckling Results of Real SAR Images

In this experiment, the despeckling performances of the proposed and existing algorithms were tested on real SAR images (see Figure 11). The despeckled results of the real SAR images obtained by the compared methods are shown in Figure 12, Figure 13 and Figure 14.
As shown in Figure 12, Figure 13 and Figure 14, the proposed MWSC outperformed its competitors on the real SAR images. The Gamma-MAP algorithm eliminated less speckle noise than other methods and introduced serious aliasing artifacts. It also blurred edges of the details, as shown in Figure 12c and Figure 13c.The Log-KSVD algorithm slightly improved the despeckling effect but retained some obvious speckle noise (see Figure 12b and Figure 13b). Although PPB effectively suppressed the speckle noise from the SAR images, it blurred some edges and strong targets Figure 12d and Figure 14d). From Figure 12e and Figure 13e, we can find that POTDF still remains much noise in both homogeneous and heterogeneous areas. SAR-BM3D and FANS preserve all structures very well, with an accuracy comparable to that of MWSC. The MWSC, however, succeed in removing speckle in both homogeneous and heterogeneous areas, providing a sharper result and contributing to a better perceived quality. To more intuitively compare the detailed processing of the various algorithms, Figure 15 and Figure 16 present local enlargements of the despeckled images obtained by the despeckling methods. Again, the proposed MWSC method achieved a higher visual quality than the other despeckling methods.
To quantitatively evaluate the despeckling results, the corresponding ENLs and EPIs of the three real images are provided in Table 2, with the best values highlighted in bold. The ENL values were calculated in homogeneous regions of the real SAR images. Regions 1 and 2 in Table 2 correspond to the red and yellow regions, respectively, in the SAR images (such as the Figure 12a). PPB and the proposed MWSC achieved higher ENL than the other despeckling methods, affirming that the PPB and MWSC algorithms have stronger speckle-reduction ability in the homogeneous regions than the remaining algorithms. This conclusion is consistent with the above visual inspection. Meanwhile, SAR-BM3D, FANS, and MWSC obtained higher EPI than PPB, Gamma-MAP, Log-KSVD and POTDF, confirming the superior performance of SAR-BM3D, FANS and MWSC in image-detail retention. In short, the proposed MWSC provided the best compromise between speckle reduction and detail preservation.

4.3. Impacts of Weights and Dictionaries on Performance

To better understand the roles of the weights Q 1 and Q 2 and the learning dictionary D in the MWSC model, we compared the performances of the proposed MWSC and several MWSC-based benchmark methods. We first replaced the weight Q 1 with the unit diagonal matrix, forming the QWSC benchmark model. Replacing the learning dictionary D with the wavelet dictionary, we then formed the DWSC benchmark model.
Figure 17 and Figure 18 present the despeckled images obtained by the different despeckling methods. Log-KSVD retained a large amount of speckle noise in the “House” image (Figure 17a). QWSC improved the visual effect of “House” because the weights Q 2 improve the sparse representation capability of the QWSC algorithm and further enhance the speckle-suppression performance (Figure 17b). MWSC achieved a higher visual effect than QWSC because the weights Q 1 effectively capture and utilize the statistical properties of the noise (Figure 18b,d). Although the result of DWSC exhibited no obvious speckle noise, a small amount of detailed information was lost (Figure 18c). Obviously, the learning dictionary in the proposed model adapted to different structures of the image to improve the speckle-reduction performance of the model.
Table 3 and Table 4 give the numerical evaluation results of the different algorithms on the “House” and real SAR2 images, respectively, with the best results marked in bold. As shown in Table 3, MWSC achieved higher PSNR and SSIM values than the benchmark algorithms on “House,” confirming that weight Q 1 , weight Q 2 , and dictionary learning all improved the PSNR and SSIM values of the recovered images to different degrees. Meanwhile, the ENL and EPI values in Table 4 clarify that the proposed BWCS algorithm well balanced the speckle noise suppression with detail preservation in the real SAR image, outperforming the Log-KSVD algorithm and the benchmark algorithms.
This comprehensive analysis demonstrates that introducing weight Q 1 improves the noise-statistical properties of SAR images, while introducing weight Q 2 captures the sparse a priori properties of SAR images. The learning dictionary is more adaptable to different image structures than the fixed dictionary, thereby enhancing the signal sparsity and achieving competitive SAR image-despeckling performance.

5. Conclusions

This paper proposed a multiweighted SAR image despeckling method based on the traditional sparse representation model and combined it into a framework that unifies dictionary learning and coefficient weighting. In this method, the structure group is constructed using the nonlocal similarity between SAR image patches and is employed as the processing object of speckle reduction in SAR images. This approach significantly improves the sparsity of the image coefficients. Borrowing the idea of coefficient weighting, we also designed a data-fidelity term for the total downscaling model and a regularization term to enhance the reliability and accuracy of image coefficient estimation. An orthogonal dictionary was trained on sets of similar image patches to elevate the sparsity of the coefficients and simplify the complexity of sparse coding. Throughout the iterations, the algorithm updates the weight matrix of each similar image patch set to improve the full-medium adaptation. By combining the statistical parameter estimation of SAR images with iterative regularization, we further improved the quality of the images during the solution process. In comparison experiments, the proposed method achieved comparable or better despeckling performance than four competing methods, in terms of speckle suppression, detail preservation, and visual effect. The results verified the reasonableness and effectiveness of the proposed method.

Author Contributions

Conceptualization, S.L.; Data curation, N.P.; Formal analysis, J.C.; Resources, K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research is partly supported by the National Natural Science Foundation of China (Grant Nos. 61701055, 42074034).

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Moreira, A.; Prats-Iraola, P.; Younis, M.; Krieger, G.; Hajnsek, I.; Papathanassiou, K.P. A tutorial on synthetic aperture radar. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–43. [Google Scholar] [CrossRef] [Green Version]
  2. Adamo, M.; De Carolis, G.; De Pasquale, V.; Pasquariello, G. Oilspill surveillance and tracking with combined use of SAR and modisimagery: A case study. In Proceedings of the 2006 IEEE International Symposium on Geoscience and Remote Sensing, Denver, CO, USA, 31 July–4 August 2006; pp. 1327–1330. [Google Scholar]
  3. Zhang, J.X.; Yang, S.C.; Zhao, Z.; Huang, G. SAR mapping technology and its application in difficulty terrain area. In Proceedings of the 2010 IEEE International Geoscience and Remote Sensing Symposium, Honolulu, HI, USA, 25–30 July 2010; pp. 3608–3611. [Google Scholar]
  4. Huang, S.Q.; Liu, D.Z.; Gao, G.Q.; Guo, X.-J. A novel method for speckle noise reduction and ship target detection in SAR images. Pattern Recognit 2019, 42, 1533–1542. [Google Scholar] [CrossRef]
  5. Khwaja, A.S.; Zhang, X.P. Motion parameter estimation and focusing from SAR images based on sparse reconstruction. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1350–1354. [Google Scholar] [CrossRef]
  6. Ma, X.; Shen, H.; Yang, J.; Zhang, L.; Li, P. Polarimetric-Spatial Classification of SAR Images Based on the Fusion of Multiple Classifiers. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 961–971. [Google Scholar] [CrossRef]
  7. Ma, X.; Shen, H.; Zhao, X.; Zhang, L. SAR Image Despeckling by the Use of Variational Methods With Adaptive Nonlocal Functionals. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3421–3435. [Google Scholar] [CrossRef]
  8. Huang, Y.M.; Lu, D.Y.; Zeng, T. Two-step approach for the restoration of images corrupted by multiplicative noise. SIAM J. Sci. Comput. 2013, 35, 2856–2873. [Google Scholar] [CrossRef]
  9. Huang, Y.-M.; Ng, M.; Zeng, T. The Convex Relaxation Method on Deconvolution Model with Multiplicative Noise. Commun. Comput. Phys. 2013, 13, 1066–1092. [Google Scholar] [CrossRef] [Green Version]
  10. Li, Z.; Lou, Y.; Zeng, T. Variational Multiplicative Noise Removal by DC Programming. J. Sci. Comput. 2016, 68, 1200–1216. [Google Scholar] [CrossRef]
  11. Lu, J.; Shen, L.; Xu, C.; Xu, Y. Multiplicative noise removal with a sparsity-aware optimization model. Inverse Probl. Imaging 2017, 11, 949–974. [Google Scholar] [CrossRef] [Green Version]
  12. Feng, H.X.; Hou, B.; Gong, M.G. SAR image despeckling based on local homogeneous-region segmentation by using pixel-relativity measurement. IEEE Trans. Geosci. Remote Sens. 2011, 49, 2724–2737. [Google Scholar] [CrossRef]
  13. Lopes, A.; Nezry, E.; Touzi, R.; Laur, H. Maximum A Posteriori Speckle Filtering And First Order Texture Models In Sar Images. In Proceedings of the 10th Annual International Symposium on Geoscience and Remote Sensing, College Park, MD, USA, 20–24 May 1990. [Google Scholar]
  14. Lee, J.S. Digital image enhancement and noise filtering by use of local statistics. IEEE Trans. Pattern Anal. Mach. Intell. 1980, PAMI-2, 165–168. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Frost, V.S.; Stiles, J.A.; Shanmugan, K.S.; Holtzman, J.C. A model for radar images and its application to adaptive digital filtering of multiplicative noise. IEEE Trans. Pattern Anal. Mach. Intell. 1982, 4, 157–166. [Google Scholar] [CrossRef] [PubMed]
  16. Kuan, D.T.; Sawchuk, A.A.; Strand, T.C.; Chavel, P. Adaptive noise smoothing filter for images with signal-dependent noise. IEEE Trans. Pattern Anal. Mach. Intell. 1985, 7, 165–177. [Google Scholar] [CrossRef] [PubMed]
  17. Lopes, A.; Touzi, R.; Nezry, E. Adaptive speckle filters and scene heterogeneity. IEEE Trans. Geosci. Remote Sens. 1990, 28, 992–1000. [Google Scholar] [CrossRef]
  18. Ozcan, C.; Sen, B.; Nar, F. Sparsity-Driven Despeckling for SAR Images. IEEE Geosci. Remote Sens. Lett. 2015, 13, 115–119. [Google Scholar] [CrossRef]
  19. Solbø, S.; Eltoft, T. Homomorphic wavelet-based statistical despeckling of SAR images. IEEE Trans. Geosci. Remote Sens. 2004, 42, 711–721. [Google Scholar] [CrossRef]
  20. Buades, A.; Coll, B.; Morel, J.-M. A non-local algorithm for image denoising. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 60–65. [Google Scholar] [CrossRef]
  21. Deledalle, C.A.; Denis, L.; Tupin, F. Iterative weighted maximum likelihood denoising with probabilistic patch-based weights. IEEE Trans. Image Process. 2009, 18, 2661–2672. [Google Scholar] [CrossRef] [Green Version]
  22. Zhang, Z.; Xuelong, L.; Yang, J.; Li, X.; Zhang, D. A Survey of Sparse Representation: Algorithms and Applications. IEEE Access 2015, 3, 490–530. [Google Scholar] [CrossRef]
  23. Lu, X.; Li, X. Group sparse reconstruction for image segmentation. Neurocomputing 2014, 136, 41–48. [Google Scholar] [CrossRef]
  24. Foucher, S. SAR Image Filtering Via Learned Dictionaries and Sparse Representations. In Proceedings of the IGARSS 2008—2008 IEEE International Geoscience and Remote Sensing Symposium, Boston, MA, USA, 7–11 July 2008. [Google Scholar]
  25. Jiang, J.; Jiang, L.; Sang, N. Non-local sparse models for SAR image despeckling. In Proceedings of the 2012 International Conference on Computer Vision in Remote Sensing, Xiamen, China, 16–18 December 2012; pp. 230–236. [Google Scholar]
  26. Chen, C.; Huang, J. The benefit of tree sparsity in accelerated MRI. Med. Image Anal. 2014, 18, 834–842. [Google Scholar] [CrossRef] [PubMed]
  27. Xu, B.; Cui, Y.; Li, Z.; Zuo, B.; Yang, J.; Song, J. Patch Ordering-Based SAR Image Despeckling Via Transform-Domain Filtering. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 8, 1682–1695. [Google Scholar] [CrossRef]
  28. Yang, J.; Zhang, Y.; Yin, W. A Fast Alternating Direction Method for TVL1-L2 Signal Reconstruction From Partial Fourier Data. IEEE J. Sel. Top. Signal Process. 2010, 4, 288–297. [Google Scholar] [CrossRef]
  29. Elad, M.; Aharon, M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans. Image Process. 2006, 15, 3736–3745. [Google Scholar] [CrossRef] [PubMed]
  30. Zhan, Z.; Cai, J.F.; Guo, D.; Liui, Y.; Chen, Z.; Qu, X. Fast multi-class dictionaries learning with geometrical directions in MRI reconstruction. IEEE Trans. Biomed. Eng. 2016, 63, 1850–1861. [Google Scholar] [CrossRef]
  31. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  32. Parrilli, S.; Poderico, M.; Angelino, C.V.; Verdoliva, L. A Nonlocal SAR Image Denoising Algorithm Based on LLMMSE Wavelet Shrinkage. IEEE Trans. Geosci. Remote Sens. 2012, 50, 606–616. [Google Scholar] [CrossRef]
  33. Cozzolino, D.; Parrilli, S.; Scarpa, G.; Poggi, G.; Verdoliva, L. Fast Adaptive Nonlocal SAR Despeckling. IEEE Geosci. Remote Sens. Lett. 2014, 11, 524–528. [Google Scholar] [CrossRef] [Green Version]
  34. Dong, W.; Zhang, L.; Shi, G.; Li, X. Nonlocally centralized sparse representation for image restoration. IEEE Trans. Image Process. 2013, 22, 1620–1630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Dong, W.; Shi, G.; Li, X. Nonlocal Image Restoration With Bilateral Variance Estimation: A Low-Rank Approach. IEEE Trans. Image Process. 2013, 22, 700–711. [Google Scholar] [CrossRef]
  36. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Weighted Nuclear Norm Minimization with Application to Image Denoising. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  37. Xie, Y.; Gu, S.; Liu, Y.; Zuo, W.; Zhang, W.; Zhang, L. Weighted Schatten pp -Norm Minimization for Image Denoising and Background Subtraction. IEEE Trans. Image Process. 2016, 25, 4842–4857. [Google Scholar] [CrossRef] [Green Version]
  38. Xu, J.; Zhang, L.; Zhang, D. A Trilateral Weighted Sparse Coding Scheme for Real-World Image Denoising. arXiv 2018, arXiv:1807.04364. [Google Scholar]
  39. Wang, F.; Zhao, X.L.; Ng, M.K. Multiplicative noise and blur removal by framelet decomposition and l1-based L-curve method. IEEE Trans. Image Process. 2016, 25, 4222–4232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Comparison of similarity levels of two groups found in the “Lena” image. (a) Group A (in the red window) and Group B (in the cyan window) include exemplary and similar patches; (b) Zoomed-in exemplary and similar patches in groups “A” and “B”.
Figure 1. Comparison of similarity levels of two groups found in the “Lena” image. (a) Group A (in the red window) and Group B (in the cyan window) include exemplary and similar patches; (b) Zoomed-in exemplary and similar patches in groups “A” and “B”.
Entropy 24 00096 g001
Figure 2. Flowchart of the MWSC algorithm for SAR image despeckling.
Figure 2. Flowchart of the MWSC algorithm for SAR image despeckling.
Entropy 24 00096 g002
Figure 3. Performance comparison with various patch sizes (a) and overlapping factors (b) for three test images.
Figure 3. Performance comparison with various patch sizes (a) and overlapping factors (b) for three test images.
Entropy 24 00096 g003
Figure 4. Original images used in the comparison experiments. (a) Lena (512 × 512); (b) House (256 × 256); (c) Cameraman (256 × 256).
Figure 4. Original images used in the comparison experiments. (a) Lena (512 × 512); (b) House (256 × 256); (c) Cameraman (256 × 256).
Entropy 24 00096 g004
Figure 5. Despeckling results of Lena corrupted by 16-look speckle. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Figure 5. Despeckling results of Lena corrupted by 16-look speckle. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Entropy 24 00096 g005
Figure 6. Despeckling results of Lena corrupted by 4-look speckle. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Figure 6. Despeckling results of Lena corrupted by 4-look speckle. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Entropy 24 00096 g006
Figure 7. Despeckling results of House corrupted by 4-look speckle. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Figure 7. Despeckling results of House corrupted by 4-look speckle. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Entropy 24 00096 g007
Figure 8. Despeckling results of cameraman corrupted by 16-look speckle. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Figure 8. Despeckling results of cameraman corrupted by 16-look speckle. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Entropy 24 00096 g008aEntropy 24 00096 g008b
Figure 9. Enlarged details of the despeckling results on the Lena image with L = 16 (left to right: PPB algorithm, POTDF algorithm, SAR-BM3D algorithm, FANS algorithm, MWSC algorithm).
Figure 9. Enlarged details of the despeckling results on the Lena image with L = 16 (left to right: PPB algorithm, POTDF algorithm, SAR-BM3D algorithm, FANS algorithm, MWSC algorithm).
Entropy 24 00096 g009
Figure 10. Enlarged details of the despeckling results on the House image with L = 4 (left to right: PPB algorithm, POTDF algorithm, SAR-BM3D algorithm, FANS algorithm, MWSC algorithm).
Figure 10. Enlarged details of the despeckling results on the House image with L = 4 (left to right: PPB algorithm, POTDF algorithm, SAR-BM3D algorithm, FANS algorithm, MWSC algorithm).
Entropy 24 00096 g010
Figure 11. Real SAR images used in the comparison experiments. (a) SAR1 (256 × 256); (b) SRA2 (256 × 256); (c) SAR3 (512 × 512).
Figure 11. Real SAR images used in the comparison experiments. (a) SAR1 (256 × 256); (b) SRA2 (256 × 256); (c) SAR3 (512 × 512).
Entropy 24 00096 g011
Figure 12. Despeckling results of the real SAR1 image. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Figure 12. Despeckling results of the real SAR1 image. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Entropy 24 00096 g012
Figure 13. Despeckling results of the real SAR2 image. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Figure 13. Despeckling results of the real SAR2 image. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Entropy 24 00096 g013
Figure 14. Despeckling results of the real SAR3 image. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Figure 14. Despeckling results of the real SAR3 image. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS; (h) MWSC.
Entropy 24 00096 g014aEntropy 24 00096 g014b
Figure 15. Despeckling results of a local enlargement of Figure 9. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS.
Figure 15. Despeckling results of a local enlargement of Figure 9. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS.
Entropy 24 00096 g015
Figure 16. Depeckling results of a local enlargement of Figure 10. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS.
Figure 16. Depeckling results of a local enlargement of Figure 10. (a) Noisy image; (b) Log-KSVD; (c) Gamma-MAP; (d) PPB; (e) POTDF; (f) SAR-BM3D; (g) FANS.
Entropy 24 00096 g016
Figure 17. Despeckling results of House corrupted by 4-look speckle. (a) Log-KSVD; (b) QWSC; (c) DWSC; (d) MWSC.
Figure 17. Despeckling results of House corrupted by 4-look speckle. (a) Log-KSVD; (b) QWSC; (c) DWSC; (d) MWSC.
Entropy 24 00096 g017
Figure 18. Despeckling results of real SAR2 image. (a) Log-KSVD; (b) QWSC; (c) DWSC; (d) MWSC.
Figure 18. Despeckling results of real SAR2 image. (a) Log-KSVD; (b) QWSC; (c) DWSC; (d) MWSC.
Entropy 24 00096 g018
Table 1. PSNR and SSIM values of the tested algorithms in despeckling of simulated SAR images.
Table 1. PSNR and SSIM values of the tested algorithms in despeckling of simulated SAR images.
ImageAlgorithmL = 4L = 16
PSNRSSIMPSNRSSIM
LenaNoise17.79870.264323.76170.4713
Log-KSVD23.05800.440629.72230.7523
Gamma-MAP25.97960.722629.75650.8146
PPB
POTDF
29.8550
29.6280
0.8025
0.8371
33.2598
34.0313
0.8786
0.8929
SAR-BM3D
FANS
31.1636
31.2764
0.8452
0.8514
34.1672
34.3950
0.8984
0.8981
MWSC31.49000.856434.12830.8949
HouseNoise17.01680.228722.99880.4362
Log-KSVD23.78370.524228.14200.7163
Gamma-MAP24.38810.673328.07290.7750
PPB
POTDF
29.6706
28.8839
0.7874
0.8244
33.2822
33.8582
0.8632
0.8746
SAR-BM3D
FANS
31.2560
31.1984
0.8355
0.8415
34.4218
34.3353
0.8886
0.8782
MWSC31.55660.846934.50000.8914
CameramanNoise17.73530.409523.73190.5629
Log-KSVD24.02670.624128.95150.7891
Gamma-MAP24.01100.723028.22640.8034
PPB
POTDF
26.9156
27.6073
0.7840
0.8262
29.1476
31.6093
0.8582
0.9023
SAR-BM3D
FANS
28.0520
28.0336
0.8359
0.8384
31.5177
31.7052
0.9086
0.9061
MWSC28.13340.842431.71150.9094
Table 2. ENL and EPI values of the tested algorithms in despeckling of real SAR images.
Table 2. ENL and EPI values of the tested algorithms in despeckling of real SAR images.
ImageAlgorithmENLEPI
Region 1Region 2
SAR1Noise9.212110.27181
Log-KSVD83.6841106.08720.3068
Gamma-MAP127.8704188.94520.1149
PPB
POTDF
246.0215
78.3594
358.4988
92.3279
0.1385
0.3227
SAR-BM3D
FANS
64.1340
102.4533
86.2054
178.5487
0.3520
0.3444
MWSC173.1734219.84000.3516
SAR2Noise14.037411.69721
Log-KSVD252.7677151.58270.4847
Gamma-MAP295.5257226.28690.2924
PPB
POTDF
1366.9521
436.5643
2939.8065
549.5423
0.1880
0.3233
SAR-BM3D
FANS
1470.8629
554.9834
727.2407
697.4763
0.3109
0.3151
MWSC1006.4874980.59040.3134
SAR3Noise17.415020.36981
Log-KSVD128.5206184.86840.2989
Gamma-MAP149.3338222.55500.2155
PPB
POTDF
488.0467
144.8686
809.7066
196.7193
0.2472
0.2902
SAR-BM3D
FANS
119.7533
156.1875
159.0166
269.8643
0.3991
0.3538
MWSC186.0285306.69400.3787
Table 3. PSNR and SSIM values of the tested algorithms in despeckling of House Image.
Table 3. PSNR and SSIM values of the tested algorithms in despeckling of House Image.
ImageAlgorithmL = 4L = 16
PSNRSSIMPSNRSSIM
HouseNoise17.01680.228722.99880.4362
Log-KSVD23.78370.524228.14200.7163
QWSC27.35450.828432.26590.8662
DWSC27.53530.717232.30590.8597
MWSC31.55660.846934.50000.8914
Table 4. ENL and EPI values of the tested algorithms in despeckling of real SAR2 Image.
Table 4. ENL and EPI values of the tested algorithms in despeckling of real SAR2 Image.
ImageAlgorithmENLEPI
Region 1Region 2
SAR2Noise14.037411.69721
Log-KSVD252.7677151.58270.4847
QWSC591.7327711.97110.2075
DWSC1527.79781029.47980.2541
MWSC1006.4874980.59040.3134
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, S.; Pu, N.; Cao, J.; Zhang, K. Synthetic Aperture Radar Image Despeckling Based on Multi-Weighted Sparse Coding. Entropy 2022, 24, 96. https://doi.org/10.3390/e24010096

AMA Style

Liu S, Pu N, Cao J, Zhang K. Synthetic Aperture Radar Image Despeckling Based on Multi-Weighted Sparse Coding. Entropy. 2022; 24(1):96. https://doi.org/10.3390/e24010096

Chicago/Turabian Style

Liu, Shujun, Ningjie Pu, Jianxin Cao, and Kui Zhang. 2022. "Synthetic Aperture Radar Image Despeckling Based on Multi-Weighted Sparse Coding" Entropy 24, no. 1: 96. https://doi.org/10.3390/e24010096

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop