Next Article in Journal
Mapping and Classification of Ecologically Sensitive Marine Habitats Using Unmanned Aerial Vehicle (UAV) Imagery and Object-Based Image Analysis (OBIA)
Previous Article in Journal
Effects of Forest Canopy Vertical Stratification on the Estimation of Gross Primary Production by Remote Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Noise Removal Based on Tensor Modelling for Hyperspectral Image Classification

Centrale Marseille, Institut Fresnel, Aix Marseille Université, CNRS, 13013 Marseille, France
*
Author to whom correspondence should be addressed.
Remote Sens. 2018, 10(9), 1330; https://doi.org/10.3390/rs10091330
Submission received: 27 June 2018 / Revised: 10 August 2018 / Accepted: 14 August 2018 / Published: 21 August 2018
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
With the current state-of-the-art computer aided manufacturing tools, the spatial resolution of hyperspectral sensors is becoming increasingly higher thus making it easy to obtain much more detailed information of the scene captured. However, the improvement of the spatial resolution also brings new challenging problems to address with signal dependent photon noise being one of them. Unlike the signal independent thermal noise, the variance of photon noise is dependent on the signal, therefore many denoising methods developed for the stationary noise cannot be applied directly to the photon noise. To make things worse, both photon and thermal noise coexist in the captured hyperspectral image (HSI), thus making it more difficult to whiten noise. In this paper, we propose a new denoising framework to cope with signal dependent nonwhite noise (SDNW), Pre-estimate—Whitening—Post-estimate (PWP) loop, to reduce both photon and thermal noise in HSI. Previously, we proposed a method based on multidimensional wavelet packet transform and multi-way Wiener filter which performs both white noise and spectral dimensionality reduction, referred to as MWPT-MWF, which was restricted to white noise. We get inspired from this MWPT-MWF to develop a new iterative method for reducing photon and thermal noise. Firstly, the hyperspectral noise parameters estimation (HYNPE) algorithm is used to estimate the noise parameters, the SD noise is converted to an additive white Gaussian noise by pre-whitening procedure and then the whitened HSI is denoised by the proposed method SDNW-MWPT-MWF. As comparative experiments, the Multiple Linear Regression (MLR) based denoising method and tensor-based Multiway Wiener Filter (MWF) are also used in the denoising framework. An HSI captured by Reflective Optics System Imaging Spectrometer (ROSIS) is used in the experiments and the denoising performances are assessed from various aspects: the noise whitening performance, the Signal-to-Noise Ratio (SNR), and the classification performance. The results on the real-world airborne hyperspectral image HYDICE (Hyperspectral Digital Imagery Collection Experiment) are also presented and analyzed. These experiments show that it is worth taking into account noise signal-dependency hypothesis for processing HYDICE and ROSIS HSIs.

1. Introduction

Hyperspectral images consist of a considerably large number of narrow spectral bands which are uniformly distributed over a wide spectral range. For each pixel, the almost continuous spectral signature adds the third orthogonal spectral dimension to the two-dimensional spatial domain and constitutes the well-known 3D data cube. Thus, hyperspectral signatures offer the capability to identify and discriminate ground cover types. To benefit from the additional spectral dimension, specific processing methods of hyperspectral images, for instance spectral signatures unmixing, target detection and classification, etc., have been developed. For these methods to operate well, the reduction of the noise affecting hyperspectral images (HSIs) should be incorporated into them.
The noise in HSIs can be distinguished into two classes [1]: random noise and fixed pattern noise. Photon and thermal noise are examples of random noise in HSIs, while striping, periodic and interference noise are examples of fixed pattern noise, which is generated by errors in the calibration process and can be removed from the HSI by suitable procedures [2]. However, random noise, due to its stochastic nature, cannot be removed by those procedures and influences the performance of the algorithms adopted in hyperspectral data exploitation. Note that, in this paper, we only focus on the random noise and will later refer to it as noise. For sensors used in hyperspectral imagery, the theory predicts that random noise mainly comes from two aspects: signal-independent (SI) electronic noise and signal-dependent (SD) photon noise [3]. The widely accepted SI noise model is white Gaussian one [1,3]. In [4,5], it was shown that the SI noise in some HSIs is colored, i.e., spectrally non-white. With the improvement of the sensitivity in the electronic components [6], the resolution of the charged coupled device (CCD) camera has improved significantly, so that the photon noise has become as dominant as the signal-independent electronic noise in HSI data collected by new-generation hyperspectral sensors [7,8,9,10]. In this case, the assumption of additive and stationary noise model is not appropriate although this hypothesis is plausible for HSIs where the SI noise is dominant while SD noise, which depends on the useful signal level, is negligible. Therefore, in this paper, we use the widely accepted noise model in [7,8,9,10] including both signal-dependent and signal-independent noise. There are few denoising algorithms based on the photon noise model. The tensor-based denoising methods were proposed only for the white noise situation. According to the different statistical properties of SI and SD noise, in this paper, we propose a method to remove the SD photon noise as well as the SI thermal noise in the HSI.
In the literature, two widely used models for the random noise in HSIs are the additive white noise along both spectral and spatial dimensions [8,11,12,13,14], and the Additive White Gaussian Noise (AWGN) along the spatial dimensions but non-stationary in the spectral dimension [15,16]; these are pretty reasonable processes when the thermal noise is dominant [8]. However, when the SD photon noise is taken into account, these two noise models become appropriate. The SD photon noise in digital images was discussed in [17,18], where the noise parameters were estimated by utilizing a scatter plot-based estimation procedure. Nonetheless, the discussion of these two papers was only limited to the pure SD noise, which is not suitable for the noise in HSIs. A more accurate generalized SD noise model for digital images was introduced in [19], where the noise parameters were estimated by the Linear Minimum Mean Square Error (LMMSE) in the wavelet domain. In addition, the same noise model was employed in [20] under the name Poissonian-Gaussian noise model, where the noise in HSIs was modeled as two parts: the Poissonian part for modeling the photon noise and the Gaussian part for modeling the remaining stationary distribution in the output data. Images were firstly transformed into the wavelet domain; then, the expectation/standard-deviation pairs were estimated by employing a local estimation method. Finally, the Maximum Likelihood (ML) method was applied to the locally estimated expectation/standard-deviation pairs to estimate the global parameters. Based on the generalized SD noise model, a scatter plot-based estimation method was proposed in [7] to estimate the parameters of the noise generated by the new-generation imaging spectrometers.
The generalized SD noise model was then introduced in the HSI framework in [3]. In this paper, the 2D fractal Brownian motion (fBm) model was employed as a statistical model for intra-band image texture correlation, which has made it possible to estimate additive noise variance locally both from homogeneous and textural images Scanning Widow (SW). Then, the noise parameters were estimated by a linear fit in each band. The spectral textural correlation was also considered, and each SW in a Multicomponent Scanning Window (MSW) was treated as a mixture of fBm-samples and noise. Finally, the global noise parameters were estimated by ML estimation.
The generalized SD noise model was also used in [8], where the HYperspectral Noise Parameter Estimation (HYNPE) algorithm was proposed. Unlike [3], HYNPE utilized the assumption that the noise in each band after being whitened followed a standard Gaussian distribution. The joint probability density function (PDF) of the whitened noise values was used to form the ML criterion. Nonetheless, the signal and noise values were assumed to be known in the ML criterion, which was not true in practical situations. Hence, the MLR-theory based approach was exploited to estimate them. However, MLR calculates the parameters by minimizing the Least Square Error (LSE), which is a biased estimator when the noise is not white. Thus, the estimates of MLR are not accurate in HYNPE, leading to the inaccuracy of the final parameter estimates. In this paper, we investigate the relevance of a new denoising framework for reducing simultaneously the SD photon noise and the SI thermal noise in HSIs on classification [21]. Considering that the noise variance is entangled with the signal, a denoising loop is proposed to remove the noise from HSIs. Each iteration of the loop consists three steps: firstly, we do the pre-estimate, i.e., use MWPT-MWF to directly denoise the HSI containing the photon and thermal noise. Secondly, use the pre-estimate to estimate the noise parameters by employing the ML criterion, and then whiten the noise. Finally, apply MWPT-MWF to the whitened HSIs to obtain the post-estimate. Then, in the next iteration, use the post-estimate of the last iteration as the pre-estimate of current iteration. After several iterations, the HSIs will be well denoised and the resulting classification results are improved.
The remainder of this paper is organized as follows: Section 2 gives the signal model used in this paper. Appendix A presents the Multilinear Algebra Tools. Appendix B overviews the HYNPE algorithm. Appendix C presents the MWPT-MWF method. Section 3 presents the proposed iterative denoising method: considering that the noise variance is dependent on the signal, a denoising loop is proposed to remove the noise from HSIs. Each iteration of the loop consists of three steps: (i) we perform a pre-estimation, i.e., use MWPT-MWF [22,23] to directly denoise the HSI containing the photon and thermal noise, (ii) we use the pre-estimation process to estimate the noise parameters by employing the ML criterion, and then whiten the noise and (iii) we apply MWPT-MWF to the whitened HSIs to obtain the post-estimate results. Then in the next iteration, we use the post-estimate of the last iteration as the pre-estimate of current iteration. After several iterations, the HSIs will be well denoised. Section 4 gives some comparative experimental results. The real-world HSI reflective optics system imaging spectrometer (ROSIS) is used in the experiments to evaluate the performances of denoising and classification processes. Finally, Section 6 provides details of the conclusions of the research undertaken.
In this paper, we denote by
x R a scalar
x R I 1 a vector
X R I 1 × I 2 a matrix
X R I 1 × I 2 × I N a N-order tensor
X n n-mode unfolding matrix
X X T a tensor obtained by calculating the square root of each element of X
I n n-mode dimension
X the Frobenius norm of X
the vector outer product
the Khatri-Rao product
the Hadamard product
< X , Y > the inner product between X and Y
E [ · ] the mathematical expectation
( . ) T the transposition.

2. Signal Modeling with Thermal and Photon Noise

An HSI is a three-dimensional data cube and can be modeled as a tensor R R I 1 × I 2 × I 3 , with the first two dimensions being the spatial domain and with the third dimension being the spectral domain. In fact, a hyperspectral sensor captures a HSI as a series of two-dimensional images of the spatial domain by a sensor array. By extending the data model in [7] to the 3D representation, a noisy HSI can be expressed as a third order tensor R R I 1 × I 2 × I 3 composed of a multidimensional signal X R I 1 × I 2 × I 3 impaired by an additive random noise N R I 1 × I 2 × I 3 :
R = X + N ,
where N accounts for both thermal and photon noise and its variance depends on the pixel x i 1 , i 2 , i 3 in the useful signal X . The photon noise is caused by the random fluctuation of photon flux arriving at the CCD sensor, and it follows a Poisson model [24]. As the pixel size becomes smaller in the new-generation hyperspectral sensor, the number of photons that reach a pixel per unit time becomes smaller as well. Hence, the photon noise cannot be neglected anymore [20]. For a given entry x i 1 , i 2 , i 3 of the pure signal HSI tensor X R I 1 × I 2 × I 3 , the corresponding photon noise element p i 1 i 2 i 3 of tensor P R I 1 × I 2 × I 3 can be expressed as [25]:
p i 1 i 2 i 3 = x i 1 i 2 i 3 u i 1 i 2 i 3 ,
where u i 1 , i 2 , i 3 is a stationary, zero-mean uncorrelated random process independent on x i 1 , i 2 , i 3 with variance σ u , i 3 2 . The thermal noise component in each sensor is electronics noise, denoted by t i 1 , i 2 , i 3 which can be modeled as an additive zero-mean white Gaussian noise in each band with variance σ t , i 3 2 , while the noise variance changes from sensor to sensor due to different states of the electronic components in the sensors. Elementwise, the data model is [7]:
r i 1 , i 2 , i 3 = x i 1 , i 2 , i 3 + x i 1 , i 2 , i 3 · u i 1 , i 2 , i 3 + t i 1 , i 2 , i 3 .
Then, we can define N = P + T , and Equation (1) can be correspondingly rewritten as
R = X + P + T .
The unfolding matrix R 3 R I 3 × M 3 of the HSI data tensor R R I 1 × I 2 × I 3 (with M 3 = I 1 I 2 ) can be expressed as:
R 3 = X 3 + N 3 ,
where X 3 is the mode-3 unfolding matrix of the multidimensional signal tensor X and
N 3 = P 3 + T 3 ,
with P 3 and T 3 being the mode-3 unfolding matrices of P and T , respectively.

3. Proposed Method

The aim of this paper is to obtain the pure signal estimate X ^ , which is necessary to determine the noise variance. Nonetheless, since the noise is signal-dependent, the noise variances of the entries of R are different from each other and are related to the signal entries x i 1 i 2 i 3 . It is worth noting that, for a given entry x i 1 i 2 i 3 , the noise n i 1 i 2 i 3 is a summation of two Gaussian-distributed variables u i 1 i 2 i 3 and t i 1 i 2 i 3 . Hence, n i 1 i 2 i 3 = x i 1 i 2 i 3 u i 1 i 2 i 3 + t i 1 i 2 i 3 is a conditional zero-mean Gaussian-distributed random variable, i.e.,
n i 1 i 2 i 3 N 0 , σ n i 1 i 2 i 3 2 ,
where N · denotes the normal distribution, and σ n i 1 i 2 i 3 2 is the noise variance, which can be expressed as [3,7,8]:
σ n i 1 i 2 i 3 2 = E [ ( x i 1 i 2 i 3 u i 1 i 2 i 3 + t i 1 i 2 i 3 ) 2 | x i 1 i 2 i 3 ] = x i 1 i 2 i 3 σ u , i 3 2 + σ t , i 3 2 .
This is to say that a precise noise variance estimate σ ^ i 1 i 2 i 3 2 needs the precise signal estimate x ^ i 1 i 2 i 3 . Hence, the signal estimate and noise variance estimate problems are inter-related, thus making the signal estimate problem difficult to solve. In HYNPE, the noise parameters are estimated by using the signal estimate generated by the MLR theory based method, which estimates the signal by minimizing LSE. However, the LSE estimator requires that the signal and noise should be statistically independent, which is not satisfied in the SD photon noise situation. Thus, the estimates of the signal and noise are not precise, which makes the parameter estimate result unreliable as well. Moreover, since there is only one step for estimating the signal, the imprecise estimate degrades the performances of noise parameter estimation.
To use the classical parameter estimation algorithms, such as LSE and LMMSE, it is necessary to make the noise “independent of” the signal.
From Equation (8), it is evident that the noise variance σ n i 1 i 2 i 3 2 is dependent on signal x i 1 i 2 i 3 . To cut off this relation, we need to whiten the noise:
n ̲ i 1 i 2 i 3 = n i 1 i 2 i 3 σ n i 1 i 2 i 3 N 0 , 1 ,
where the underlined is used to distinguish the whitened data from the original data. After the whitening operation, we can consider that the noise n ̲ i 1 i 2 i 3 is independent from the whitened signal x ̲ i 1 i 2 i 3 = x i 1 i 2 i 3 σ n i 1 i 2 i 3 .
It is worth noting that, in the likelihood function Equation (A13), the signal value x i 1 i 2 i 3 is assumed to be known. However, we cannot get this prior information in realistic situations; therefore, the signal value x i 1 i 2 i 3 should be replaced by its estimate x ˜ i 1 i 2 i 3 , as is presented in Appendix B.
Referring to [22], the wavelet-tensor-based algorithm MWPT-MWF yields the most accurate signal estimate x ˜ i 1 i 2 i 3 . In addition, the well-known HYNPE algorithm permits obtaining the ML estimates of σ ^ u , i 3 2 and σ ^ t , i 3 2 . Hence, in this paper, we choose to combine these two methods to get more accurate estimation of noise variance for each element of N , which can be calculated by:
σ ^ n i 1 i 2 i 3 2 = x ˜ i 1 i 2 i 3 σ ^ u , i 3 2 + σ ^ t , i 3 2 .
When the noise variance of each entry of R is obtained, the noise can be whitened by:
r ̲ i 1 i 2 i 3 = x ̲ i 1 i 2 i 3 + n ̲ i 1 i 2 i 3 = x i 1 i 2 i 3 σ ^ n i 1 i 2 i 3 + n i 1 i 2 i 3 σ ^ n i 1 i 2 i 3
and the whitened hyperspectral image can be written as
R ̲ = X ̲ + N ̲ .
However, MWPT-MWF was proposed for the white noise situation, therefore, when we use it to estimate x ˜ i 1 i 2 i 3 directly without noise whitening, the estimate result is not accurate. To distinguish it with the signal estimate after noise whitening, x ˜ i 1 i 2 i 3 is named as pre-estimate. Correspondingly, the estimate x ^ i 1 i 2 i 3 obtained after the noise whitening procedure is more accurate than x ˜ i 1 i 2 i 3 , so we name it as post-estimate. The PWP process needs to be repeated several times to improve the performance of estimation. In fact, HYNPE only takes the first pre-estimation step in the PWP procedure, which degrades its performance when estimating the parameters. However, we utilize the more accurate post-estimate of current PWP iteration as the pre-estimate of the next PWP iteration. Therefore, the estimate accuracy can be improved in the PWP loop.
To adaptively stop the PWP loop according to the processed HSI, we need to find a stop criterion. The RMSE between the pre-estimate X ˜ and the post-estimate X ^ is given as follows:
R M S E X = X ^ X ˜ 2 I 1 I 2 I 3 X ^ 2 ,
where X ˜ and X ^ are the tensor forms of the pre-estimate and post-estimate, respectively. With the iteration times increasing, the R M S E X becomes asymptotically stable. Hence, we can use the relative error of R M S E X between two adjacent iterations as the stop criterion:
e = | R M S E X R M S E X 0 | R M S E X 0 ,
where R M S E X 0 is the RMSE of last iteration. If e is less than a given value ϵ , the loop should be terminated. This newly proposed method is called Signal-Dependent-Noise-Whitening MWPT-MWF(SDNW-MWPT-MWF), and, to make it easy to understand, its pseudo-code and flowchart are also supplied in Algorithm 1 and Figure 1, respectively.
Algorithm 1: SDNW-MWPT-MWF algorithm
  • procedure  SDNW-MWPT-MWF  Tensor R
  • Set the maximum iteration times J.
  • Set R M S E X 0 = 1 .
  • Compute the signal pre-estimate X ˜ by performing the MWPT-MWF to the data tensor R .
  • for j = 1; j <= J; j++  do
  • Compute the SD and SI noise variance estimates σ ^ u , i 3 2 and σ ^ t , i 3 2 by using Equation (A12).
  • Compute the noise variance σ ^ n i 1 i 2 i 3 2 of each element of R : σ ^ n i 1 i 2 i 3 2 = x ˜ i 1 i 2 i 3 σ ^ u , i 3 2 + σ ^ t , i 3 2 .
  • Compute the whitened tensor R ̲ by whitening each element r i 1 i 2 i 3 of R : r ̲ i 1 i 2 i 3 = r i 1 i 2 i 3 σ ^ n i 1 i 2 i 3 .
  • Compute the whitened signal post-estimate X ^ ̲ by performing the MWPT-MWF to the whitened tensor R ̲ .
  • Compute the signal post-estimate X ^ by performing the inverse whitening operation to X ^ ̲ : x ^ i 1 i 2 i 3 = x ^ ̲ i 1 i 2 i 3 × σ ^ n i 1 i 2 i 3 .
  • Compute R M S E X by using Equation (13).
  • Compute   e   by using Equation (14).
  • if   e < ϵ    then
  •     Break.
  • end if
  • Use the post-estimate X ^ in this iteration as the pre-estimate X ^ in next iteration: X ˜ X ^ .
  • Refresh the value of R M S E X 0 : R M S E X 0 R M S E X .
  • end for
  • return tensor X ^ .
  • end procedure

4. Experimental Results

The data set used in the experiments is the HSI captured by the ROSIS during a flight campaign over Pavia University, Northern Italy.
The ROSIS owns 103 spectral bands and 610 × 340 pixels with the geometric resolution being 1.3 m. In this paper, only a part of 250 × 250 pixels of this image is used. Hence, it is modeled as a 250 × 250 × 103 tensor in the experiments. The SD photon noise P and the SI thermal noise T are both taken into account. In order to reproduce different noise scenarios, the SNR ranged from 20 dB to 40 dB with a step of 5 dB. As the power of the SD photon noise and that of the SI thermal noise are of the same level, in this paper, only the case E [ X X T P 2 ] = E [ T 2 ] is taken into account. The random noise is generated with a variance depending on the value of the useful signal according to Equation (8) and added into the signal X as Equation (3) to create the noisy HSI data R . The raw HSI has SNR between 35 and 40 dB [26,27]. This high-SNR HSI could be viewed as a noise-free data cube, so, in this experiment, the raw image can be taken as a reference data cube X .
The RGB composites of X and R are shown in Figure 2. Correspondingly, Figure 3 presents the curve of the mean noise variance versus the band number.
In the experiments, the wavelet db3 and transform level [1 1 0] are employed in the SDNW-MWPT-MWF. It is worth noting that various denoising methods can be used to replace the MWPT-MWF in the proposed SDNW-MWPT-MWF method. MWF is a classical tensor-based denoising method and MLR was used in the HYNPE method. Therefore, MWF and MLR have been considered to replace the MWPT-MWF as comparative experiments and are named as SDNW-MWF and SDNW-MLR, respectively. In this paper, the value of ϵ is set to 10 3 for these three methods.

4.1. Noise-Whitening Performance Evaluation and Comparison

According to the relationship given in Equation (10), the noise variance estimate σ ^ n i 1 i 2 i 3 2 relies on the SD noise variance estimate σ ^ u , i 3 2 and the SI noise variance estimate σ ^ t , i 3 2 . Hence, the performance of estimating σ ^ u , i 3 2 and σ ^ t , i 3 2 influences directly the noise variance estimation result. Thus, we firstly consider the evolution of σ ^ u , i 3 2 and σ ^ t , i 3 2 in the estimation loop. The RMSE is employed to analyze the accuracy of σ ^ u , i 3 2 and σ ^ t , i 3 2 . The RMSE of the SD photon noise variance and the SI thermal noise variance are calculated by
R M S E S D = 1 I 3 i 3 = 1 I 3 σ ^ u , i 3 2 σ u , i 3 2 σ u , i 3 2 2 ,
R M S E S I = 1 I 3 i 3 = 1 I 3 σ ^ t , i 3 2 σ t , i 3 2 σ t , i 3 2 2 .
Notice that low values for R M S E S D and R M S E S I denote good estimation accuracy.
A comparative analysis of SDNW-MWF, SDNW-MLR and SDNW-MWPT-MWF have been carried out by analyzing R M S E S D and R M S E S I with the maximum iteration times being set as 10. Figure 4 and Figure 5 present the evolution of R M S E S D and R M S E S I (in logarithmic scale) with the iteration times. From these two figures, it can be seen that, in the case where S N R I N P U T = 20 dB, SDNW-MLR performs better than SDNW-MWF in estimating both σ u , i 3 2 and σ t , i 3 2 . However, in the case where S N R I N P U T = 40 dB, SDNW-MWF outperforms SDNW-MLR. Nonetheless, in both cases, the proposed SDNW-MWPT-MWF can improve the estimation performance significantly according to the lowest R M S E S D and R M S E S I it obtains. Moreover, R M S E S D and R M S E S I are greater than the initial error in SDNW-MLR and SDNW-MWF, whereas R M S E S D and R M S E S I are well constrained in SDNW-MWPT-MWF.
Apart from σ u , i 3 2 and σ t , i 3 2 , the estimate of the signal also influences the accuracy of the noise variance estimate of a pixel (see Equation (10)). Since the post-estimate x ^ i 1 i 2 i 3 of current PWP iteration is used as the pre-estimate x ˜ i 1 i 2 i 3 of the next PWP iteration, we only analyze the estimation performance of the post-estimate x ^ i 1 i 2 i 3 . To assess the performance of the signal estimator x ^ i 1 i 2 i 3 , we resort to the S N R I N P U T , which will be presented in Section 4.2.
To intuitively present the noise whitening results, Figure 6 shows the mean noise variance of each band after the noise whitening operation. It is evident that the mean noise variance generated by SDNW-MWPT-MWF changes slightly around 1 and is quite constant with respect to the band number. However, the mean noise variance generated by SDNW-MLR and SDNW-MWF is not very satisfactory. In Figure 6a, it can be seen that the trend of the mean noise variances in the lower bands is similar to that in Figure 3 thus implying that the noise in the lower bands (from 1 to 20) is not well whitened. On the other hand, in the bands from 20 to 100, the noise variances are relatively constant, the mean noise variance value is not 1. In Figure 6b, the noise whitening results by SDNW-MLR and SDNW-MWF are worse than that in Figure 6a. Hence, the whitening results in Figure 6 are strongly in favor of the proposed method SDNW-MWPT-MWF.
The normal probability plot is usually used to visually ascertain whether or not a dataset is approximately normally distributed, which means that the values in the dataset have the same noise variance. Hence, we supply the normal probability plots of the noise before and after whitening in Figure 7. It is obvious that the noise before whitening (Figure 7a) is not normally distributed. After whitening by SDNW-MWF (Figure 7b), the noise approaches towards the normal distribution though there still remain some values not well whitened as can be observed in the interval [−5, 0]. Nonetheless, the noise values after whitening by SDNW-MLR (Figure 7c) and SDNW-MWPT-MWF (Figure 7d) form a straight line and as such they can be considered as normally distributed. Correspondingly, Figure 8a presents the noise distribution situation in the noise environment where S N R I N P U T = 40 dB. It can be seen that the noise values after whitening by SDNW-MLR (Figure 8b) and SDNW-MWF (Figure 8c) are still not normally distributed, while the noise values after whitening by SDNW-MWPT-MWF (Figure 8d) are well normally distributed. The results in Figure 7 and Figure 8 validate once again that the proposed SDNW-MWPT-MWF performs well in both lower ( S N R I N P U T = 20 dB) and higher ( S N R I N P U T = 40 dB) SNR noise environments.

4.2. Denoising Performance

In Section 4.1, we have discussed in detail the SD and SI noise variance estimates R M S E S D and R M S E S I . In this subsection, we show some results about the denoising performance. Figure 9 presents the noise removed in band 10 from the ROSIS HSI by SDNW-MLR, SDNW-MWF and SDNW-MWPT-MWF in the noise environment where S N R I N P U T = 20 dB. It is obvious that the removed noise is SD noise by comparing visually Figure 9 with the original image Figure 2a. The experimental results in Figure 9 imply that the proposed denoising framework is efficient in removing the SD photon noise in the HSI.
Additionally, we have assessed the denoising performance of various methods by analyzing the criterion S N R O U T P U T . Figure 10 presents the evolution of the S N R O U T P U T in various noisy environments. It is obvious that the S N R O U T P U T generated by SDNW-MWPT-MWF reaches a higher stable value after several iterations. Conversely, the S N R O U T P U T of SDNW-MLR and SDNW-MWF are relatively lower than that of SDNW-MWPT-MWF due to the highest estimation errors of R M S E S D and R M S E S I as shown in Figure 4 and Figure 5, respectively. When S N R I N P U T = 20 dB, the S N R O U T P U T generated by SDNW-MWF is only improved marginally compared to the S N R I N P U T . On the other hand, when S N R I N P U T = 40 dB SDNW-MWF improves the SNR significantly but the S N R O U T P U T is not stable in the evolution when the iterations are increased. The SDNW-MLR performs well when S N R I N P U T = 20 dB, but when S N R I N P U T = 40 dB the denoising performance of SDNW-MLR is not good compared to that of SDNW-MWF and SDNW-MWPT-MWF. This trend becomes worse as the S N R O U T P U T of SDNW-MLR is lower than the S N R I N P U T . Moreover, Figure 11 compares the S N R O U T P U T of each method from 20 dB to 40 dB with a step of 5 dB. It shows that SDNW-MLR can improve the SNR from 20 to 30 dB, while degrading the SNR from 35 to 40 dB. The SDNW-MWF is even worse as there is only a marginal improvement of the SNR in 20 dB. From 25 to 40 dB, it degrades the SNR. Nonetheless, from the results presented in Figure 10 and Figure 11, we can conclude that SDNW-MWPT-MWF is a stable and reliable denoising method in various noisy environments.

4.3. Classification after Denoising

In Section 4.2, we have mainly compared various methods with the capability of improving the SNR. However, some methods might also modify the useful signal severely in the denoising process and which cannot be reflected by the SNR. The classification is employed to distinguish different materials in HSIs [21], and it is sensitive to the signal distortion. Hence, in this paper, we take into account the classification improvement ability of the various methods considered.
Two real-world images are considered for this investigation. The first one, referred to as ROSIS HSI, is described in the beginning of this section.
In the ROSIS HSI, there are nine classes: bitumen, self-blocking bricks, trees, shadows, gravel, bare soil, asphalt, painted metal sheets, and meadows, which are shown in Figure 2b and in the ground truth in Figure 12 with different colors. A proportion of 10 % of the reference data of each class is randomly selected as the training samples. The numbers of training and testing samples are shown in Table 1. The SVM classifier is employed to do the classification and its kernel function is RBF with γ = 1 and the penalty parameter C = 100 .
The second one, referred to as HYDICE HSI, was acquired by the HYperspectral Digital Imagery Collection Experiment (HYDICE) and has 148 spectral bands (from 435 to 2326 nm), 310 rows, and 220 columns. The scene is shown in Figure 13a. This HSI is modeled as a tensor R R 310 × 220 × 148 and its ground truth is shown in Figure 13b. According to the ground truth, there are seven land cover classes in HYDICE HSI: field, trees, road, shadow and three different targets. A proportion of 30 % of the reference data of each class is randomly selected as the training samples. The numbers of training and testing samples are shown in Table 2.
The classification is applied to the denoised HSIs using various methods in order to compare their abilities of improving the classification performance. Figure 14 presents the classification results obtained in the noisy environment where S N R I N P U T = 30 dB. It can be seen from the “bare soil” class that there are less misclassified pixels in the results obtained by SDNW-MWPT-MWF compared to the other results. To make it easy to compare, we give the Overall Accuracy (OA) and Kappa coefficient (K) results for the various methods used in Table 3. It can be seen that if the classification is applied directly to the noisy HSI, the OA is only 91.33 % and K = 0.71 . After denoising by SDNW-MLR, there is a marginal improvement of OA resulting in 91.88 % and K = 0.81 . SDNW-MWF performs better than SDNW-MLR, and its OA is increased to 94.20 % and K to 0.94 . The proposed SDNW-MWPT-MWF makes the most significant improvement among the denoising methods. Indeed, its OA is 98.51 % and its K is 0.97 . The classification result clearly shows that the proposed SDNW-MWT-MWF is efficient in improving the classification performance.
To investigate how the number of training samples affects the performance of our method and other comparison methods, we considered a proportion of 50 % of the reference data of each class is randomly selected as the training samples. Table 4 shows the OA and Kappa coefficient results. As shown from Table 3 and Table 4, our proposed method outperforms all other methods. In particular, with the number of training samples reducing ( 10 % ), our method could obtain bigger accuracy gains than other comparison methods.
Figure 15 shows the OA values obtained from the denoised HYDICE HSI and shows that the SDNW-MWPT-MWF method permits reduction of the noise, which is of great interest for SVM classifier. The comparison of the OA and Kappa values (see Table 5) calculated for each preprocessing of denoising shows that the multilinear algebra-based method SDNW-MWPT-MWF leads to better classification results than the considered methods in this experiment.
To make the experimental results more convincing, we have compared the classification improvement performance of SDNW-MLR, SDNW-MWF and SDNW-MWPT-MWF in varying noisy environments, i.e., S N R I N P U T varies from 20 dB to 40 dB with a step of 5 dB. The classification result of the noisy HSI without denoising is also supplied as a benchmark. Figure 16 presents the curves of OA versus S N R I N P U T . In the cases where the HSI is impaired severely ( S N R I N P U T = 20 and 25 dB), the classification result can be improved significantly after denoising, which implies that the denoising is a necessary preprocessing procedure prior to the classification. Above 30 dB, SDNW-MLR can only improve the classification result marginally and the same is valid for SDNW-MWF with more than 35 dB. Nonetheless, SDNW-MWPT-MWF performs better than SDNW-MLR and SDNW-MWF and it improves the OA significantly for S N R I N P U T values from 20 dB to 40 dB.

5. Discussion

Hyperspectral sensors collect data in hundreds of narrow contiguous spectral bands, providing a powerful means to discriminate different materials by their spectral features. In the last decade, for improving the classification, several denoising algorithms for hyperspectral images were proposed. Most were derived assuming the spatial stationarity of the noise that affects hyperspectral images, meaning that the noise characteristics are assumed to be the same in each hyperspectral image region. The existing algorithms are proposed to remove the signal independent noise for enhancing hyperspectral image applications. In this study, we have proved that assumption is not valid for new-generation hyperspectral sensors, where photon noise, which depends on the spatially varying signal level, is not negligible. We are thus studying the possible impacts of signal-dependent noise on the performance of existing classification algorithms. By assuming a signal-dependent noise model, we have proposed a novel multilinear algebra-based algorithm to remove simultaneously the signal-independent noise and signal-dependent noise. This result has inspired a new iterative procedure noise-whitening (SDNW-MWPT-MWF) which improves the SVM’s robustness in the presence of signal-dependent noise. The noise-whitening is achieved by exploiting estimates of the noise variance for each band and for each pixel. The performance of the proposed SDNW-MWPT-MWF method are validated on the simulated HSIs disturbed by both SD and SI noise and on the real-world ROSIS and HYDICE HSIs. From the analysis and the comparative study against other similar methods in the experiments, it can be concluded that SDNW-MWPT-MWF method can effectively reduce both SD and SI noise from HSIs. It is also necessary to take into account the signal-dependent noise in the denoising when dealing with HSIs that were collected by a new-generation airborne hyperspectral sensors. Indeed, this study demonstrated that the signal-dependent noise may affect the properties of existing classification algorithms, thus encouraging future work about the impact of SD noise. Our ongoing activity is aimed at demonstrating the benefits arising from using the SDNW-MWPT-MWF algorithm in mitigating the impact of the SD noise in different algorithms for hyperspectral data exploitation.
Since in this study the optimal parameter combination is found by time-consuming brute force searching, future works will be focused on the reduction of the computational load. A heuristic algorithm can be used to search for the optimal (sub-optimal) parameter combination such as a genetic algorithm.

6. Conclusions

In this paper, a denoising method SDNW-MWPT-MWF has been proposed under the PWP denoising framework for the purpose of reducing the SD photon and SI thermal noise in HSIs. Two other adapted denoising methods SDNW-MLR, SDNW-MWF are also used as comparative methods. The performances of these methods are compared in the case of photon noise whitening, denoising and improvement of classification. The photon noise whitening experiment, the denoising experiment and the classification experiment are designed to assess the parameter estimation performance, the improvement of the SNR and the distortion of the spectra in the HSI, respectively. From the experimental results obtained, it can be concluded that the SDNW-MLR performs well in removing the noise, though it also changes the signal severely in the denoising process as shown from the lower classification result OA. In contrast, SDNW-MWF performs well in preserving the signal though it cannot remove the noise well from the images. However, the proposed SDNW-MWPT-MWF is able to obtain high S N R O U T P U T as well as the high OA, thus implying that it is able to remove noise well while still preserving the signal.
These promising results encourage us to extend our experiments on other hyperspectral data such as Indian Pines and Salinas HSIs.
In this study, we employed the HYNPE algorithm to estimate the noise parameters, and then filtered noise by the proposed method SDNW-MWPT-MWF. Nonetheless, SDNW-MWPT-MWF converts SD noise to additive white Gaussian noise by a pre-whitening procedure and then applied MWPT-MWF. Due to the parameter estimate errors, the whitened noise is only approximately white, therefore the performance of filters developed for the white noise might degrade. Thus, it is interesting to develop a filter that can directly process the SD noise without the noise whitening procedure. For example, the tensor-based filtering method using a PARAFAC tensor decomposition could be carried out in the future.

Author Contributions

All authors contributed to the conception and design of the methodology.

Funding

This research received no external funding.

Acknowledgments

The authors would like to thank the reviewers for their careful reading and helpful comments, which improved the quality of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Multilinear Algebra Tools

Since multilinear algebra is used in this paper, we present some basic definitions in this section to make it easy to understand the discussions in the following sections. More details about multilinear algebra can be found in [28,29,30,31].

Appendix A.1. Definition of a Tensor

An N-th order tensor is an N-dimensional array, X R I 1 × × I N , in which R indicates the real manifold, and N is the number of dimensions. The elements in this tensor can be expressed as x i 1 i N , with i 1 = 1 , , I 1 ; ; i N = 1 , , I N . The nth dimension of this tensor is called mode-n.

Appendix A.2. Unfolding

Unfolding is also known as matricization or flattening. It can convert a tensor into its matrix form for the conveniency of using matrix-based data analyzing methods. A tensor can be unfolded in different ways according to the mode in which the unfolding is performed. The unfolding in mode-n is called mode-n unfolding.
Let X n R I n × M n denote the mode-n unfolding matrix of a tensor X R I 1 × × I N , where M n = I n + 1 I n + 2 I N I 1 I n 1 . The columns of X n are the I n -dimensional vectors obtained from X by varying index i n while keeping the other indices fixed.

Appendix A.3. Mode-n Tensor Product × n

The mode-n product is defined as the product between a data tensor X R I 1 × × I N and a matrix B R J × I n in mode n. This mode-n product is denoted by C = X × n B , whose entries are given by
c i 1 i n 1 j i n + 1 i N i n = 1 I n x i 1 i n 1 i n i n + 1 i N b j i n ,
where C R I 1 × × I n 1 × J × I n + 1 × × I N .

Appendix A.4. Element Extraction

For a given tensor X R I 1 × I 2 × × I N , we define the following element extraction operation:
X ( i 1 , i 2 , , i N ) x i 1 i 2 i N .

Appendix A.5. Hadamard Product

The Hadamard product is defined as the product of two equal-size tensors X R I 1 × I 2 × × I N and Y R I 1 × I 2 × × I N . The entries of the Hadamard product Z = X Y can be computed as:
z i 1 i n 1 i n i n + 1 i N = x i 1 i n 1 i n i n + 1 i N y i 1 i n 1 i n i n + 1 i N ,
where Z R I 1 × I 2 × × I N .

Appendix A.6. Mode-n Rank of a Tensor

The term mode-n rank K n of a tensor X R I 1 × I 2 × × I N , denoted by K n = rank n ( X ) , is the dimension of the column space of the mode-n unfolding matrix X n [29], i.e.,
K n = rank n ( X ) = rank ( X n ) .
The term mode-n rank is another way to extend the notion of the matrix rank to the tensor case. The mode-n rank is actually a rank of a matrix, therefore it can be analyzed by using matrix based techniques. Nevertheless, for a matrix X , the ranks of its column space and row space are similar, rank ( X ) = rank ( X T ) , but, for the tensor X , the different mode-n ranks are not necessarily the same. This means that the mode-n rank of tensor X is not a unique value but a set of N values K 1 , , K N .

Appendix B. Overview of HYNPE

HYNPE was proposed in [8] to estimate the parameters of the photon noise which has aroused recently much research interest in the new generation hyperspectral sensors. HYNPE assumes that the noise sources are independent from one another and spectrally uncorrelated, which is the same noise model as the one introduced in Section 2. There are mainly two steps of HYNPE for estimating the noise parameters.
The first step is to estimate the noise and signal by employing the MLR theory. For a given noisy HSI R , we consider its mode-3 unfolding matrix R 3 R I 3 × I 1 I 2 , which consists of I 3 row vectors (see Appendix A):
R 3 = r 1 T r 2 T r I 3 T T .
Denote the estimate of pure HSI X by X ^ and noise N by N ^ . Then, the mode-3 unfolding matrix of X ^ and N ^ are X ^ 3 and N ^ 3 , respectively, and their row-vector forms can be expressed as:
X ^ 3 = x ^ 1 T x ^ 2 T x ^ I 3 T T , N ^ 3 = n ^ 1 T n ^ 2 T n ^ I 3 T T .
The MLR theory estimates the signal by exploiting the strong spectral correlation of the useful signal and the weak between-band correlation of the random noise in the HSI. It assumes that the signal estimate row vector x ^ i 3 T in band i 3 can be expressed as a linear combination of the noisy data row vectors r j 3 T ( j 3 = 1 , , i 3 1 , i 3 + 1 , , I 3 ) in the other I 3 1 bands:
x ^ i 3 = Θ i 3 w i 3 ,
where Θ i 3 = r 1 , , r i 3 1 , r i 3 + 1 , , r I 3 and w i 3 R ( I 3 1 ) is the combination weight vector. Then, the optimal weight vector can be estimated by minimizing the LSE:
w ^ i 3 = arg min w i 3 r i 3 x ^ i 3 2 .
The solution of the LSE problem is well known and can be expressed as:
w ^ i 3 = Θ i 3 T Θ i 3 1 Θ i 3 T r i 3 ,
and the corresponding signal estimate x ^ i 3 and noise estimate n ^ i 3 can be computed by:
x ^ i 3 = Θ i 3 w ^ i 3 ,
n ^ i 3 = r i 3 x ^ i 3 .
After the estimation of the signal and noise, the noise variances σ ^ u , i 3 2 and σ ^ t , i 3 2 are estimated by maximizing the likelihood function [8]:
{ σ ^ u , i 3 2 , σ ^ t , i 3 2 } = arg max σ u , i 3 > 0 σ t , i 3 > 0 ln σ u , i 3 , σ t , i 3 ,
with
ln σ u , i 3 , σ t , i 3 = M 2 ln ( 2 π ) 1 2 i 1 = 1 I 1 i 2 = 1 I 2 ln σ u , i 3 2 · x i 1 i 2 i 3 + σ t , i 3 2 1 2 i 1 = 1 I 1 i 2 = 1 I 2 n i 1 i 2 i 3 2 σ u , i 3 2 · x i 1 i 2 i 3 + σ t , i 3 2 .
Since x i 1 i 2 i 3 and n i 1 i 2 i 3 are unknown in a realistic scenario, they are replaced by the estimates computed in Equations (A10) and (A11), respectively.

Appendix C. Multiway Wiener Filter in Multidimensional Wavelet Packet Domain

In this section, we remind the main principles of the multidimensional wavelet packet transform (MWPT) and of the MWF in multidimensional wavelet packet domain [22].

Appendix C.1. MWPT

MWPT can be computed by performing 1D wavelet packet transform in each mode [32]. Therefore, the wavelet packet coefficient tensor C l R can be computed as
C l R = R × 1 W ˜ 1 l 1 × 2 W ˜ 2 l 2 × 3 W ˜ 3 l 3 ,
and the reconstruction can be written as
R = C l R × 1 ( W ˜ 1 l 1 ) T × 2 ( W ˜ 2 l 2 ) T × 3 ( W ˜ 3 l 3 ) T ,
where l = [ l 1 , l 2 , l 3 ] T , and l 1 , l 2 , l 3 0 . In particular, when l 1 , l 2 , l 3 > 0 , MWPT indicates the 3D wavelet packet transform. W ˜ k l k denotes the l k level wavelet packet transform to kth mode of R . C l , m R is defined as the coefficient subtensor of C l R , where m = [ m 1 , m 2 , m 3 ] T is the index vector, and 0 m k 2 l k 1 , k = 1 , 2 , 3 . Then, for each element of C l , m R , we can define:
C l , m R ( j 1 , j 2 , j 3 ) C i R ( J 1 ( j 1 ) , J 2 ( j 2 ) , J 3 ( j 3 ) ) ,
where
J n = m n I n 2 l , , ( m n + 1 ) I n 2 l 1 T , n = 1 , 2 , 3
and
j n 1 , , I n 2 l , n = 1 , 2 , 3 .
The notation C l , m R ( j 1 , j 2 , j 3 ) indicates the element of tensor C l , m R in position ( j 1 , j 2 , j 3 ) as defined in Equation (A2). From the properties of the wavelet packet transform, we know that m n indicates the “frequency” of mode n. Thus, m is the frequency index of coefficient block C l , m R . For convenience, a component tensor of R is referred to as C l , m R in this paper.

Appendix C.2. Multiway Wiener Filter in Multidimensional Wavelet Packet Domain

In the existing MWF algorithm, the filter is applied to the whole hyperspectral image R . As the calculation of the filters needs the estimation of the signal subspace or rank in each mode for suppressing the smallest eigenvalues [33], some weak signal might be removed in this procedure. Therefore, the SNR is an important factor influencing the rank. When SNR is higher, the rank estimated is greater, therefore more signal is preserved in the filtering process. In the contrast condition, more signal is lost. When the noise is white, the power of noise in each component C l , m R is the same, whereas the signal concentrates in the lower frequency component. That is to say, in different components, the SNR is different. When MWF is applied to each component, more signal can be preserved. Performing MWPT to tensor R , X and N in Equation (1), we obtain:
R × 1 W ˜ 1 l 1 × 2 W ˜ 2 l 2 × 3 W ˜ 3 l 3 = X + N × 1 W ˜ 1 l 1 × 2 W ˜ 2 l 2 × 3 W ˜ 3 l 3 = X × 1 W ˜ 1 l 1 × 2 W ˜ 2 l 2 × 3 W ˜ 3 l 3 + N × 1 W ˜ 1 l 1 × 2 W ˜ 2 l 2 × 3 W ˜ 3 l 3 .
The coefficient tensor of each part
C l R = R × 1 W ˜ 1 l 1 × 2 W ˜ 2 l 2 × 3 W ˜ 3 l 3 ,
C l X = X × 1 W ˜ 1 l 1 × 2 W ˜ 2 l 2 × 3 W ˜ 3 l 3 ,
C l N = N × 1 W ˜ 1 l 1 × 2 W ˜ 2 l 2 × 3 W ˜ 3 l 3 ,
and the coefficient tensor of the estimate X ^ :
C ^ l X = X ^ × 1 W ˜ 1 l 1 × 2 W ˜ 2 l 2 × 3 W ˜ 3 l 3 .
Extracting the components of each frequency C l , m R , C l , m X and C l , m N from C l R , C l X and C l N respectively by using Equation (A16), we obtain:
C l , m R = C l , m X + C l , m N .
From Parseval’s theorem, the following expression can be obtained:
X X ^ 2 = C l X C ^ l X 2 = m C l , m X C ^ l , m X 2 ,
which means that minimizing the MSE between X and its estimate X ^ is equivalent to minimizing the MSE between C l , m X and C ^ l , m X for each m . If C ^ l , m X is estimated by Tucker3 decomposition of C l , m R :
C ^ l , m X = C l , m R × 1 H 1 , m × 2 H 2 , m × 3 H 3 , m ,
then H 1 , m , H 2 , m , H 3 , m are the mode-n filters of the multiway Wiener filter [33]. After estimating C ^ l , m X for each m , we obtain C ^ l X by concatenating C ^ l , m X . Furthermore, the estimate X ^ can be obtained by inverse MWPT, i.e.,
X ^ = C ^ l X × 1 ( W ˜ 1 l 1 ) T × 2 ( W ˜ 2 l 2 ) T × 3 ( W ˜ 3 l 3 ) T .

Appendix C.3. Best Transform Level and Basis Selection

In MWPT-MWF algorithm, several parameters should be determined:
  • Level of transform: the performance of the algorithm is affected by the level of transform, which depends on the size of tensor R . The maximum level can be calculated by:
    N L k = log 2 I k 5 , k = 1 , 2 , 3 ,
    where · rounds a number upward to its nearest integer, and the constant 5 is reduced from log 2 I k to make sure there are enough elements in each mode so that the transform is meaningful. Then, the set of possible transform levels can be expressed as:
    L k = { 0 , 1 , , N L k } , k = 1 , 2 , 3 ,
    where { · } denotes a set.
  • Basis of transform: there are many wavelet bases designed for different cases. For the simplicity of expression, we define:
    W = { w 1 , w 2 , , w N W }
    to denote the set of possible wavelet bases, where N W is the number of wavelets in this set.
The best transform level and basis should minimize the MSE or risk R c ( X , X ^ ) = E X X ^ 2 [34], whose equivalent form using the coefficients can be expressed as
R c ( X , X ^ ) = m E C l , m X C ^ l , m X 2 .
Then, the best transform level and basis can be selected by
l , w = arg min l k L k , w W m E C l , m X C ^ l , m X 2 , k = 1 , 2 , 3 .
As the selection of the optimal l , w depends on X , which is generally unknown, to overcome this drawback, an alternative solution should be found. Denoting by C ^ l , m X [ d ] the estimate of C l , m X at the dth ALS loop and noticing that when C ^ l , m X [ d ] C ^ l , m X [ d 1 ] 2 is minimized, C ^ l , m X C ^ l , m X [ d ] is the optimal estimate of C l , m X obtained by MWF, and at the same time E C l , m X C ^ l , m X 2 is minimized. Therefore, Equation (A32) can be replaced by:
l , w = arg min l k L k , w W R c ^ , k = 1 , 2 , 3 ,
where
R c ^ = m C ^ l , m X [ d ] C ^ l , m X [ d 1 ] 2 .

Appendix C.4. Summary of the MWPT-MWF Method

The algorithm MWPT−MWF, can be summarized as presented here.
  • Find the optimal l 1 , l 2 , l 3 L and w W . Loop l 1 , l 2 , l 3 and w:
    (a)
    Decompose the data R by MWPT: C l R = R × 1 W ˜ 1 l 1 × 2 W ˜ 2 l 2 × 3 W ˜ 3 l 3 .
    (b)
    Extract component C l , m R from C l R using Equation (A16), for m = [ m 1 , m 2 , m 3 ] T , where 0 m k 2 l k 1 , k = 1 , 2 , 3 .
    (c)
    Filter component C l , m R by MWF: C ^ l , m X = C l , m R × 1 H 1 , m × 2 H 2 , m × 3 H 3 , m .
    (d)
    Calculate the risk R c ^ using Equation (A34). If R c ^ reaches a fixed threshold, return the optimal l 1 , l 2 , l 3 , w and C ^ l , m X .
  • Concatenate C ^ l , m X to obtain C l X and perform inverse MWPT: X ^ = C ^ l X × 1 ( W ˜ 1 l 1 ) T × 2 ( W ˜ 2 l 2 ) T × 3 ( W ˜ 3 l 3 ) T .

References

  1. Acito, N.; Diani, M.; Corsini, G. Signal-dependent noise modeling and model parameter estimation in hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2011, 8, 2957–2971. [Google Scholar] [CrossRef]
  2. Rakwatin, P.; Takeuchi, W.; Yasuoka, Y. Stripe noise reduction in modis data by combining histogram matching with facet filter. IEEE Trans. Geosci. Remote Sens. 2007, 45, 1844–1856. [Google Scholar] [CrossRef]
  3. Uss, M.L.; Vozel, B.; Lukin, V.V.; Chehdi, K. Local signal-dependent noise variance estimation from hyperspectral textural images. IEEE J. Sel. Top. Signal Process. 2011, 3, 469–486. [Google Scholar] [CrossRef]
  4. Roger, R.E. Principal components transform with simple, automatic noise ajustment. INT J. Remote Sens. 1996, 17, 2719–2727. [Google Scholar] [CrossRef]
  5. Chang, C.-I.; Du, Q. Interference and noise adjusted principal components analysis. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2387–2396. [Google Scholar] [CrossRef]
  6. Rhodes, H.; Agranov, G.; Hong, C.; Boettiger, U.; Mauritzson, R.; Ladd, J.; Karasev, I.; McKee, J.; Jenkins, E.; Quinlin, W.; et al. Cmos imager technology shrinks and image performance. In Proceedings of the 2004 IEEE Workshop on Microelectronics and Electron Devices, Boise, ID, USA, 16 April 2004; pp. 7–18. [Google Scholar]
  7. Alparone, L.; Selva, M.; Aiazzi, B.; Baronti, S.; Butera, F.; Chiarantini, L. Signal-dependent noise modelling and estimation of new-generation imaging spectrometers. In Proceedings of the Workshop on Hyperspectral Image and Signal Processing (WHISPERS), Grenoble, France, 26–28 August 2009; pp. 1–4. [Google Scholar]
  8. Liu, X.; Bourennane, S.; Fossati, C. Denoising of hyperspectral images using the parafac model and statistical performance analysis. IEEE Trans. Geosci. Remote Sens. 2012, 50, 3717–3724. [Google Scholar] [CrossRef]
  9. Gao, L.; Yao, D.; Li, Q.; Zhuang, L.; Zhang, B.; Bioucas-Dias, J.M. A New Low-Rank Representation based hyperspectral image denoising method for mineral mapping. Remote Sens. 2017, 9, 1145. [Google Scholar] [CrossRef]
  10. Fan, Y.R.; Huang, T.Z.; Zhao, X.L.; Deng, L.J.; Fan, S. Multispectral Image Denoising via Nonlocal Multitask Sparse Learning. Remote Sens. 2018, 10, 116. [Google Scholar] [CrossRef]
  11. Atkinson, I.; Kamalabadi, F.; Mohan, S.; Jones, D. Wavelet-based 2D multichannel signal estimation. In Proceedings of the IEEE International Conference on Image Processing, Barcelona, Spain, 14–17 September 2003; pp. 743–745. [Google Scholar]
  12. Pizurica, A.; Philips, W. Estimating the probability of the presence of a signal of interest in multiresolution single-and multiband image denoising. IEEE Trans. Image Process. 2006, 3, 654–665. [Google Scholar] [CrossRef]
  13. Letexier, D.; Bourennane, S. Noise removal from hyperspectral images by multidimensional filtering. IEEE Trans. Geosci. Remote Sens. 2008, 7, 2061–2069. [Google Scholar] [CrossRef]
  14. Renard, N.; Bourennane, S. Improvement of target detection methods by multiway filtering. IEEE Trans. Geosci. Remote Sens. 2008, 8, 2407–2417. [Google Scholar] [CrossRef]
  15. Aiazzi, B.; Alparone, L.; Barducci, A.; Baronti, S.; Marcoinni, P.; Pippi, I.; Selva, M. Noise modelling and estimation of hyperspectral data from airborne imaging spectrometers. Ann. Geophys. 2006, 1, 1–9. [Google Scholar]
  16. Gao, L.R.; Zhang, B.; Zhang, X.; Zhang, W.-J.; Tong, Q.-X. A new operational method for estimating noise in hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2008, 1, 83–87. [Google Scholar] [CrossRef]
  17. Aiazzi, B.; Alparone, L.; Baronti, S. A robust method for parameter estimation of signal-dependent noise models in digital images. In Proceedings of the IEEE International Conference on Digital Signal Processin, Santorini, Greece, 2–4 July 1997; pp. 601–604. [Google Scholar]
  18. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A. Coherence estimation from multilook incoherent sar imagery. IEEE Trans. Geosci. Remote Sens. 2003, 11, 2531–2539. [Google Scholar] [CrossRef]
  19. Argenti, F.; Torricelli, G.; Alparone, L. Mmse filtering of generalised signal-dependent noise in spatial and shift-invariant wavelet domains. Signal Process. 2006, 8, 2056–2066. [Google Scholar] [CrossRef]
  20. Foi, A.; Trimeche, M.; Katkovnik, V.; Egiazarian, K. Practical poissonian-Gaussian noise modeling and fitting for single-image raw-data. IEEE Trans. Image Process. 2008, 10, 1737–1754. [Google Scholar] [CrossRef] [PubMed]
  21. Cheng, G.; Li, Z.; Han, J.; Yao, X.; Guo, L. Exploring Hierarchical Convolutional Features for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 99, 1–11. [Google Scholar] [CrossRef]
  22. Lin, T.; Bourennane, S. Hyperspectral image processing by jointly filtering wavelet component tensor. IEEE Trans. Geosci. Remote Sens. 2013, 6, 3529–3541. [Google Scholar] [CrossRef]
  23. Lin, T.; Bourennane, S. Hyperspectral image denoising with rare signal preserving by jointly filtering image component. In Proceedings of the 2013 4th European Workshop on Visual Information Processing (EUVIP), Paris, France, 10–12 June 2013; pp. 265–269. [Google Scholar]
  24. Faraji, H.; MacLean, W.J. CCD noise removal in digital images. IEEE Trans. Image Process. 2006, 9, 2676–2685. [Google Scholar] [CrossRef]
  25. Aiazzi, B.; Alparone, L.; Baronti, S.; Butera, F.; Chiarantini, L.; Selva, M. Benefits of signal-dependent noise reduction for spectral analysis of data from advanced imaging spectrometers. In Proceedings of the Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Lisbon, Portugal, 6–9 June 2011; pp. 1–4. [Google Scholar]
  26. Othman, H.; Qian, S.-E. Noise reduction of hyperspectral imagery using hybrid spatial-spectral derivative-domain wavelet shrinkage. IEEE Trans. Geosci. Remote Sens. 2006, 2, 397–408. [Google Scholar] [CrossRef]
  27. Chen, G.; Qian, S.-E. Denoising of hyperspectral imagery using principal component analysis and wavelet shrinkage. IEEE Trans. Geosci. Remote Sens. 2011, 3, 973–980. [Google Scholar] [CrossRef]
  28. Kolda, T.G.; Bader, B.W. Tensor decompositions and applications. SIAM Rev. 2009, 3, 455–500. [Google Scholar] [CrossRef]
  29. De Lathauwer, L.; De Moor, B.; Vandewalle, J. A multilinear singular value decomposition. SIAM J. Matrix Anal. Appl. 2000, 4, 1253–1278. [Google Scholar] [CrossRef]
  30. De Lathauwer, L.; De Moor, B.; Vandewalle, J. On the best rank-1 and rank-(r1, r2, ..., rn) approximation of higher-order tensors. SIAM J. Matrix Anal. Appl. 2000, 4, 1324–1342. [Google Scholar] [CrossRef]
  31. Cichocki, A.; Zdunek, R.; Phan, A.; Amari, S. Nonnegative Matrix and Tensor Factorizations: Applications to Exploratory Multi-Way Data Analysis and Blind Source Separation; Wiley: Hoboken, NJ, USA, 2009. [Google Scholar]
  32. Mallat, S. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 7, 674–693. [Google Scholar] [CrossRef]
  33. Muti, D.; Bourennane, S. Multidimensional filtering based on a tensor approach. Signal Process. 2005, 12, 2338–2353. [Google Scholar] [CrossRef]
  34. Donoho, D.; Johnstone, I. Ideal denoising in an orthonormal basis chosen from a library of bases. Comptes Rendus de l’Academie des Sciences-Serie I-Mathematique 1994, 12, 1317–1322. [Google Scholar]
Figure 1. Flowchart of the SDNW-MWPT-MWF algorithm.
Figure 1. Flowchart of the SDNW-MWPT-MWF algorithm.
Remotesensing 10 01330 g001
Figure 2. RGB composites of X and R (band 20, 35 and 45 for red, green and blue): (a) RGB composites of X ; (b) RGB composites of R .
Figure 2. RGB composites of X and R (band 20, 35 and 45 for red, green and blue): (a) RGB composites of X ; (b) RGB composites of R .
Remotesensing 10 01330 g002
Figure 3. Mean noise variance in each band ( S N R I N P U T = 20 dB).
Figure 3. Mean noise variance in each band ( S N R I N P U T = 20 dB).
Remotesensing 10 01330 g003
Figure 4. Evolution of R M S E S D with iteration times according to different values of S N R I N P U T : (a) 20 dB; (b) 40 dB.
Figure 4. Evolution of R M S E S D with iteration times according to different values of S N R I N P U T : (a) 20 dB; (b) 40 dB.
Remotesensing 10 01330 g004
Figure 5. Evolution of R M S E S I with iteration times according to different values of S N R I N P U T : (a) 20 dB; (b) 40 dB.
Figure 5. Evolution of R M S E S I with iteration times according to different values of S N R I N P U T : (a) 20 dB; (b) 40 dB.
Remotesensing 10 01330 g005
Figure 6. Curves of noise variance versus band number with S N R I N P U T : (a) 20 dB; (b) 40 dB.
Figure 6. Curves of noise variance versus band number with S N R I N P U T : (a) 20 dB; (b) 40 dB.
Remotesensing 10 01330 g006
Figure 7. Normal probability plot for the noise: (a) before whitening, and whitening after (b) SDNW-MWF; (c) SDNW-MLR; (d) SDNW-MWPT-MWF, with S N R I N P U T = 20 dB.
Figure 7. Normal probability plot for the noise: (a) before whitening, and whitening after (b) SDNW-MWF; (c) SDNW-MLR; (d) SDNW-MWPT-MWF, with S N R I N P U T = 20 dB.
Remotesensing 10 01330 g007aRemotesensing 10 01330 g007b
Figure 8. Normal probability plot for the noise: (a) before whitening, and after whitening using (b) SDNW-MLR; (c) SDNW-MWF; (d) SDNW-MWPT-MWF, with S N R I N P U T = 40 dB.
Figure 8. Normal probability plot for the noise: (a) before whitening, and after whitening using (b) SDNW-MLR; (c) SDNW-MWF; (d) SDNW-MWPT-MWF, with S N R I N P U T = 40 dB.
Remotesensing 10 01330 g008
Figure 9. Noise removal using: (a) SDNW-MLR; (b) SDNW-MWF; (c) SDNW-MWPT-MWF, band 10, S N R I N P U T = 20 dB.
Figure 9. Noise removal using: (a) SDNW-MLR; (b) SDNW-MWF; (c) SDNW-MWPT-MWF, band 10, S N R I N P U T = 20 dB.
Remotesensing 10 01330 g009
Figure 10. Evolution of S N R O U T P U T with iteration times according to different values of S N R I N P U T : (a) 20 dB; (b) 40 dB.
Figure 10. Evolution of S N R O U T P U T with iteration times according to different values of S N R I N P U T : (a) 20 dB; (b) 40 dB.
Remotesensing 10 01330 g010
Figure 11. Comparison of denoising results based on S N R O U T P U T versus S N R I N P U T .
Figure 11. Comparison of denoising results based on S N R O U T P U T versus S N R I N P U T .
Remotesensing 10 01330 g011
Figure 12. Classification reference data:(a) ground truth of the area; (b) nine classes in ROSIS HSI.
Figure 12. Classification reference data:(a) ground truth of the area; (b) nine classes in ROSIS HSI.
Remotesensing 10 01330 g012
Figure 13. HSI images: (a) ground truth; (b) classes in HYDICE HSI.
Figure 13. HSI images: (a) ground truth; (b) classes in HYDICE HSI.
Remotesensing 10 01330 g013
Figure 14. Classification results of the ROSIS HSI after denoising by: (a) SDNW-MLR; (b) SDNW-MWF; and (c) SDNW-MWPT-MWF. The classification result without denoising (d) is supplied as a benchmark. ( S N R I N P U T = 30 dB).
Figure 14. Classification results of the ROSIS HSI after denoising by: (a) SDNW-MLR; (b) SDNW-MWF; and (c) SDNW-MWPT-MWF. The classification result without denoising (d) is supplied as a benchmark. ( S N R I N P U T = 30 dB).
Remotesensing 10 01330 g014
Figure 15. Classification results of the HYDICE HSI after denoising by: (a) SDNW-MLR; (b) SDNW-MWF; and (c) SDNW-MWPT-MWF. The classification result without denoising (d) is supplied as a benchmark. ( S N R I N P U T = 30 dB).
Figure 15. Classification results of the HYDICE HSI after denoising by: (a) SDNW-MLR; (b) SDNW-MWF; and (c) SDNW-MWPT-MWF. The classification result without denoising (d) is supplied as a benchmark. ( S N R I N P U T = 30 dB).
Remotesensing 10 01330 g015
Figure 16. Classification OA with respect to S N R I N P U T .
Figure 16. Classification OA with respect to S N R I N P U T .
Remotesensing 10 01330 g016
Table 1. Training and testing samples used in the classification.
Table 1. Training and testing samples used in the classification.
IDClassTraining SamplesTesting Samples
1Bitumen1331330
2Self-Blocking Bricks1711709
3Trees70697
4Shadows49486
5Gravel93929
6Bare Soil5035029
7Asphalt1871868
8Painted metal sheets1351345
9Meadows2012005
Total154215,398
Table 2. Training and testing samples used in the classification.
Table 2. Training and testing samples used in the classification.
IDClassTraining SamplesTesting SamplesColor
1Field12,17440,811Green
2Trees13615537Sea green
3Road11463226White
4Shadow13635036Maroon
5Target 1138519Red
6Target 278285Blue
7Target 367223Yellow
Total16,32755,637
Table 3. OA (%) and Kappa of the classification of the denoised ROSIS HSI, S N R I N P U T = 30 dB.
Table 3. OA (%) and Kappa of the classification of the denoised ROSIS HSI, S N R I N P U T = 30 dB.
MethodsWithout DenoisingSDNW-MLRSDNW-MWFSDNW-MWPT-MWF
OA91.3391.8894.2098.51
Improvement00.552.877.18
Kappa0.790.810.910.96
Improvement00.020.120.17
Table 4. OA (%) and Kappa of the classification of the denoised ROSIS HSI, S N R I N P U T = 30 dB.
Table 4. OA (%) and Kappa of the classification of the denoised ROSIS HSI, S N R I N P U T = 30 dB.
MethodsWithout DenoisingSDNW-MLRSDNW-MWFSDNW-MWPT-MWF
OA91.9892.7395.9299.06
Improvement00.753.947.08
Kappa0.810.840.940.98
Improvement00.030.130.16
Table 5. OA (%) and Kappa of the classification of the denoised HYDICE HSI, S N R I N P U T = 30 dB.
Table 5. OA (%) and Kappa of the classification of the denoised HYDICE HSI, S N R I N P U T = 30 dB.
MethodsWithout DenoisingSDNW-MLRSDNW-MWFSDNW-MWPT-MWF
OA95.6095.6697.2299.94
Improvement00.061.624.34
Kappa0.8940.8950.9450.989
Improvement00.010.510.95

Share and Cite

MDPI and ACS Style

Bourennane, S.; Fossati, C.; Lin, T. Noise Removal Based on Tensor Modelling for Hyperspectral Image Classification. Remote Sens. 2018, 10, 1330. https://doi.org/10.3390/rs10091330

AMA Style

Bourennane S, Fossati C, Lin T. Noise Removal Based on Tensor Modelling for Hyperspectral Image Classification. Remote Sensing. 2018; 10(9):1330. https://doi.org/10.3390/rs10091330

Chicago/Turabian Style

Bourennane, Salah, Caroline Fossati, and Tao Lin. 2018. "Noise Removal Based on Tensor Modelling for Hyperspectral Image Classification" Remote Sensing 10, no. 9: 1330. https://doi.org/10.3390/rs10091330

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop