Next Article in Journal
Adhering Solid Precipitation in the Current and Pseudo-Global Warming Future Climate over the Canadian Provinces of Manitoba and Saskatchewan
Next Article in Special Issue
Bayesian Analysis of Spatial Model for Frequency of Tornadoes
Previous Article in Journal
Long-Term Exposure to Ambient Fine Particles and Heart Rate in Northwestern China: Findings from 1.8 Million Adults of the Kashgar Prospective Cohort Study (KPCS)
Previous Article in Special Issue
Study on the Peak Factor of the Wind-Induced Response of Super-High-Rise Buildings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Two-Stage Decomposition Multi-Scale Nonlinear Ensemble Model with Error-Correction-Coupled Gaussian Process for Wind Speed Forecast

1
School of Management Science and Engineering, Nanjing University of Information Science and Technology, Nanjing 210044, China
2
Institute of Climate Economy and Low-Carbon Industry, Nanjing University of Information Science and Technology, Nanjing 210044, China
*
Author to whom correspondence should be addressed.
Atmosphere 2023, 14(2), 395; https://doi.org/10.3390/atmos14020395
Submission received: 29 December 2022 / Revised: 9 February 2023 / Accepted: 15 February 2023 / Published: 17 February 2023
(This article belongs to the Special Issue Advances in Computational Wind Engineering and Wind Energy)

Abstract

:
Wind power has great potential in the fields of electricity generation, heating, et cetera, and the precise forecasting of wind speed has become the key task in an effort to improve the efficiency of wind energy development. Nowadays, many existing studies have investigated wind speed prediction, but they often simply preprocess raw data and also ignore the nonlinear features in the residual part, which should be given special treatment for more accurate forecasting. Meanwhile, the mainstream in this field is point prediction which cannot show the potential uncertainty of predicted values. Therefore, this paper develops a two-stage decomposition ensemble interval prediction model. The original wind speed series is firstly decomposed using a complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN) and the decomposed subseries with the highest approximate entropy is secondly decomposed through singular-spectrum analysis (SSA) to further reduce the complexity of the data. After two-stage decomposition, auto-encoder dimensionality reduction is employed to alleviate the accumulated error problem. Then, each reconstructed subsequence will generate an independent prediction result using an elastic neural network. Extreme gradient boosting (Xgboost) is utilized to integrate the separate predicted values and also carry out the error correction. Finally, the Gaussian process (GP) will generate the interval prediction result. The case study shows the best performance of the proposed models, not only in point prediction but also in interval prediction.

1. Introduction

Wind power is an important clean green energy, which has a great potential in the field of electricity generation. Therefore, the assessment of wind resources plays an important role in the effective design of wind turbines, wind farm design and the unit commitment of operational wind farms [1]. As is well-known, wind speed is the most influential factor in the efficient use of wind power, which means once the rules of wind speed are grasped, the efficiency of wind power generation will be greatly enhanced. As a result, wind power (or wind speed) prediction and trend modification have become a crucial research topic. However, due to the complexity and volatility of wind speed, accurate wind speed prediction has always been a tough task.
In recent years, multitudinous models have been proposed in the wind speed prediction field and a decomposition ensemble framework has been the frequently used model and has tended to be increasingly popular. There are two main parts in this model, the decomposition period and the prediction period.
The classic decomposition techniques are mainly categorized into wavelet transform (WT), empirical mode decomposition (EMD) and its extensions, variational mode decomposition (VMD), and singular spectrum analysis (SSA). WT is a basic time-frequency analysis method which is essentially equivalent to a multi-channel bandpass filter [2]. It can perform reasonably well when the frequency bands are clearly separated with noise [3,4]. However, it is a pity that WT has an unproductive adaptation and requires decomposition layers to be set manually. In the case of EMD, it was proposed by Huang et al. in 1998 [5] to improve capacity based on the local characteristics of the time series, such as the maximum, minimum and zero-crossings [6]. It can smooth complex signals, obtain subsequences with different features, and also be applied to various data samples [7,8,9,10]. In addition, the extensions of EMD such as ensemble empirical mode decomposition (EEMD) [11,12], complementary ensemble empirical mode decomposition (CEEMD) [13], and complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN) [14] have successively been developed to solve the problems of model mixing, lower computational efficiency, and residual noise. VMD, proposed by Dragomiretskiy, can concurrently extract the modes, and adaptively implement the frequency domain division of signals and the effective separation of intrinsic mode functions [15,16,17]. SSA is another commonly used decomposition method. It can obtain effective information in the sequence. In a variational mode decomposition–singular-spectrum analysis–autoregressive-integrated moving average (VMD–SSA–ARIMA) model, the overall results of all predictions obtained prove the stability of the prediction errors, which shows the good performance of SSA [18].
The prediction methods can be categorized into statistical approaches and machine learning models. Classic statistical models play an important role in linear and stationary time-series predictions. They are easy to understand and operate, but also they cannot handle complex and nonstationary data well. The representative statistical models are the autoregressive-integrated moving average model (ARIMA) [19,20], generalized autoregressive conditional heteroskedasticity model (GARCH) [21,22], and vector auto-regressive (VAR) model [23]. When data satisfy the statistical assumptions, the classic statistical methods can have a splendid performance. However, in most cases, the collected data are nonlinear and nonstationary, which cannot be well applied to traditional statistical methods.
Machine learning models have shown their superiority in extracting nonlinear features and nonlinear integration tasks. The commonly used machine learning methods are artificial neural networks (ANN) [24,25], support vector machine (SVM) and eXtreme gradient boosting (Xgboost). ANN can be subdivided into back propagation (BP) [26,27,28,29,30] and radial basis function (RBF) [31,32,33,34]. The common structure of a BP neural network includes three layers and each node has the function of calculating the inner product of the input and weight vector to provide an output. Similarly, RBF also has three layers, but it is worth stating that each neuron in the hidden layer of a RBF is centered at a particular point and the distance between the input vector and its own center is calculated and transformed using some basis function afterwards. However, using ANN poses a challenging task because the parameters needs to be manually set without a scientific mindset. Song et al. selected four types of ANN model as the individual forecasting models to construct a hybrid model and it turned out that the prediction ability was significantly enhanced when the parameters were suitable [35], which means traditional neural networks still have room for improvement.
SVM was developed to further increase the prediction ability as well as improve efficiency. SVM is a supervised machine learning model which was first developed by Vapnik [36]. It is based on statistical learning theory, and is widely utilized in predicting fields because of its great generalization ability, outstanding non-linear mapping ability, and small sample size [37]. An enhanced model of SVM, the least-square support vector machine (LS-SVM), was developed to alleviate the convex quadratic programming related to SVM and has also been applied in many studies [38,39]. However, SVM, as with ANN, needs quantities of parameters and the input can seriously influence the performance [40].
Extreme gradient boosting (Xgboost) is another frequently used prediction method which is effective, flexible, and movable. It can solve problems speedily and accurately as it offers parallel tree boosting. Compared with other methods, such as support vector regression (SVR) [41], Xgboost can prevent the overfitting problem through regularization items, thus achieving a higher accuracy [42,43,44]. It can also specify the default direction of the branch for missing values as well as specified values in an effort to improve the efficiency.
Overall, there is still room for improvement for abovementioned models. To start with, some of the current literature predicted wind speed directly through initial data, and it is difficult to find a proper model with low prediction errors. In addition, even though they used decomposition techniques, there are still the problems of pattern aliasing, unsteady signals in the first decomposition components, adding complexing, and the accumulation of estimation errors. For the residual component especially, it is always neglected by researchers, but it also contains useful and nonlinear information. Thirdly, the previous papers often focus on linear integration, which ignores nonlinear features. Finally, most of the existing literature only concentrates on point prediction which has a high uncertainty. In other words, point prediction cannot reflect the reliability of the predicted values.
To solve the abovementioned problems, this study proposes a two-stage decomposition multi-scale nonlinear ensemble model with a Gaussian process. The original data series will firstly go through CEEMDAN, and the decomposed subsequence with the highest approximate entropy (AE) value will be decomposed again through singular-spectrum analysis (SSA) to fully separate signals with different characteristics. After secondary decomposition, there are too many subsequences to predict which one brings low efficiency and an error accumulation problem, so an auto encoder is employed for the dimensionality reduction. Next, because the elastic network has a fast convergence rate and always produces effective solutions, it is utilized to predict each subseries and all the results will be integrated by extreme gradient boosting (Xgboost). If the integrated predictive results pass through the white noise test, that means there is no more useful residual information. Conversely, if the white noise test is not positive, the errors will be input into Xgboost again to obtain corrected predictive results. The last step is the use of the Gaussian process (GP) in order to gain interval prediction results. The innovations of this paper are as follows:
(1)
Many researchers used a single decomposition technique to obtain subsequences with different characteristics, which makes research easier because subsequences have more regular and simpler patterns than the original data series. However, there are still a few complex subsequences after once decomposition. Therefore, two-step decomposition is adopted to preprocess the original data and fully extract complicated features to the largest extent.
(2)
With the increase in the number of subsequences, the error accumulation problem is also increasingly apparent as each subsequence will be fed into a model separately and each model has its own errors. To alleviate the error accumulation problem, the high dimensional data are reduced to low dimensional ones by an auto encoder, which can not only save useful information in a data series but also simplify the modelling process.
(3)
The proposed model is based on the idea of “divide and conquer”, which can effectively handle distinct subseries through training different models. Instead of linear integration which is simply adding the results of subsequences, Xgboost is used to integrate the predictive results of subseries, which is an effective nonlinear ensemble process and shows superior performance.
(4)
In previous studies, the integrated results are seen as the final predictive results. However, in our proposed model there is an error correction strategy which is used for correcting the predictive values. Therefore, if the residual values include useful information, the proposed model still has the potential to dig them out.
(5)
The model proposed in this study does not stop at point prediction, but utilizes the Gaussian process to generate the interval prediction, which can show the potential uncertainty and the reliability of predictive results.
The remainder of this paper has four sections. The second section shows the detailed information on the related methods. The third section is the introduction of the proposed model. The case study, including the discussion, is in section four. The last section is the conclusion.

2. Related Methodology

There are eight related methods which are used in the proposed model. These methods are, respectively, CEEMDAN, AE, SSA, the white noise test, an auto encoder, an elastic neural network, Xgboost, and GP.

2.1. Complete Ensemble Empirical Mode Decomposition with Adaptive Noise

CEEMDAN is an extension of EMD and EEMD. E k ( ) is defined as the operator and its function is to solve the k t h modal component I M F k . Then,  w i  represents the white noise which uses the normal distribution of  N ( 0 , 1 ) . The white noise is added and  ε k  is its amplitude coefficient for the  K t h  time. Then, the whole process of CEEMDAN is expressed as follows:
The white noise is added at the  t t h  sample to the raw signals:
X ( t ) + ε 0 w i ( t )
The  I t h  EMD decomposition is performed and the average operation completed to obtain  I M F 1 :
I M F 1 = 1 I i = 1 1 E 1 ( X ( t ) + ε 0 w i ( t ) )
The first stage residual component is calculated:
r 1 ( t ) = X ( t ) I M F 1
White noise is added and EMD is performed in an effort to calculate  I M F 2  and the  K t h  residual component:
I M F 2 = 1 I i = 1 1 E 1 ( r 1 ( t ) + ε 1 E 1 ( w i ( t ) ) )
r k ( t ) = r k 1 ( t ) I M F k
Similarly to (2),  I M F k + 1  can be calculated through:
I M F k + 1 = 1 I i = 1 1 E 1 ( r k ( t ) + ε k E k ( w i ( t ) ) )
The last two processes are repeated until the value of residual component is less than two extremes. The final formula of the residual variable will be:
r ( t ) = X ( t ) k = 1 K I M F k
where  K  denotes mode number, and the signal is eventually:
X ( t ) = r ( t ) + k = 1 K I M F k

2.2. Approximate Entropy

Approximate entropy (AE) is an index which can quantify the regularity and unpredictability of a time series. It utilizes a non-negative number to indicate the complexity of a time series, which can reflect the occurrence of new information. The more complex a time series is, the larger the approximate entropy is.
Using  u ( 1 ) , u ( 2 ) , , u ( N )  represents an N-dimensional time series sampling in equal intervals. The parameter  m  represents the length of the comparison vector;  r  as a real number indicates a measure of similarity. Then, it is required to reconstruct the vector:  X ( 1 ) , X ( 2 ) , , X ( N m + 1 ) , where  X ( i ) = [ u ( i ) , u ( i + 1 ) , , u ( i + m 1 ) ] .
For  1 i N m + 1 , the number of vectors is counted if it meets:
C i m ( r ) = ( n u m b e r   o f   X ( j )   s u c h   t h a t   d [ X ( i ) , X ( j ) r ] ) N m + 1
where the symbols are:
d [ X , X ] = max a | u ( a ) u ( a ) | ,
u ( a ) : element of vector  X ,
d : distance between  X ( i )  and  X ( j )
X : adjoint matrix of  X .
Then, define:
Φ m ( r ) = i = 1 N m + 1 log ( C i m ( r ) ) N m + 1
So the AE value will be:
A p E n = Φ m ( r ) Φ m + 1 ( r )

2.3. Singular Spectrum Analysis

Singular spectrum analysis can decompose nonlinear time series data well. When the trajectory matrix of a time series is decomposed and reconstructed, distinct subseries (long-term components, seasonal component, noise, etc.) are extracted for the tasks of analysis or de-noising. Singular spectrum analysis mainly includes four steps: embedding, decomposition, grouping and reconstruction.
Step 1: For embedding, assume there is a time series which is one-dimensional and has a finite length:  [ x 1 , x 2 , , x N ]  ( N  is the number of samples). First, it is necessary to select the appropriate window length  L  to lag raw data and obtain the trajectory matrix:
X = [ x 1   x 2     x N L + 1 x 2   x 3     x N L + 2     x L   x L + 1         x N ]
usually L is smaller than half of N.
Then, K = NL + 1 is defined and the matrix is expressed as:
X = [ x 1   x 2             x K x 2   x 3           x K + 1     x L   x L + 1           x N ]
Step 2: For decomposition, a singular value decomposition is performed on the trajectory matrix. Note that here, there is a singular value decomposition (SVD) decomposition on the trajectory matrix.
X = U Σ V T
where  U  is called the left matrix;  Σ  only has a value on the main diagonal, which is the singular value, and the other elements are all zero;  V  is called the right matrix; and  V T  is the transpose matrix. In addition,  U  and  V  are unit orthogonal arrays.
Because it is a challenging task to directly decompose the trajectory matrix, first its covariance matrix S is calculated:
S = X X T
Next, an eigenvalue decomposition is performed on  S  to obtain eigenvalues:  λ 1 > λ 2 > > λ L 0  and the corresponding feature vector:  U 1 , U 2 , U L . Then,  [ U 1 , U 2 , , U L ]  and  λ 1 > λ 2 > > λ L 0  are the singular spectrum and we also have:
X = m = 1 L λ m U m V m T
V m = X T U m λ m
where  m = 1 , 2 , , L .
Step 3: For grouping, all  L  components are separated into  c  disjoint groups, which represents distinct components. In this way, some selected components are reconstructed to obtain a new sequence:
X = X l 1 + + X l c
where  X l = m l λ m U m V m T = X ( m l U m U m T ) .
Step 4: For reconstruction, first the projection of the hysteresis sequence  X i  is calculated onto  U m  (referred to as  a i m ):
a i m = X i U m = j = 1 L x i + j U m , j , 0 i N L
where  X i  represents column  i  of trajectory matrix  X a i m  is the weight of the time evolution reflected by  X i  in the  x i + 1 , x i + 2 , , x i + L  period of the original sequence.
Next, the time empirical orthogonal function and time principal components are used to complete the reconstruction process. The details of the reconstruction process are as follows:
x i k = { 1 i j = 1 i a i j k U k , j , 1 i L 1 1 L j = 1 L a i j k U k , j , L i N L + 1 1 N i + 1 j = i N + L L a i j K E k , j , N L + 2 i N
When all reconstructed sequences are added, a raw sequence should be obtained, which is:
x i = k = 1 L x i k i = 1 , 2 , , N

2.4. Auto Encoder

An auto encoder is an expansion of principal component analysis (PCA), which is also introduced here. PCA is mainly used for data preprocessing, reducing dimensions, and extracting key features (removing redundant features). For the original sample  x , it is encoded first through:
c = f ( x )
Then, another decoding function is constructed:
x = g ( f ( x ) ) ,   where   one   can   define   g ( x ) = D c
The loss function is minimized:
min c | x D c | 2
by reducing it to zero, producing the encoding function:
c = D T x
Furthermore, for a given  X , the covariance matrix will be:
C x = 1 n X X T
where  n  denotes the feature dimension.
The encoding function is applied and  Y = P X  is defined. Then, the covariance matrix of  Y  is constructed:
C y = 1 n Y Y T = P C X P T
Then, the next step is to calculate  P .
After making  Y = 1 n X T P  can be calculated through:
Y T Y = 1 n X X T = C x = P Σ T Σ V T
where  C x  is a symmetric matrix,  Σ  is the transition matrix.
An auto encoder has only one hidden layer and the original signal is reproduced as much as possible to obtain effective features. The main process is shown in Figure 1.

2.5. Elastic Neural Network

An elastic neural network is a linear regression model combined with two norms as a priori regular terms. It only needs a few non-zero sparse parameters, but it still retains some regular properties. Parameters can be used to control the convex combination of two norms. The theory is introduced in this section.
First, the expression of partial regression coefficients is defined:
θ = ( X T X ) 1 X T Y
Assuming there are  m  samples, each sample  i  contains  n  features:  x i 1 , x i 2 , , x i n  and an output  y i  as follows:
( x 1 ( 1 ) , x 2 ( 1 ) , , x n ( 1 ) , y 1 ) ( x 1 ( 2 ) , x 2 ( 2 ) , , x n ( 2 ) , y 2 ) ( x 1 ( m ) , x 2 ( m ) , , x n ( m ) , y m )
When a new  ( x 1 ( m + 1 ) , x 2 ( m + 1 ) , , x n ( m + 1 ) )  is given and the corresponding  y m + 1  is determined through the following model:
h θ ( x 1 , x 2 , , x n ) = θ 0 + θ 1 x 1 + + θ n x n = a d d   x 0 = 1 i = 0 n θ i x i
where the matrix expression is  h θ ( X ) = X θ , in which  h θ ( X )  is the  m × 1  vector and there are  n  model parameters of the algebraic method.  X  is the  m × n  matrix with  m  samples and  n  features of the sample.
Then, the loss function is expressed as follows:
J ( θ 0 , θ 1 , , θ n ) = i = 1 m ( h θ ( x 0 ( i ) , x 1 ( i ) , , x n ( i ) ) y i ) 2           = m a t r i x   f o r m 1 2 ( X θ Y ) T ( X θ Y )
where  Y  is the  m × 1  matrix representing the output matrix made up by  y i ( i = 1 , 2 , , n ) .
To apply a regularization norm,  L 1  is the regularization term for lasso regression and  L 2  is regularization term for ridge regression. Then, the loss function of elastic neural network will be:
J ( θ ) = 1 2 ( X θ Y ) T ( X θ Y ) + r α θ 1 + 1 r 2 α θ 2 2
where  α  and  r  denote constant coefficients, representing the regularization coefficient;  θ 1  is the norm of  L 1 θ 2  is the norm of  L 2 .
This kind of model is as sparse as pure lasso regression and possesses the same regularization ability as ridge regression.

2.6. White Noise Test

The white noise series has no correlation between the values and is therefore meaningless to investigate. Stationary series usually have a short-term correlation which can be described by the autocorrelation coefficient. With the increase in the number of delay periods, the autocorrelation coefficient of the stationary series rapidly decays to 0.
In particular, the formula for calculating the correlation coefficient for delay is as follows:
ρ = i = 1 n t ( x i υ ) ( x i + t υ ) i = 1 n ( x i υ ) 2
where  t  is the delay period,  n  is the number of the value and  υ  is the mean of all the values.
In a white noise series, the distribution of the autocorrelation coefficient will be:
ρ 0 ~ N ( 0 , 1 n )

2.7. Extreme Gradient Boosting

The extreme gradient boosting (Xgboost) algorithm is applied to finish the integration task. The main steps are as follows.
Step 1: The objective function of the boosting method is defined.
For each regression tree, the model can be expressed as:
f t ( x ) = w q x ,   w R T .   q : R d 1 , 2 T
where  w  represents the vector of the leaves and  q  means the structure of the tree while  T  is the number of leaf nodes in the tree. Then, the tree complexity can be expressed as:
Ω ( f t ) = γ T + 1 2 λ j = 1 T w j 2
where γ is the regular penalty of  L 1 , while λ is the regular penalty of  L 2 .
Step 2: A cycle is established.
Step 3: The gradient vector of the loss function of the  m t h  iteration is calculated.
As with the traditional boosting tree model, the lifting model of Xgboost also uses residuals (or negative gradient direction). The difference is that the selection of split nodes is not necessarily the least square loss. Here, the goal function can be expressed as:
O b j ( t ) i = 1 n [ l ( y i , y ^ i ( t 1 ) ) + g i f t ( x i ) + 1 2 h i f t 2 ( x i ) ] + Ω ( f t ) + C
Step 4: The Hessian matrix of the m-th iteration is calculated.
The objective function is only dependent on the first two derivatives of each data point on the error function. The reason for this is obvious. Since the previous objective function is only appropriate for the square loss function when seeking the optimal solution, it becomes very complicated for other loss functions. Through the transformation of the second-order Taylor expansion, the other solutions are solved in this way. Then, the objective function is transformed into the form of a leaf node in the  t t h  tree:
O b j ( t ) j = 1 T [ ( i I j g i ) w j + 1 2 ( i I j h i + λ ) w j 2 ] + γ T
where  I j = i | q ( x i = j )  means the sample  i  lands on the  j t h  leaf node and  w j  is the score of the  j t h  leaf node. When the partial derivative of  w j  is found and its derivative function is made equal to 0, we have:
w j = G j H j + λ
O b j = 1 2 j = 1 T G j 2 H j + λ + γ T
Step 5: The decision tree parameters for minimizing loss function are calculated.
Step 6: The weights of each base learner are combined to obtain the base learner model.
Additionally, this will be the loss function we need. Then, the estimate of the efficiency will be expressed in the form of the score:
G a i n = 1 2 [ G L 2 H L + λ + G R 2 H R + λ ( G L + G R ) 2 H L + H R ] γ
In theory, the general gain will be the mixture of the score of the left and right sub-tree minus the score if the split is not applied.
Step 7: The basic learner models obtained from the loop are all added.

2.8. Gaussian Process

The Gaussian process is usually described by a function. If two points are:  x 0  = 0,  x 1  = 1, then their function values will satisfy a Gaussian distribution:
( y 0 y 1 ) ~ N ( ( 0 0 ) , ( 1 0 0 1 ) )
where  y 0  and  y 1  are the function values.
Then, we have the kernel function:
k ( x , x ) = exp ( x x 2 / 2 σ 2 )
where  σ  is the standard deviation.
This is seen as the distance. If there are N data points, these function values will satisfy the N-dimensional Gaussian distribution with the mean of 0. Each element in it corresponds to (K means covariance matrix):
K n m = k ( x n , x m )
Then, the Gaussian process can be applied to learn any function. Assuming a given  x y  follows:
p ( y | x ) = N ( y | 0 , K )
where  K = k ( x , x ) .
In the introduced training set,  x # y #  will be the joint conditional distribution ( y  is the output):
( y # y ) ~ N ( ( 0 # 0 ) , ( K # K K T ( K , K ) ) )
Then, we have a conditional distribution ( x  is the output):
p ( y | x , x , y ) = N ( y | μ , E )
Here,  μ  and  E  can be calculated through:
μ = K T K 1 y
E = ( K , K ) K T K 1 K

3. Proposed Model and the Evaluation Criteria

This section gives the general procedure of the proposed model and the evaluation criteria used to assess the prediction ability.

3.1. Proposed Model

Figure 2 demonstrates the flowchart and the main parts are introduced as well:
(1)
CEEMDAN combined with AE is the first decomposition step where the raw data are decomposed in an effort to extract the features.
(2)
SSA decomposition is the secondary decomposition process where the focus is on the subsequence with the highest AE value. The first two steps are seen as the whole decomposition system which stresses the different features of the data series.
(3)
When the decomposition process is finished, the number of the subsequences will be too large which will bring a low efficiency in computation. Therefore, an auto encoder is used to realize the dimensionality reduction task to reduce the number of the subsequences.
(4)
Each subsequence will go through the elastic neural network to generate the sub-result, after obtaining these sub-results, Xgboost will generate the first prediction result based on sub-results. This process can be seen as a two-step nonlinear learning process.
(5)
Error correction is another innovation in the proposed model, which can correct the first predictive results and achieve a higher accuracy. Here, the white noise test is used to verify whether the predictive results should be corrected.
(6)
The last step is to generate the interval prediction. GP has the ability of interval generation which will be used in this paper, and this will bring a high fault tolerance and low uncertainty.

3.2. Evaluation Criteria

There are five measures selected to assess the predicting ability of the models (see Table 1). Three of them are used for evaluating the point prediction performance, which are mean absolute error (MAE), root mean squared error (RMSE), and mean absolute percentage error (MAPE), respectively. Because the three indexes quantify the number of errors, the smaller is the better. The rest is coverage probability (CP) and mean width percentage (MWP), which are mainly used to evaluate the performance of interval prediction. The larger the CP is, the better the interval prediction is. The MWP value cannot be too large or small, otherwise there will not be a suitable interval.
In Table 1 n 1  is the sample number,  w ( i )  is the real wind speed at time point  i  and  w ^ ( i )  is the prediction speed,  λ 1 μ  is the number of the actual speed falling in the gap at the confidence of  1 μ U ( x i )  and  L ( x i )  are the upper/lower bounds;  r i  is the  i t h  real value.

4. Case Study

4.1. Data Collection

This study applied data from four places which are Dingxin, Yongchang, Jingtai, and Wushan. They are, respectively, referred to case 1, case 2, case 3, and case 4. The four places are all located in Gansu, China. The data were sourced from the website of China Meteorological Data Service Center, http://data.cma.cn/ (accessed on 12 April 2022). The basic information of datasets is shown in Table 2. It is apparently shown that the empirical distribution of the studied datasets is far away from normal distribution based on the results of a normal test. The standard deviation represents the fluctuation degree and it can be seen that the volatility in the four datasets was very close. The daily average wind speed was selected and 75% data were used as the training set and the rest was used as the test set (the ratio is 3:1).

4.2. Point Prediction

4.2.1. Case I: Wind Speed from Dingxin

The data series was firstly divided into subsequences using CEEMDAN. Then, for each subsequence, the AE value was calculated, as is shown in Figure 3.
It is obvious that s1 and s2 are apparently higher than others (which are larger than 2), so they were decomposed again, separately, through SSA. After the two-stage decomposition, the original data series was decomposed into 27 subseries. The subsequences were reconstructed through the auto encoder and the reconstructed sequences are shown in Figure 4.
Once the decomposition and dimension reconstruction were finished, the integration process was carried out. Firstly, an elastic neural network was employed in each subsequence to obtain the sub-result and then Xgboost was used to integrate all the sub-results into one prediction result. Subsequently, the white noise test was conducted on the residual series to determine whether to use an error correction strategy. The Q-values are shown in Figure 5 and clearly the residual series was proven not to be a white noise series.
For the residual correction part, Xgboost was applied again on the residual series. The process is described in Figure 6. To further demonstrate the necessity of residual correction, Figure 6 shows the comparison of the first residual series from Xgboost and the final residual series from error correction. It can be seen in the figure that the residual series after error correction was closer to zero than the previous one, which means the residual correction strategy can effectively improve accuracy.
The numerical results are shown in the next section. Table 3 is the related parameters.

4.2.2. Case II: Test with Wind Speed from Wushan

The second dataset was from Wushan, China. The data series was firstly divided into subsequences using CEEMDAN. Then, for each subsequence, the AE value was calculated, as is shown in Figure 7.
The highest value was over 3, which was then decomposed again, separately, through SSA. When the process above was completed, the auto encoder was applied to carry out the dimension reconstruction, which is shown in Figure 8.
When the decomposition and dimension reconstruction were finished, the integration process was carried out. Similarly, the Q-values are shown in Figure 9 and they proved the necessity to carry out residual correction.
The integration process and the residual comparison are shown in Figure 10.
The measures were also calculated and listed as well. Table 4 is the related parameters.

4.2.3. Case III: Test with Wind Speed from Yongchang

In case 3, the samples were from Yongchang, China. The original data series was firstly divided into subsequences using CEEMDAN. Then, for each subsequence, the AE value was calculated, as is shown in Figure 11.
The highest value was over 5, which was decomposed again, separately, through SSA. When the subsequences were prepared, the auto encoder was applied to carry out the dimension reconstruction work, which is shown in Figure 12.
When the decomposition and dimension reconstruction were finished, the integration process was carried out. The result of the test is shown in Figure 13.
After the residual correction part was completed, the final point prediction was generated and the process above is described in Figure 13. The detailed point prediction results are shown in Figure 14.
Table 5 is the related parameters.

4.2.4. Case IV: Test with Wind Speed from Jingtai

For the case of Jingtai, China, the original data series was firstly divided into subsequences using CEEMDAN. Then, for each subsequence, the AE value was calculated, as is shown in Figure 15.
The highest value was over 2.5, which was decomposed again, separately, through SSA. Then, the auto encoder was applied to reconstruct the dimension and the output subsequences’ number was 5, which is shown in Figure 16.
After integration, the white noise test showed the residual series was not noisy. The values are shown in Figure 17.
When the residual correction part was finished, the final point prediction result was generated. The process above is described in Figure 18.
The parameters are shown in Table 6.

4.3. Interval Prediction

GP was used as the interval forecast tool. Figure 19 displays the GP results of the four datasets as well as some related measures. The parameters are displayed in Table 7. The measures of the four cases are shown in Table 8.

4.4. Analysis and Discussion

In order to show the rationality of model construction, we designed the following comparative experiments: (1) The first model used a one-step decomposition which is different to our secondary decomposition model; (2) the second model discarded the whole decomposition process and directly built the model on the original dataset; (3) the third model contained the whole decomposition process, but the integration process only contained the elastic neural network; (4) the fourth model used linear accumulation to replace nonlinear integration; (5) the fifth model was designed to show the significance of residual correction and it was the proposed model without the residual correction process.
Based on the above comparative experiments, we can obtain the following conclusions:
(1)
The proposed model applies a secondary decomposition process which contains CEEMDAN and SSA while some existing studies only use one-step decomposition. For this reason, the first cases applied a one-step decomposition. The next steps remained the same as the proposed model. The RMSE and MAE values showed obvious increases with the secondary decomposition. More specifically, the RMSE values of the four cases were, respectively, 8 times, 6 times, 6.5 times and 13 times larger than that of the proposed model; the MAE values of the comparative model were also several times larger than the proposed one. The increase in the MAPE values was also apparent in the four cases and the value increased by 45.49%, 32.03%, 31.58%, and 40.62%, respectively. Then, for the interval prediction, the CP values of the first two cases decreased by about 0.2, while the MWP values increased by about 0.2 and 0.03. It can be seen that in the third and fourth cases, the CP values showed a slight increase while the MWP values showed a multifold increase. The experiment demonstrates that the one-step decomposition will neglect some information in the series which means the feature extraction still has room for improvement. Moreover, it proves that the two-step decomposition is necessary.
(2)
There are a large number of studies investigating the study process on the original dataset, so in the second model, the whole decomposition process was deleted entirely as a comparative group. The RMSE values of the second comparative model were dozens of times larger than the proposed model as well as the MAE values in all cases. Similarly, the MAPE values were several times larger than the proposed model’s. In addition, the CP value of the first case showed an apparent decrease by 0.334, and the MWP value increased to 2.4638. For the next three cases, the CP values showed no great difference, but the MWP values all became far from satisfactory. Obviously, using the original dataset directly means that all the features are given the same weight, which can greatly affect the prediction ability in practice.
(3)
The third model turned the original two-step learning process to a one-step integration. For certain, the process became easier to operate, but the accuracy and fault tolerance were both influenced. The most obvious changes were the increase in the MAPE value in case 1 by 33.66% and the increase in the MWP value in case 3, which was multiplied 9.96 times. The simplified process brought not only the decline in the point prediction accuracy, but also a negative effect on the certainty. Sometimes, a simple one-step integration does offer a good performance, but it also causes risks that the features are not fully captured and the learning process is insufficient. Thus, in an effort to complete the learning process and pay as much attention as possible to all the features, it is highly recommended to apply a two-step learning process.
(4)
The proposed model applies nonlinear integration as an important process in the point prediction period. In some classic methods, linear accumulation is always utilized, so the fourth model replaced the Xgboost integration with linear accumulation. It was learned from the results that the accuracy greatly decreased. The RMSE values were much larger than that in the proposed model in the four cases. The MAE values increased by the similar multiples accordingly. The MAPE values showed a distinct increase as well and the values of the four cases increased by 129.99%, 88.50%, 72.54% and 55.81%, respectively. Meanwhile, in the interval prediction process, differences in the CP values were not really evident, but when it came to MWP values, the values increased by 7.6663, 2.0659, 4.2418 and 4.5723, which means the gap between the boundaries widened and this will bring large uncertainty. Linear integration takes the least time to operate and it is the easiest way to generate results, but doing this means the ignorance of nonlinear features, and, as a result, the prediction accuracy is far from satisfactory.
(5)
Finally, as the residual correction strategy is another notable part in the proposed model, to further demonstrate the significance of this, the fifth model eliminated the residual correction process. Drawn from the table, the RMSE values of the four cases increased by 0.06, 0.04, 0.02, and 0.01, respectively, together with an increase in the MAE values in the first three cases, by 0.05, 0.04, and 0.02, respectively. While for the interval prediction, the CP values showed a great decline, and the gap narrowed further, with decreases of 1165.50 times, 535.38 times, 236.50 times, and 127.47 times in four cases. It has been said that MWP values cannot be too large or too small, and in this experiment, the MWP with too small values will turn the interval prediction to the results that are quite alike with the point prediction. In other words, the residual correction must not be neglected. The residual series is always neglected by researchers while, as a matter of fact, some features in the residual series remain unnoticed. The application of the residual correction strategy means the capturing of all the possible information and, by doing this, the performance will be the most precise.
To sum up, the proposed model shows improvement in many aspects which are often not given much attention. In this model, the data is decomposed fully to separate the features and a two-step learning process captures the nonlinear features to a large extent. In addition, the residual series is given the correct attention. According to the results of the experiments above, these techniques are noteworthy and effective.

5. Conclusions

The accurate prediction of wind speed is of great significance. This study proposed a two-stage decomposition multi-scale nonlinear ensemble model and this model contains a secondary decomposition feature reconstruction process and a two-stage nonlinear process including a residual correction strategy. The proposed model is capable of providing point and interval prediction results. Both the case studies and the comparison models show that the proposed model can improve not only the accuracy of the point prediction but the performance of interval prediction.
The proposed model combines various methods effectively and reasonably and it can be applied in other fields of data prediction and trend modification. In addition, the proposed model also has room for improvement. For example, some different integration and decomposition methods can be applied as well.

Author Contributions

J.W., M.H. and S.Q. conceived the presented idea, developed the theory, performed the computations, discussed the results, wrote the paper, and approved the final manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (Grant Nos. 71971122 and 71501101).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data is available at the website of China Meteorological Data Service Center, http://data.cma.cn/ (accessed on 12 April 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chinmoy, L.; Iniyan, S.; Goic, R. Modeling wind power investments, policies and social benefits for deregulated electricity market—A review. Appl. Energy 2019, 242, 364–377. [Google Scholar] [CrossRef]
  2. Zhang, Y.G.; Chen, B.; Pan, G.F.; Zhao, Y. A novel hybrid model based on VMD-WT and PCA-BP-RBF neural network for short-term wind speed forecasting. Energy Convers. Manag. 2019, 195, 180–197. [Google Scholar] [CrossRef]
  3. Chen, C.; Liu, H. Dynamic ensemble wind speed prediction model based on hybrid deep reinforcement learning. Adv. Eng. Inform. 2021, 48, 101290. [Google Scholar] [CrossRef]
  4. Liu, H.; Yu, C.M.; Yu, C.Q.; Chen, C.; Wu, H.P. A novel axle temperature forecasting method based on decomposition, reinforcement learning optimization and neural network. Adv. Eng. Inform. 2020, 44, 101089. [Google Scholar] [CrossRef]
  5. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. Lond. Ser. A Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  6. Ren, Y.; Suganthan, P.N.; Srikanth, N. A Comparative Study of Empirical Mode Decomposition-Based Short-Term Wind Speed Forecasting Methods. IEEE Trans. Sustain. Energy 2017, 6, 236–244. [Google Scholar] [CrossRef]
  7. Zhou, J.G.; Xu, X.L.; Huo, X.J.; Li, Y.S. Forecasting Models for Wind Power Using Extreme-Point Symmetric Mode Decomposition and Artificial Neural Networks. Sustainability 2019, 11, 650. [Google Scholar] [CrossRef] [Green Version]
  8. Fei, S.W. A hybrid model of EMD and multiple-kernel RVR algorithm for wind speed prediction. Int. J. Electr. Power Energy Syst. 2016, 78, 910–915. [Google Scholar] [CrossRef]
  9. Ye, R.; Suganthan, P.N.; Srikanth, N. A Novel Empirical Mode Decomposition With Support Vector Regression for Wind Speed Forecasting. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 1793–1798. [Google Scholar] [CrossRef]
  10. Qolipour, M.; Mostafaeipour, A.; Mohammad, S.M.; Arabnia, H.R. Prediction of wind speed using a new Grey-extreme learning machine hybrid algorithm: A case study. Energy Environ. 2019, 30, 44–62. [Google Scholar] [CrossRef]
  11. Cheng, Z.S.; Wang, J.Y. A new combined model based on multi-objective salp swarm optimization for wind speed forecasting. Appl. Soft Comput. 2020, 92, 106294. [Google Scholar] [CrossRef]
  12. Huang, Y.S.; Liu, S.J.; Yang, L. Wind Speed Forecasting Method Using EEMD and the Combination Forecasting Method Based on GPR and LSTM. Sustainability 2018, 10, 3693. [Google Scholar] [CrossRef] [Green Version]
  13. Yang, Z.S.; Wang, J. A combination forecasting approach applied in multistep wind speed forecasting based on a data processing strategy and an optimized artificial intelligence algorithm. Appl. Energy 2018, 230, 1108–1125. [Google Scholar] [CrossRef]
  14. Liu, H.; Mi, X.W.; Li, Y.F. Comparison of two new intelligent wind speed forecasting approaches based on Wavelet Packet Decomposition, Complete Ensemble Empirical Mode Decomposition with Adaptive Noise and Artificial Neural Networks. Energy Convers. Manag. 2018, 155, 188–200. [Google Scholar] [CrossRef]
  15. Dragomiretskiy, K.; Zosso, D. Variational Mode Decomposition. IEEE Trans Signal Process. 2014, 62, 531–544. [Google Scholar] [CrossRef]
  16. Fu, W.L.; Wang, K.; Zhou, J.Z.; Xu, Y.H.; Tan, J.W.; Chen, T. A Hybrid Approach for Multi-Step Wind Speed Forecasting Based on Multi-Scale Dominant Ingredient Chaotic Analysis, KELM and Synchronous Optimization Strategy. Sustainability 2019, 11, 1804. [Google Scholar] [CrossRef] [Green Version]
  17. Wu, Q.L.; Lin, H.X. Short-Term Wind Speed Forecasting Based on Hybrid Variational Mode Decomposition and Least Squares Support Vector Machine Optimized by Bat Algorithm Model. Sustainability 2019, 11, 652. [Google Scholar] [CrossRef] [Green Version]
  18. Moreno, S.R.; Mariani, V.C.; Coelho, L.D. Hybrid multi-stage decomposition with parametric model applied to wind speed forecasting in Brazilian Northeast. Renew. Energy 2021, 164, 1508–1526. [Google Scholar] [CrossRef]
  19. Shukur, B.O.; Lee, M.H. Daily wind speed forecasting through hybrid KF-ANN model based on ARIMA. Renew. Energy 2015, 76, 637–647. [Google Scholar] [CrossRef]
  20. Aasim, S.S.; Mohapatra, A. Repeated wavelet transform based ARIMA model for very short-term wind speed forecasting. Renew. Energy 2019, 136, 758–768. [Google Scholar] [CrossRef]
  21. Sharma, S.K.; Ghosh, S. Short-term wind speed forecasting: Application of linear and non-linear time series models. Int. J. Green Energy 2016, 13, 1490–1500. [Google Scholar] [CrossRef]
  22. Li, L.X.; Miao, S.H.; Tu, Q.Y.; Duan, S.M.; Li, Y.W.; Han, J. Dynamic dependence modelling of wind power uncertainty considering heteroscedastic effect. Int. J. Electr. Power Energy Syst. 2020, 116, 105556. [Google Scholar] [CrossRef]
  23. Lucheroni, C.; Boland, J.; Ragno, C. Scenario generation and probabilistic forecasting analysis of spatio-temporal wind speed series with multivariate autoregressive volatility models. Appl. Energy 2019, 239, 1226–1241. [Google Scholar] [CrossRef]
  24. Baomar, H.; Bentley, P.J. Autonomous flight cycles and extreme landings of airliners beyond the current limits and capabilities using artificial neural networks. Appl. Intell. 2021, 51, 6349–6375. [Google Scholar] [CrossRef]
  25. Luo, X.J.; Oyedele, L.O.; Ajayi, A.O.; Monyei, C.G.; Akinade, O.O.; Akanbi, L.A. Development of an IoT-based big data platform for day-ahead prediction of building heating and cooling demands. Adv. Eng. Inform. 2019, 41, 100926. [Google Scholar] [CrossRef]
  26. Zhang, Y.G.; Pan, G.F.; Zhang, C.H.; Zhao, Y. Wind speed prediction research with EMD-BP based on Lorenz disturbance. J. Electr. Eng. 2019, 70, 198–207. [Google Scholar] [CrossRef] [Green Version]
  27. Ren, C.; An, N.; Wang, J.Z.; Li, L.; Hu, B.; Shang, D. Optimal parameters selection for BP neural network based on particle swarm optimization: A case study of wind speed forecasting. Knowl.-Based Syst. 2014, 56, 226–239. [Google Scholar] [CrossRef]
  28. Liu, H.; Tian, H.Q.; Li, Y.F.; Zhang, L. Comparison of four Adaboost algorithm based artificial neural networks in wind speed predictions. Energy Convers. Manag. 2015, 92, 67–81. [Google Scholar] [CrossRef]
  29. Wang, S.X.; Zhang, N.; Wu, L.; Wang, Y.M. Wind speed forecasting based on the hybrid ensemble empirical mode decomposition and GA-BP neural network method. Renew. Energy 2016, 94, 629–636. [Google Scholar] [CrossRef]
  30. Jiang, P.; Li, R.R.; Zhang, K.Q. Two combined forecasting models based on singular spectrum analysis and intelligent optimized algorithm for short-term windspeed. Neural Comput. Appl. 2016, 30, 1–19. [Google Scholar] [CrossRef]
  31. Zhang, C.; Wei, H.K.; Xie, L.P.; Shen, Y.; Zhang, K.J. Direct interval forecasting of wind speed using radial basis function neural networks in a multi-objective optimization framework. Neurocomputing 2016, 205, 53–63. [Google Scholar] [CrossRef]
  32. Dalibor, P.; Shahaboddin, S.; Nor, B.A.; Hadi, S.; Ainuddin, W.A.W.; Milan, P.; Erfan, Z.; Seyed, M.A.M. An appraisal of wind speed distribution prediction by soft computing methodologies: A comparative study. Energy Convers. Manag. 2014, 84, 133–139. [Google Scholar] [CrossRef]
  33. Zhang, Y.G.; Zhang, C.H.; Sun, J.B.; Guo, J.J. Improved Wind Speed Prediction Using Empirical Mode Decomposition. Adv. Electr. Comput. Eng. 2018, 18, 3–10. [Google Scholar] [CrossRef]
  34. Wu, Z.Q.; Jia, W.J.; Zhao, L.R.; Wu, C.H. Maximum wind power tracking based on cloud RBF neural network. Renew. Energy 2016, 86, 466–472. [Google Scholar] [CrossRef]
  35. Song, J.J.; Wang, J.Z.; Lu, H.Y. A novel combined model based on advanced optimization algorithm for short-term wind speed forecasting. Appl. Energy 2018, 215, 643–658. [Google Scholar] [CrossRef]
  36. Thoranin, S. A new class of MODWT-SVM-DE hybrid model emphasizing on simplification structure in data pre-processing: A case study of annual electricity consumptions. Appl. Soft Comput. 2017, 54, 150–163. [Google Scholar] [CrossRef]
  37. Fu, C.; Li, G.Q.; Lin, K.P.; Zhang, H.J. Short-Term Wind Power Prediction Based on Improved Chicken Algorithm Optimization Support Vector Machine. Sustainability 2019, 11, 512. [Google Scholar] [CrossRef] [Green Version]
  38. Zhao, H.R.; Zhao, H.R.; Guo, S. Short-Term Wind Electric Power Forecasting Using a Novel Multi-Stage Intelligent Algorithm. Sustainability 2018, 10, 881. [Google Scholar] [CrossRef] [Green Version]
  39. Yang, J.X. A novel short-term multi-input-multi-output prediction model of wind speed and wind power with LSSVM based on improved ant colony algorithm optimization. Cluster Comput. 2019, 22, S3293–S3300. [Google Scholar] [CrossRef]
  40. Sun, W.; Liu, M.H.; Liang, Y. Wind Speed Forecasting Based on FEEMD and LSSVM Optimized by the Bat Algorithm. Energies 2015, 8, 6585–6607. [Google Scholar] [CrossRef] [Green Version]
  41. Zhang, X.; Yan, M.; Xie, B.L.; Yang, H.Q.; Ma, H. An automatic real-time bus schedule redesign method based on bus arrival time prediction. Adv. Eng. Inform. 2021, 48, 101295. [Google Scholar] [CrossRef]
  42. Zheng, H.; Wu, Y.H. A XGBoost Model with Weather Similarity Analysis and Feature Engineering for Short-Term Wind Power Forecasting. Appl. Sci. 2019, 9, 3019. [Google Scholar] [CrossRef] [Green Version]
  43. Cai, R.; Xie, S.; Wang, B.Z.; Yang, R.J.; Xu, D.S.; He, Y. Wind Speed Forecasting Based on Extreme Gradient Boosting. IEEE Access. 2020, 8, 175063–175069. [Google Scholar] [CrossRef]
  44. Chakraborty, D.; Elhegazy, H.; Elzarka, H.; Gutierrez, L. A novel construction cost prediction model using hybrid natural and light gradient boosting. Adv. Eng. Inform. 2020, 46, 101201. [Google Scholar] [CrossRef]
Figure 1. Auto encoder process.
Figure 1. Auto encoder process.
Atmosphere 14 00395 g001
Figure 2. Flowchart of the proposed model.
Figure 2. Flowchart of the proposed model.
Atmosphere 14 00395 g002
Figure 3. AE calculation.
Figure 3. AE calculation.
Atmosphere 14 00395 g003
Figure 4. The six subfigures represent six subsequences which are generated through decomposition and dimension reduction from the original time series in case 1.
Figure 4. The six subfigures represent six subsequences which are generated through decomposition and dimension reduction from the original time series in case 1.
Atmosphere 14 00395 g004
Figure 5. Box_Ljung value of each lag order.
Figure 5. Box_Ljung value of each lag order.
Atmosphere 14 00395 g005
Figure 6. The first six subfigures are the point prediction results of subseries. The seventh subfigure titled “Xgboost integration” shows the integrated results. The eighth subfigure represents the final point prediction results which are generated based on the Xgboost integration and error correction strategy. The last subfigure is the residual comparison between the model with error correction strategy and the model without it. All the experimental results here used data in case 1.
Figure 6. The first six subfigures are the point prediction results of subseries. The seventh subfigure titled “Xgboost integration” shows the integrated results. The eighth subfigure represents the final point prediction results which are generated based on the Xgboost integration and error correction strategy. The last subfigure is the residual comparison between the model with error correction strategy and the model without it. All the experimental results here used data in case 1.
Atmosphere 14 00395 g006
Figure 7. AE calculation.
Figure 7. AE calculation.
Atmosphere 14 00395 g007
Figure 8. The five subfigures represent five subsequences which are generated through decomposition and dimension reduction from the original time series in case 2.
Figure 8. The five subfigures represent five subsequences which are generated through decomposition and dimension reduction from the original time series in case 2.
Atmosphere 14 00395 g008
Figure 9. Box_Ljung value of each lag order.
Figure 9. Box_Ljung value of each lag order.
Atmosphere 14 00395 g009
Figure 10. The first five subfigures are the point prediction results of subseries. The sixth subfigure titled “Xgboost integration” shows the integrated results. The seventh subfigure represents the final point prediction results which are generated based on the Xgboost integration and error correction strategy. The last subfigure is the residual comparison between the model with error correction strategy and the model without it. All the experimental results here used data in case 2.
Figure 10. The first five subfigures are the point prediction results of subseries. The sixth subfigure titled “Xgboost integration” shows the integrated results. The seventh subfigure represents the final point prediction results which are generated based on the Xgboost integration and error correction strategy. The last subfigure is the residual comparison between the model with error correction strategy and the model without it. All the experimental results here used data in case 2.
Atmosphere 14 00395 g010
Figure 11. CEEMDAN and AE calculation.
Figure 11. CEEMDAN and AE calculation.
Atmosphere 14 00395 g011
Figure 12. The three subfigures represent three subsequences which are generated through decomposition and dimension reduction from the original time series in case 3.
Figure 12. The three subfigures represent three subsequences which are generated through decomposition and dimension reduction from the original time series in case 3.
Atmosphere 14 00395 g012
Figure 13. Box_Ljung value of each lag order.
Figure 13. Box_Ljung value of each lag order.
Atmosphere 14 00395 g013
Figure 14. The first three subfigures are the point prediction results of subseries. The fourth subfigure titled “Xgboost integration” shows the integrated results. The fifth subfigure represents the final point prediction results which are generated based on the Xgboost integration and error correction strategy. The last subfigure is the residual comparison between the model with error correction strategy and the model without it. All the experimental results here used data in case 3.
Figure 14. The first three subfigures are the point prediction results of subseries. The fourth subfigure titled “Xgboost integration” shows the integrated results. The fifth subfigure represents the final point prediction results which are generated based on the Xgboost integration and error correction strategy. The last subfigure is the residual comparison between the model with error correction strategy and the model without it. All the experimental results here used data in case 3.
Atmosphere 14 00395 g014
Figure 15. CEEMDAN and AE calculation.
Figure 15. CEEMDAN and AE calculation.
Atmosphere 14 00395 g015
Figure 16. The five subfigures represent five subsequences which are generated through decomposition and dimension reduction from the original time series in case 4.
Figure 16. The five subfigures represent five subsequences which are generated through decomposition and dimension reduction from the original time series in case 4.
Atmosphere 14 00395 g016
Figure 17. Box_Ljung value of each lag order.
Figure 17. Box_Ljung value of each lag order.
Atmosphere 14 00395 g017
Figure 18. The first five subfigures are the point prediction results of subseries. The sixth subfigure titled “Xgboost integration” shows the integrated results. The seventh subfigure represents the final point prediction results which are generated based on the Xgboost integration and error correction strategy. The last subfigure is the residual comparison between the model with error correction strategy and the model without it. All the experimental results here used data in case 4.
Figure 18. The first five subfigures are the point prediction results of subseries. The sixth subfigure titled “Xgboost integration” shows the integrated results. The seventh subfigure represents the final point prediction results which are generated based on the Xgboost integration and error correction strategy. The last subfigure is the residual comparison between the model with error correction strategy and the model without it. All the experimental results here used data in case 4.
Atmosphere 14 00395 g018
Figure 19. The interval prediction estimate of four cases.
Figure 19. The interval prediction estimate of four cases.
Atmosphere 14 00395 g019
Table 1. The formula of evaluation criteria.
Table 1. The formula of evaluation criteria.
NameEquation
MAE   1 n 1 i = 1 n 1 | w ( i ) w ^ ( i ) |
RMSE   1 n 1 i = 1 n 1 ( w ( i ) w ^ ( i ) ) 2
MAPE   1 n 1 i = 1 n 1 | w ( i ) w ^ ( i ) w ( i ) | × 100 %
MWP   1 n 1 i = 1 n 1 U ( x i ) L ( x i ) r i
CP   λ 1 μ n 1 × 100 %
Table 2. The basic information of datasets.
Table 2. The basic information of datasets.
DatasetDingxinWushanYongchangJingtai
Starting date1 January 19551 January 19691 January 19601 October 1956
Ending date31 December 201631 December 201631 December 201631 December 2016
Samples22,64517,53120,81921,975
Mean30.79029.74029.15725.677
Min0000
Max143103139127
Std.16.39812.35813.97015.082
Normal test8136.282528.8115,631.9712,534.76
Std. means standard deviation; the normal test is Jarque-Bera test.
Table 3. Parameters used in case 1.
Table 3. Parameters used in case 1.
Sub 1Sub 2Sub 3Sub 4Sub 5Sub 6
alpha0.0010.0010.0010.0010.0010.001
ratio0.0010.0010.0010.8990.0010.001
Xgboost-integrationXgboost-residual correction
max_depthsubsamplecolsample_bytreemax_depthsubsamplecolsample_bytree
50.80.7700.750.6
Table 4. Parameters used in case 2.
Table 4. Parameters used in case 2.
Sub 1Sub 2Sub 3Sub 4Sub 5
alpha0.010.010.010.010.01
ratio0.61610.010.011
Xgboost-integrationXgboost-residual correction
max_depthsubsamplecolsample_bytreemax_depthsubsamplecolsample_bytree
30.70.651050.70.65
Table 5. Parameters used in case 3.
Table 5. Parameters used in case 3.
Sub 1Sub 2Sub 3
alpha0.010.010.01
ratio0.1920.011
Xgboost-integrationXgboost-residual correction
max_depthsubsamplecolsample_bytreemax_depthsubsamplecolsample_bytree
120.750.851000.50.8
Table 6. Parameters used in case 4.
Table 6. Parameters used in case 4.
Sub 1Sub 2Sub 3Sub 4Sub 5
alpha0.010.010.010.010.04
ratio0.0110.010.010.01
Xgboost-integrationXgboost-residual correction
max_depthsubsamplecolsample_bytreemax_depthsubsamplecolsample_bytree
130.550.71100.60.5
Table 7. Parameters of cases used in GP.
Table 7. Parameters of cases used in GP.
Meancovlik
Case 11−1,1,−75
Case 21−5,6,−13
Case 31−1,0,−75
Case 411,−3,−23
Table 8. Measures of models.
Table 8. Measures of models.
ModelCaseMeasure
RMSEMAEMAPECPMWP
Proposed modelCase 10.020.015.74%0.99931.3986
Case 20.020.015.08%0.98810.6960
Case 30.020.015.96%0.99920.3784
Case 40.010.016.25%0.94980.2422
CEEMDAN-elastic neural network
-two-step Xgboost
Case 10.160.1251.23%0.79281.2404
Case 20.120.0937.11%0.78500.7338
Case 30.130.0937.54%15.0851
Case 40.130.0946.87%13.6869
Elastic neural network-
two-step Xgboost
Case 10.150.1149.88%0.66532.4638
Case 20.120.0935.93%12.2888
Case 30.120.0834.47%0.95904.4455
Case 40.110.0841.74%12.8217
CEEMDAN-AE-
SSA-AE-elastic neural network
Case 10.120.0939.40%0.96102.9545
Case 20.080.0729.03%0.94001.6599
Case 30.070.0623.67%0.97603.7672
Case 40.080.0929.03%12.9151
CEEMDAN-AE-
SSA-AE-elastic neural network-
linear accumulation
Case 10.270.23135.73%19.0649
Case 20.210.1893.58%0.94702.7619
Case 30.250.2378.50%0.95604.5932
Case 40.190.1662.06%0.99704.8145
CEEMDAN-AE-
SSA-AE-elastic neural network-
Xgboost integration
Case 10.080.0624.28%0.67970.0012
Case 20.060.0518.99%0.75080.0013
Case 30.040.0312.56%0.28100.0016
Case 40.020.018.12%0.15780.0019
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.; He, M.; Qiu, S. Two-Stage Decomposition Multi-Scale Nonlinear Ensemble Model with Error-Correction-Coupled Gaussian Process for Wind Speed Forecast. Atmosphere 2023, 14, 395. https://doi.org/10.3390/atmos14020395

AMA Style

Wang J, He M, Qiu S. Two-Stage Decomposition Multi-Scale Nonlinear Ensemble Model with Error-Correction-Coupled Gaussian Process for Wind Speed Forecast. Atmosphere. 2023; 14(2):395. https://doi.org/10.3390/atmos14020395

Chicago/Turabian Style

Wang, Jujie, Maolin He, and Shiyao Qiu. 2023. "Two-Stage Decomposition Multi-Scale Nonlinear Ensemble Model with Error-Correction-Coupled Gaussian Process for Wind Speed Forecast" Atmosphere 14, no. 2: 395. https://doi.org/10.3390/atmos14020395

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop