Next Article in Journal
Wind Forces and Flow Patterns of Three Tandem Prisms with a Small Height–Width Ratio
Previous Article in Journal
Research on the Flow and Transmission Performance of Magnetorheological Fluid between Two Discs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SFINet: Shuffle–and–Fusion Interaction Networks for Wind Power Forecasting

1
Department of Technical Development, AI Forward Technology Co., Ltd., Beijing 100801, China
2
School of Electronic and Control Engineering, North China Institute of Aerospace Engineering, Langfang 065000, China
3
Faculty of Engineering and Information Sciences, University of Wollongong, Wollongong, NSW 2522, Australia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(4), 2253; https://doi.org/10.3390/app12042253
Submission received: 29 December 2021 / Revised: 11 February 2022 / Accepted: 17 February 2022 / Published: 21 February 2022
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
Wind energy is one of the most important renewable energy sources in the world. Accurate wind power prediction is of great significance for achieving reliable and economical power system operation and control. For this purpose, this paper is focused on wind power prediction based on a newly proposed shuffle–and–fusion interaction network (SFINet). First, a channel shuffle is employed to promote the interaction between timing features. Second, an attention block is proposed to fuse the original features and shuffled features to further increase the model’s sequential modeling capability. Finally, the developed shuffle–and–fusion interaction network model is tested using real-world wind power production data. Based on the results verified, it was proven that the proposed SFINet model can achieve better performance than other baseline methods, and it can be easily implemented in the field without requiring additional hardware and software.

1. Introduction

With the increasing global warming threats, the United Nations has called for the reduction of carbon dioxide emissions and hence set out the goals of reducing greenhouse gas emissions by 45 percent by 2030 and to net zero emissions by 2050 [1]. In line with the United Nations’ goals, the developed and most developing countries have started to take actions to develop realistic plans toward the reduction of carbon dioxide emissions. For instance, the Chinese government announced their goals to reach the peak of carbon dioxide emissions before 2030 and strive to achieve carbon neutrality before 2060 at the 75th session of the United Nations General Assembly (UNGA 75) in September 2020 [2]. In 2019, the total carbon dioxide emissions in China were estimated at 10.5 billion tons, of which the carbon emissions from energy consumption were about 9.8 billion tons, accounting for around 87% of the total emissions [3]. With rapid economic and social development, the transition to a green and low-carbon society is accelerated, and the transition of the country’s energy structure brooks no delay. Cleanliness is an important direction for carbon emission reduction in energy production. The way to accomplish the “dual carbon” task is to develop green and low-carbon renewable energy. During the 14th Five-Year Plan period, coal consumption in China will continue to decline. In the plan, it is forecasted that the installed capacity of renewable energy and nuclear power will reach 1200 GW by 2030, of which wind power will reach 500 GW in China [4].
As the installed capacity of wind power continues to grow and wind power is connected to the grid on a large scale, the overall grid performance is affected by the output power from wind farms due to the intermittency of wind [5]. In order to ensure the safety and stability of the operation of the power system, the power grid needs to prepare a sufficient spinning reserve capacity. However, the increase in the reserve capacity will increase the operating cost of wind power. Therefore, accurate wind power forecasting (WPF) is required for providing a basis to develop a grid dispatch schedule, and it also helps to greatly reduce the operating costs of wind farms and improve the competitiveness of wind energy in the overall energy market [6,7,8].
Historically, there have been different wind power forecasting (WPF) methods, which can be divided into four categories: physical, statistical, hybrid, and deep-learning methods. A summary report of these four categories of methods in terms of their features and limitations in application is given in [9]. The physical method is based on a mesoscale weather model or a numerical weather prediction system (NWP). NWP represents a variety of mathematical expression models of geographic and meteorological information [10,11]. Although this method has a good effect on short or medium-term forecasts of more than 3 h, it is difficult for it to collect all relevant geographic or meteorological data [12,13,14], so it has limitations in application. The physical forecasting method is generally used to select/determine new wind farms, but not for wind turbine power production prediction.
The statistical prediction method is based on the historical data collected by the SCADA system to establish a linear/non-linear relationship between relevant index data and power to predict the output power of a wind farm. Statistical prediction methods can be categorized as conventional statistical methods and those based on artificial neural networks (ANN). The conventional statistical methods have limitations in forecasting due to the demand for non-linear expression in wind power forecasting (WPF), while the methods based on artificial neural networks can effectively represent a large number of non-linear relationships and complex characteristics among wind speed, temperature, and other parameters in power generation. Therefore, statistical prediction methods based on ANN have become widely applied. Ref. [15] proposed a shallow model for wind speed prediction (WSF) based on artificial neural networks, which is more accurate than physical or traditional statistical methods.
The hybrid method integrates the physical and statistical models to improve forecast performance by preserving the advantages of each approach [16,17,18], but the hybrid models may not have the capability to achieve stable prediction, as their complex learning architecture may cause low efficiency in training and even underfitting [9].
With the development of deep learning techniques in recent years, and because wind power prediction possesses a natural time-series feature, some deep neural network (DNN)-based time series forecast methods have been developed and used for wind power estimation, such as the methods based on recurrent neural networks (RNNs) [19], long short-term memory (LSTM) [20], Transformer [21], temporal convolutional networks (TCNs) [22], sample convolution and interaction networks (SCINets) [23], etc. These form the deep learning-based method for wind power forecast. It is promising in terms of new model development for time series forecasting; however, none of the existing methods can claim to be perfect in time series prediction, which depends on the available data and data quality.
SCINet is a novel framework proposed by Liu et al. [23] very recently that has been applied to time series forecasting problems. It performs sample convolution and interaction at multiple resolutions for time-series modeling. Although good prediction results can be achieved by SCINet models, the SCINet framework has some shortcomings that affect the prediction performance, i.e., the prediction accuracy. One of the shortcomings is the strict binary tree structure taken in SCINet causing information blockage as the number of network levels increases. To address this issue, this present paper proposes a novel framework with a shuffle–and–fusion interaction network, named SFINet, to avoid the information blocking of SCINet sequence channels and develops an improved algorithm for wind power forecasting. The developed models based on SFINet have been proven effective to achieve the economic dispatch of energy production and reliable operation of the power system, providing an opportunity for reducing the operation costs of wind farms. The main contributions are as follows: (1) considering the sequence interaction modeling of time series tasks, we introduced the shuffle operation to increase the dependence between sequences; (2) in order to further promote interactive learning, we proposed feature fusion based on the time series attention mechanism; and (3) the developed models are applied to wind power forecasting using real wind power production data collected from a wind farm in China, verifying the outperformance of our models by comparing them with other baseline approaches.

2. Deep Learning-Based Method for Wind Turbine Operation and Power Forecasting

2.1. Channel Interleave Operation

To the best of our knowledge, the first real use of channel alternate operation was in IGCNets [24], and channel interleave in the form of shuffle was proposed in shufflenet [25], which aimed to break the information blockage between group convolutions. Subsequently, channel shufflenet has been widely utilized as a basic backbone network [26,27], with applications in semantic segmentation, Multi-Person Pose Estimation, Image Processing, and other tasks [28,29,30]. However, channel shufflenet is mostly used on the basis of grouped convolution and lightweight models. In this paper, we apply it to the construction of the sequence channels for time series forecasting to improve the interaction capabilities of different time series features.

2.2. Attention Mechanisms

Attention is essentially a tool to filter and focus important information from a large number of available processing resources, while ignoring non-important information [31,32]. It is usually combined with threshold functions, such as softmax and sigmoid, or sequential techniques [33,34]. In both computer vision and sequence tasks, it has shown superior performance [35,36]. In these applications, the attention mechanism usually acts on one or more top layers to further reshape the characteristics of the higher level. Attention mechanisms have provided a lot of benefits in many applications, e.g., image classification [37], object detection [38], multi-modal task [39], few-show learning [40], and machine translation [21].
The more common attentions are channel attention [37,41], spatial attention [37,42], temporal attention [43], and branch attention [44]. Channel attention adaptively recalibrates the weight of each channel and can be viewed as an object selection process, thus determining what to pay attention to. Hu et al. [41] introduced a lightweight attention operation with a Squeeze–and–Excitation block to model channel-wise relationships. Spatial attention can be seen as an adaptive spatial region selection mechanism for determining where to pay attention to. Dai et al. [42] proposed deformable convolutional networks (deformable ConvNets) to be invariant to geometric transformations, but they paid attention to the important regions in a different manner. Self attention [21] is also used as a spatial attention mechanism to capture global information. Temporal attention is a dynamic time-selection mechanism. Li et al. [43] proposed a global–local temporal representation (GLTR) to exploit multi-scale temporal cues in a video sequence. In a multi-branch structure, branch attention is used for branch selection. Reference [44] proposed an automatic selection operation called selective kernel (SK) convolution implemented using three operations: split, fuse, and select.
The above-mentioned attention methods are often combined in application. Chen et al. [45] dynamically modulated the sentence generation context in multi-layer feature maps using encoding channel attention and spatial attention. Reference [46] identified spatial saliency associated with image pixels and executed temporal intensity filtering and predictive coding to filter spatiotemporal redundancies from images.
On the basis of the above overview, we propose a feature fusion method based on time series channel attention, aiming to enhance the model’s long-term forecasting ability in wind power forecasting.

2.3. Deep Learning-Based Wind Power Forecasting

The wind speed and power indicators collected through the wind turbine SCADA system are all time series data. Time series forecasting can estimate their future development based on indicators or events. At the same time, there are complex nonlinear relationships among other indicator data related to power. From the previously published research works, it has been realized that deep learning-based time series forecasting has higher forecasting accuracy than the traditional methods [47], so the deep learning-based time series forecasting (TSF) method has been widely utilized.
The recurrent neural network (RNN)-based TSF method given in [48,49] compactly summarizes the past information in the internal memory used for prediction, where the memory state is recursively updated with new inputs at each time step, as shown in Figure 1a below. Ref. [50] proposed a long and short-term memory-based recurrent neural network (LSTM-RNN) to predict the wind power from 1 to 24 h. Transformer relies on the attention mechanism to model the global dependency of input and output, and breaks the non-parallelization problem of RNN-based methods, so it is gradually replacing the RNN model in almost all sequence modeling tasks. Therefore, various Transformer-based TSF methods were presented in [51], as shown in Figure 1b. The multi-head self-attention mechanism is used to extract the spatial correlation between wind farms [52]. Models based on convolutional neural networks (CNNs), such as temporal convolutional networks (TCNs), are also used in time series forecasting (TSF) [53,54]. The TCN uses a series of causal convolutional layers stacked to make full use of convolution. Parallel operation with efficient modeling of the dependency relationship between multiple sequence features is shown in Figure 1c. Long-term prediction of wind power with a mean absolute percentage error of 10% was carried out in [9] by using the temporal convolutional network (TCN).
Ref. [23] proposed a new neural network structure named SCINet, as shown in Figure 1d, specifically designed for time series forecasting, which lifts causal convolutional layers and the forced numbers of network input and output to be the same constraints of TCNs, and achieved very good performance in TSF tasks. To the best of our knowledge, however, SCINet has not been applied to the field of wind power forecasting (WPF). At the same time, the binary tree structure of SCINet causes information blockage as the number of network levels increases. For this reason, we propose shuffle–and–fusion interaction networks (SFINet) to overcome this issue.

3. SFINet: Shuffle–and–Fusion Interaction in Convolution Networks

As mentioned above, SCINet follows the strictly binary tree structure, and the time-series feature information will no longer have the opportunity for information interaction after passing through the parent node of the binary tree. Although there is an interactive learning operation in SCI-Block, it can fuse information between time-series, but this interactive process only exists at the node of the parent tree, which means that the subsequent layers of different depths can only come from the first interactive learning of the parent node for the most primitive timing input. As the number of tree layers deepens, this information will be transmitted more insignificantly. We think that this feature is very unfavorable for capturing the dependencies between long sequences.
We use Figure 2 to illustrate this feature of the original SCINet structure, where the most basic unit is SCI-Block, as shown in Figure 2a. SCI-Block contains interactive learning modules responsible for the interaction between two different timing features. The SCINet is composed of basic SCI-Blocks according to the strictly binary tree structure, as shown in Figure 2b, and the SCI-Block always averages the split input features in the timing dimension; finally, SCINet is stacked to form Stacked SCINet. We named the input feature of an SCI-Block as SSF (split-sequence features), then the input to the k-th SCI-Block of the l-th layer can be defined as SSF(l, k), where l = 1 ,   2 , , L ,   and   k = 2 0 , 2 1 , , 2 L 1 . Due to the average split characteristic in the timing dimension, for a timing sequence whose input length is known to be S, then L [ 1 , n ] ,   and   2 n ≤ S.
Obviously, in the SCINet structure, SSF(l, k) is with l > 1 , and   k ( 1 , 2 l 1 ) . Only part of the output from the upper layer of the module will interact with other outputs in this layer. As shown in Figure 2b, each of SSF(3, 1), SSF(3, 2), SSF(3, 3), SSF(3, 4) is input into the corresponding SCI-Block for further reasoning, and they are mutually isolated. Their interactive learning only exists in the first layer of SCI-Block. Hence, this property blocks information flow between sequential channels and weakens representation.

3.1. Shuffle Split-Sequence Features

If SCI-Block is allowed to obtain an input from different split-sequence features, the split-sequence features that characterize different time dimensions will obtain better interaction. Therefore, the shuffle operation was introduced on the basis of SCINet. Specifically, for the structural characteristics of SCINet, for each layer of input split-sequence features, the inputs are naturally presented in groups at each level. First, the channels of each group can be divided into sub-groups, and then different sub-groups can be evenly allocated to each group as the input to the next layer. This operation is a sequential operation, as shown in Figure 3a.
The above operations can be efficiently implemented through a channel shuffle operation, as shown in Figure 3b. Use 2 l groups to form new sequence features, and its output channel has 2 2 l sub-groups. First, reshape the output channel size to   ( 2 l , 2 l ) , transposing and then flatten it back as the input of the next layer. Channel shuffling is also differentiable, which means that it can be embedded into network structures for end–to–end training. The shuffle operation makes it possible to build more powerful structures with sequential interactive learning.
On this basis, the shuffle operation was embedded in the SCINet structure, and therefore, the SFINet structure was designed as shown in Figure 3c. The shuffle operation acted on the output of all leaf nodes of all SCI-Blocks in each layer. There was only one shuffle operation in each layer, which had nothing to do with the number of SCI-Blocks in that layer.

3.2. Fusion with Channel Attention

Taking into account the natural law of features in timing, this paper does not directly use the feature SSF ¯ after shuffle operation to replace the original SSF, but merges the two. Here, a simple method of adding corresponding positions was used to realize the fusion of the two parts; that is, to ensure the sequence relationship of the original SSF in time series and to ensure the communication of features in different slice groups.
Before feature fusion, the attention operation was further performed on SSF ¯ , which was completed by the attention block shown in Figure 4a.
First of all, in order to ensure that the features of each dimension on the timing channel are fully utilized and can participate in subsequent predictions, we propose squeezing global spatial information into a channel descriptor. Exploiting such information is prevalent in feature engineering work. We opted for the simplest. This was achieved by using global average pooling to generate channel-wise statistics. Formally, a statistic z c   was generated by shrinking SSF ¯ through spatial dimensions W, where the c-th element of z is calculated by:
z c = F g ( SSF ¯ ) = 1 W i = 1 w SSF ¯ ( i )
To make use of the information aggregated in the squeeze operation, we followed it with a second operation that aims to fully capture channel-wise dependencies. We performed one-dimensional convolution again after Z, and the output was E. At the same time, it was to ensure that the characteristic channel output was the original length, and we finally used Sigmoid to activate E. Decoupling realizes the release of the dependency between different channels, and outputs of the final attention coefficient S = [ s 1 , s 2 , , s c ] . The final output   X ¯   = [ x 1 ¯ , x 2 ¯ , , x c ¯ ] was obtained by rescaling the transformation output SSF ¯ with the activation, such that the c-th element of     X ¯     was calculated by
x c ¯ = F scale ( S c   ,   SSF c ¯ ) = S c · SSF c ¯
where   F scale ( S c   , SSF c ¯ ) refers to the channel-wise multiplication between the feature map SSF c ¯ w and the scalar S c .
Finally, we embedded the Attention Block into SCINet, as shown in Figure 4b; the output after the Attention Block will be added and fused with the original SSF and sent to the next layer of inference.

3.3. SFINet Architecture

After each layer of SCINet is the output, it is connected to the shuffle operation and attention block. Hence, a new network structure SFINet—the shuffle–and–fusion interaction network is proposed. This structure represents different timing information features. After passing through each layer, they will be shuffled and further processed at the attention block. After the above operations are completed, they will be fused with the previous features, and then enter the next layer of reasoning, thus breaking the information blockage in the original structure. After the above operations, not only the capture of short-term dependencies in timing is guaranteed, but also the ability to build long-term dependencies in timing is further improved.
It should be noted that, when level = 1, SFINet has the same structure as the original SCINet. Because the original input is directly connected to the output after a layer of reasoning, it does not go through the shuffle operation and attention block modules; however, in actual tasks, most tasks require a structure with larger than 2 levels.

4. Wind Power Forecast

This section provides a test of the proposed SFINet for wind power forecasting. Section 4.1 introduces the data sets included in this study. There were five data sets utilized: two of them were the collected real wind power data and three were the published data sets. Section 4.2 introduces detailed power prediction and other experimental settings. Section 4.3 shows the performance and usability of the proposed method, as well as the testing results. The effectiveness of SFINet was verified by comparing it with various other methods including SCINet.

4.1. Data Set Selection

We empirically perform the test of the established models using five data sets: two of them were collected from a wind farm and the other three were selected from the published benchmark data sets.
The data collected from a wind farm represent the operation data of two independent wind turbines, each with a rated power of 1.5 MW, for one year with a sampling frequency of once per 10 min, i.e., 10 min data. The two data sets were marked as WPm1 and WPm2, respectively. The data for each sampling point include 12 variables—see Table 1 below. These variables were most relevant to wind power generation, including the wind speed, generator output power, pitch angle, nacelle position (or yaw angle), wind direction (vane direction), etc. These variables were selected by referring to the previously published research articles for wind power forecast and through discussions with the wind farm operation manager and engineers. There were two wind turbine operation modes: Mode 1 was marked as 20 when there was power output to the grid and Mode 2 was marked as 6 when there was no power output or the wind turbine stopped running. Other values represent the wind turbine running in a transition status between the two operation modes. The value was calculated based on the time length when it was running in Mode 1 and the time length when it was running in Mode 2 in a 10 min time step. Similarly, there were two wind turbine braking modes: Under braking and no braking. When it was under braking, the variable Turbine Brake Level was assigned a value of 51, whereas it was given 0 if there was no braking. Other values assigned to this variable represent the braking is in a transition status between the two braking modes. The value was calculated based on the time length when it was in braking mode and the time length when there was no braking. Each value was calculated in average on a 10 min time step.
The ratio of the training data set to the validation set and the test set was 5:2:3. See Table 2 for detailed information. These two data sets were used to verify the effectiveness of the proposed method in wind power forecasting.
Electricity Transformer Temperature (ETT) data were collected and used in [55]. The ETT data cover 2 years’ data collected from two separate counties in China. They were split into two data sets marked as ETTh1 and ETTh2, respectively, with a sampling frequency of once per hour. The ETTm1 data set was 15 min data, i.e., the sampling frequency was once per 15 min. Each data point consisted of the target value of “oil temperature” (°C) and six power load features—see Table 3 below. The ratio of the training data set to the validation set and the test set was 3:1:1. See Table 2 for detailed information. The ETT data sets were used to demonstrate the general validity of the proposed method.

4.2. Experiment Implementation

In order to evaluate the performance of the proposed method in different aspects, a variety of tasks were defined based on the wind power data set, including prediction tasks of different horizons and univariate or multivariate predictions.
In terms of horizon, similar to ETT data, in the case of a fixed sampling time, different output sequence lengths characterized different prediction times and also showed the difficulty of the task. The prediction lengths of WPh1 and WPh2 were divided into 6, 12, 24, and 48, and the corresponding prediction times were 1 h, 2 h, 4 h, and 8 h. In terms of variables, two forms of multivariate and univariate were used for evaluation. Univariate prediction takes the value of the generated wind power per second (A1gr Gen Power for Process_1sec), and multivariate prediction takes the values from all variables.
For ETT data, the prediction lengths of ETTh1 and ETTh2 were 24, 48, 96, 288, and 720, respectively, and the corresponding prediction time lengths were 24 h, 48 h, 96 h, 288 h, and 720 h. The predicted lengths of ETTm1 were 24, 48, 96, 288, and 672, corresponding to the predicted times of 6 h, 12 h, 24 h, 72 h, and 168 h, respectively. Univariate prediction takes the value of oil temperature, and multivariate prediction takes the values from all variables.
(1)
Evaluation index
The Mean Absolute Error (MAE) and Weighted Mean Absolute Percentage Error (WMAPE) were used as evaluation criteria. Because some variables in the task may have had negative values (the data sets in the format as Table 1 and Table 2), the optimized WMAPE was calculated as follows,
WMAPE = i = 0 τ |   x ¯ i x i | i = 0 τ | x i |
where τ is the total number of data points and   x ¯ i is the mean of the data in the sample.
The relative improvement of performance (RIP) with MAE and absolute improvement of performance (AIP) with WMAPE were used for comparison. They were calculated as follows:
RIP = MAE O MAE MAE O ,
AIP = WMAPE o WMAPE ,
where MAE = 1 τ i = 0 τ |   x ¯ i x i | ,   MAE o   and WMAPE o are obtained using other competitive methods for prediction, and MAE and WMAPE are obtained using the newly proposed method based on the SFINet model.
(2)
Data processing
In order to evaluate the performance of our proposed algorithm in wind turbine power prediction, we conducted experiments based on the proposed WP data sets and compared the prediction results with those given by the SCINet models. At the same time, in order to further verify the general applicability of the algorithm, we also verified the performance of the algorithm using the ETT data sets and compared it with other methods, including SCINet.
Our task does not specify the look-back windows corresponding to a certain prediction sequence horizon. In the wind power forecasting task, the original wind power data have singular values. In order to eliminate the singular values shown in Table 1, we chose to skip singular values when constructing the training, validation, and testing for different tasks. If there was a singular value in a pair, this piece of data was discarded. Therefore, in this way, for different tasks, the number of samples for training, validation, and testing were be produced, as shown in Table 4. Table 4 also shows the input data time length for a certain prediction length; for example, if the prediction length is for 6 h, the input data covers a time length of 128 h.
The singular value issue did not appear in the ETT data set, and these data could be used normally. In the ETT task, the settings are shown in Table 5.
All data were normalized. In terms of the loss function and optimizer, we followed the same settings for the SCINet model given in [23].
(1)
Hardware platform
Training GPU: Single Nvidia 3080Ti 16 GB.
CPU Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20 GHz.
Memory: 256 G.

4.3. Prediction Experiment and Result Analysis

(1)
Wind power forecasting
Applying SFINet to the wind power data sets, the forecasting performance obtained is shown in Table 6. It can be seen from Table 6 that the prediction results of the SFINet model proposed in this paper were generally better than those of the SCINet model. The evaluation criteria of MAE using the multivariate and univariate prediction results with the WPm1 data set were reduced by up to 10.07% and 6.90%, respectively; and up to 9.20 and 6.87%, respectively, for the WPm2 data set.
The forecast results evaluated using WMAPE are shown in Table 7. From the results given in Table 7, the superiority of the SFINet model algorithm was verified in univariate and multivariate wind turbine power prediction. The evaluation indices of WMAPE with the WPm1 data set were reduced by up to 5.79% using multivariate evaluation and 4.93% using univariate evaluation, and the WMAPE with the WPm2 data set were reduced by up to 4.86 and 4.35%, respectively. The performance improvement trend of the SFINet model under different horizon tasks was similar to the results using MAE metric.
We then performed qualitative analysis of the prediction results using the wind power data set by selecting a piece of wind turbine power data for a sampling length of 200 in the WPh1 data set and WPh2 data set, as shown in Figure 5a,b. It can be seen that the forecasted wind power had the characteristics of large variation, violent fluctuations and no obvious laws to follow. The prediction result of the SFINet model could better fit the actual power curve. At the same time, at the peak and valley points of the power change, the prediction result of the SFINet model was generally better than that of the SCINet model (see also Figure 5c) by using the published data set.
(2)
Generalization study
The ETT data set given in [55] were used to evaluate the performance of a time series forecasting task e.g., [23,55]. In this paper, we used the same dataset to evaluate the performance of time series forecasting by different approaches, and the results of multivariate and univariate prediction are shown in Table 8 and Table 9.
As shown in Table 8, in multivariate prediction, the prediction effects of Transformer-based methods other than Reformer [56], such as LogTrans [52] and Informer [55], outperformed the RNN-based methods, such as LSTMa [5]; the performance of TCN [22] further outperformed Transformer-based methods; compared with these methods, SCINet model achieved better performance, because the downsample–convolve–interact architecture enabled multi-resolution analysis, which facilitated extracting temporal relation features with enhanced predictability. Overall, in this paper, as shown in all subtasks with ETT data, the prediction performance using SFINet was all better than that using SCINet—see the relative performance improvement given by RIP as shown in green color.
By comparison with the multivariate prediction, the performance of these methods in discussion for univariate prediction was gradually improved. N-Beats [57] outperforms the above methods, and it is observed that SCINet is superior to other baseline methods. However, the performance of SFINet in time series forecasting is even better than that of SCINet.
Specifically, for the two different tasks using the ETTh1 data set, the evaluation criterion of MAE was improved by 6.33% and 0.76% or more, respectively, while using the ETTh2 and ETTm1 data sets, the prediction performance was improved by 6.94% and 1.76%, and 3.06% and 1.14% or more, respectively. When increasing the horizon, the improvement in MAE showed an increasing trend. The results further confirmed the effectiveness and universality of the algorithm proposed in this paper.
In order to make further comparison between SFINet and SCINet in time series forecasting, we performed the prediction using the ETT data set and the prediction performance was evaluated using the metric of WMAPE—see Table 9. SFINet also achieved better performance than SCINet. More specifically, in the multivariate task and univariate task on the ETTh1 data set, the WMAPE was improved by at least 2.88% and 1.04%, respectively, while, on the ETTh2 and ETTm1 data sets, the above-mentioned performance was improved by 1.11% and 1.47%, and 1.13% and 0.43% or more, respectively.
Finally, the prediction results using ETTh1 data set for a horizon of 168 are shown in Figure 5c as an illustration example. The predicted results could very well fit the actual values, and the degree of fitting was better than that of the SCINet model.

5. Conclusions

In this paper, we propose a new framework named SFINet: shuffle–and–fusion interaction network. The SFINet model included a shuffle operation and a feature fusion function based on the attention mechanism. The shuffle operation was succinctly embedded in between the adjacent layers of SFINet, which increased the interaction of the different time series features of the model. At the same time, in order to more effectively integrate the features of different parts, a feature fusion function based on the attention mechanism was proposed to enhance the feature interaction capabilities of different parts of the model. The developed SFINet models were tested using real data of wind power generation. It was verified that the SFINet models provided better performance than the network algorithm based on SCINet. At the same time, in order to verify the universality of the proposed framework, we evaluated the performance of SFINet models using the ETT datasets, which are open and published datasets. The model we proposed presented wind power forecast performance which was better than that of seven other types of algorithms in comparison. In future work, we will focus on further improvement of the SFINet structure with more applications to show the powerfulness of SFINet models in time series forecasting.

Author Contributions

X.Z. was responsible for theoretical development and developed the algorithms; C.X. contributed to the methodology and verified the modelling; X.Z., C.X. and T.Z. drafted the manuscript; T.Z. verified the algorithms and the results, finalized the paper, and was responsible for paper revision and submission. Each author contributed to the research approach development. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the Key Project of North China Institute of Aerospace Engineering (ZD202003); the Excellent Going Abroad Experts Training Program of Hebei Province; Domestic Visiting Program for Young Scholars in Universities in the Midwest of China; and the Australian Government 2022 NCP grant (No. 34223).

Data Availability Statement

Not applicable.

Acknowledgments

The wind turbine operation data were collected from a wind farm in China. The authors are grateful to the wind farm managers and engineers for their kind support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Summit, C.A. Report of the Secretary-General on the 2019 Climate Action Summit and the Way Forward in 2020; United Nations: New York, NY, USA, 2019. [Google Scholar]
  2. Xi, J.P. Statement by H.E. Xi Jinping, President of the People’s Republic of China at the General Debate of the 75th Session of The United Nations General Assembly [EB/OL]. Available online: https://www.fmprc.gov.cn/ce/cgmb/eng/zxhd/t1817304.html (accessed on 16 December 2021).
  3. Grant, M.; Larsen, K. Preliminary China Emissions Estimates for 2019 [EB/OL]. Available online: https://rhg.com/research/preliminary-china-emissions-2019 (accessed on 16 December 2021).
  4. United Nations Development Program China. China’s 14th Five-Year Plan Spotlighting Climate and Environment [EB/OL]. Available online: https://www.carbonbrief.org/qa-what-does-chinas-14th-five-year-plan-mean-for-climate-change (accessed on 16 December 2021).
  5. Bird, L.; Milligan, M.; Lew, D. Integrating Variable Renewable Energy: Challenges and Solutions; Technical Report NREL/TP-6A20-60451; National Renewable Energy Lab (NREL): Jefferson County, CO, USA, 2013.
  6. Kamath, C. Understanding wind ramp events through analysis of historical data. In Proceedings of the IEEE PES Transmission and Distribution Conference and Exposition, New Orleans, LA, USA, 19–22 April 2010. [Google Scholar] [CrossRef] [Green Version]
  7. Guo, Z.H.; Wu, J.; Lu, H.Y.; Wang, J.Z. A case study on a hybrid wind speed forecasting method using BP neural network. Knowl.-Based Syst. 2011, 24, 1048–1056. [Google Scholar] [CrossRef]
  8. Wang, Y.; Zhang, K.; Fu, J.Y.; Pang, X.; Geng, J. Optimization control method of wind/storage system for suppressing wind power ramp rate. Autom. Electr. Power Syst. 2013, 37, 17–23. [Google Scholar] [CrossRef]
  9. Lin, W.H.; Wang, P.; Chao, K.M.; Lin, H.C.; Yang, Z.Y.; Lai, Y.H. Wind power forecasting with deep learning networks: Time-series forecasting. Appl. Sci. 2021, 11, 10335. [Google Scholar] [CrossRef]
  10. Nielsen, H.A.; Nielsen, T.S.; Madsen, H. Optimal combination of wind power forecasts. Wind. Energy 2007, 10, 471–482. [Google Scholar] [CrossRef]
  11. Lange, M.; Focken, U. Physical Approach to Short-Term Wind Power Prediction; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar] [CrossRef]
  12. Giebel, G.; Brownsword, R.; Kariniotakis, G.; Denhard, M.; Draxl, C. State-of-the-Art in Short-Term Prediction of Wind Power: A Literature Overview; Technical Report of Project ANEMOS.plus and SafeWind; Project SafeWind: Paris, France, 2011. [Google Scholar] [CrossRef]
  13. Ren, Y.; Suganthan, P.; Srikanth, N. Ensemble methods for wind and solar power forecasting-a state-of-the-art review. Renew. Sustain. Energy Rev. 2015, 50, 82–91. [Google Scholar] [CrossRef]
  14. Maatallah, O.A.; Achuthan, A.; Janoyan, K.; Marzocca, P. Recursive wind speed forecasting based on hammerstein auto-regressive model. Energy 2015, 145, 191–197. [Google Scholar] [CrossRef]
  15. Cadenas, E.; Rivera, W. Short term wind speed forecasting in la venta, Oaxaca, México, using artificial neural networks. Renew. Energy 2009, 34, 274–278. [Google Scholar] [CrossRef]
  16. Mana, M.; Astolfi, D.; Castellani, F.; Meißner, C. Day-ahead wind power forecast through high-resolution mesoscale model: Local computational fluid dynamics versus artificial neural network downscaling. J. Sol. Energy Eng. 2020, 142, 034502. [Google Scholar] [CrossRef]
  17. Emeksiz, C.; Tan, M. Multi-step wind speed forecasting and Hurst analysis using novel hybrid secondary decomposition approach. Energy 2022, 238, 121764. [Google Scholar] [CrossRef]
  18. Jiang, P.; Liu, Z.; Niu, X.; Zhang, L. A combined forecasting system based on statistical method, artificial neural networks, and deep learning methods for short-term wind speed forecasting. Energy 2021, 217, 119361. [Google Scholar] [CrossRef]
  19. Medsker, L.R.; Jain, L.C. Recurrent neural networks. Des. Appl. 2011, 5, 64–67. [Google Scholar]
  20. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  21. Shaw, P.; Uszkoreit, J.; Vaswani, A. Self-attention with relative position representations. arXiv 2018, arXiv:1803.02155. [Google Scholar]
  22. Bai, S.J.; Zico Kolter, J.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  23. Liu, M.H.; Zeng, A.L.; Lai, Q.X. Time series is a special sequence: Forecasting with sample convolution and interaction. arXiv 2021, arXiv:2106.09305. [Google Scholar]
  24. Zhang, T.; Qi, G.J.; Xiao, B. Interleaved group convolutions for deep neural networks. arXiv 2017, arXiv:1707.02725v2. [Google Scholar]
  25. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An extremely efficient convolutional neural network for mobile devices. arXiv 2017, arXiv:1707.01083v2. [Google Scholar]
  26. Ma, N.N.; Zhang, X.Y.; Zheng, H.T.; Sun, J. ShuffleNet V2: Practical guidelines for efficient CNN architecture design. arXiv 2018, arXiv:1807.11164v1. [Google Scholar]
  27. Yang, Q.L.; Yang, Y.B. SA-Net: Shuffle attention for deep convolutional neural networks. arXiv 2021, arXiv:2102.00240v1. [Google Scholar]
  28. Wang, Y.; Zhou, Q.; Liu, J.; Xiong, J.; Gao, G.; Wu, X.; Latecki, L.J. Lednet: A lightweight encoder-decoder network for real-time semantic segmentation. In Proceedings of the 2019 International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1860–1864. [Google Scholar]
  29. Li, W.; Li, S.M.; Liu, R.H. Channel shuffle reconstruction network for image compressive sensing. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 2880–2884. [Google Scholar]
  30. Su, K.; Yu, D.D.; Xu, Z.Q.; Geng, X.; Wang, C. Multi-person pose estimation with enhanced channel-wise and spatial information. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  31. Laurent, I.; Koch, C. Computational modelling of visual attention. Nat. Rev. Neurosci. 2001, 2, 194–203. [Google Scholar]
  32. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA; pp. 6000–6010. [Google Scholar]
  33. Cao, C.S.; Liu, X.M.; Yang, Y.; Yu, Y.; Wang, J.; Wang, Z.; Huang, Y.; Huang, C.; Xu, W.; Ramanan, D.; et al. Look and think twice: Capturing top-down visual attention with feedback convolutional neural networks. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 2956–2964. [Google Scholar]
  34. Jaderberg, M.; Simonyan, K.; Zisserman, A.; Kavukcuoglu, K. Spatial transformer networks. In Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; MIT Press: Cambridge, MA, USA; Volume 2, pp. 2017–2025. [Google Scholar]
  35. Bluche, T. Joint line segmentation and transcription for end-to-end handwritten paragraph recognition. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; Curran Associates Inc.: Red Hook, NY, USA; pp. 838–846. [Google Scholar]
  36. Antoine, M.; Laptev, I.; Sivic, J. Learnable pooling with context gating for video classification. arXiv 2018, arXiv:1706.06905V2. [Google Scholar]
  37. Woo, S.H.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  38. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar]
  39. Su, W.; Zhu, X.; Cao, Y.; Li, B.; Lu, L.; Wei, F.; Dai, J. Vl-bert: Pre-training of generic visual-linguistic representations. arXiv 2019, arXiv:1908.08530, 2019. [Google Scholar]
  40. Peng, Y.; He, X.; Zhao, J. Object-part attention model for fine-grained image classification. IEEE Trans. Image Processing 2017, 27, 1487–1500. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; Volume 42, pp. 2011–2023. [Google Scholar]
  42. Dai, J.F.; Qi, H.Z.; Xiong, Y.W.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar] [CrossRef] [Green Version]
  43. Li, J.N.; Zhang, S.L.; Wang, J.D.; Gao, W.; Tian, Q. Global-local temporal representations for video person re-identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 3957–3966. [Google Scholar] [CrossRef] [Green Version]
  44. Li, X.; Wang, W.H.; Hu, X.L.; Yang, J. Selective kernel networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 510–519. [Google Scholar] [CrossRef] [Green Version]
  45. Chen, L.; Zhang, H.; Xiao, J.; Nie, L.; Shao, J.; Liu, W.; Chua, T.-S. SCA-CNN: Spatial and channel-wise attention in convolutional networks for image captioning. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5659–5667. [Google Scholar] [CrossRef] [Green Version]
  46. Bhowmik, P.; Pantho, J.H.; Mbongue, J.M.; Bobda, C. ESCA: Event-based Split-CNN architecture with data-level parallelism on ultraScale+ FPGA. In Proceedings of the 2021 IEEE 29th Annual International Symposium on Field-Programmable Custom Computing Machines (FCCM), Orlando, FL, USA, 9–12 May 2021; pp. 176–180. [Google Scholar]
  47. Lim, B.; Zohren, S. Time-series forecasting with deep learning: A survey. Philos. Trans. R. Soc. A-Math. Phys. Eng. Sci. 2021, 379, 20200209. [Google Scholar] [CrossRef]
  48. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  49. Lai, G.; Chang, W.C.; Yang, Y.; Liu, H. Modeling long-and short-term temporal patterns with deep neural networks. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 95–104. [Google Scholar] [CrossRef] [Green Version]
  50. Cali, U.; Sharma, V. Short-term wind power forecasting using long-short term memory based recurrent neural network model and variable selection. Int. J. Smart Grid Clean Energy 2019, 8, 103–110. [Google Scholar] [CrossRef] [Green Version]
  51. Li, S.Y.; Jin, X.Y.; Xuan, Y.; Zhou, X.; Chen, W.; Wang, Y.; Yan, X. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Curran Associates Inc.: Red Hook, NY, USA; pp. 5243–5253. [Google Scholar]
  52. Fu, X.B.; Gao, F.; Wu, J.; Wei, X.; Duan, F. Spatiotemporal attention networks for wind power forecasting. In Proceedings of the 2019 IEEE International Conference on Data Mining Workshops (ICDMW), Beijing, China, 8–11 November 2019. [Google Scholar] [CrossRef] [Green Version]
  53. Wu, Z.H.; Pan, S.R.; Long, G.D.; Jiang, J.; Zhang, C. Graph wavenet for deep spatial-temporal graph modeling. arXiv 2019, arXiv:1906.00121. [Google Scholar]
  54. Yu, B.; Yin, H.T.; Zhu, Z.X. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. arXiv 2017, arXiv:1709.04875. [Google Scholar]
  55. Zhou, H.Y.; Zhang, S.H.; Peng, J.Q.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, 2–9 February 2021; Volume 35, pp. 11106–11115. Available online: https://ojs.aaai.org/index.php/AAAI/article/view/17325 (accessed on 17 December 2021).
  56. Kitaev, N.; Kaiser, L.; Levskaya, A. Reformer: The efficient transformer. arXiv 2020, arXiv:2001.04451. [Google Scholar]
  57. Oreshkin, B.N.; Carpov, D.; Chapados, N.; Bengio, Y. N-BEATS: Neural basis expansion analysis for interpretable time series forecasting. arXiv 2019, arXiv:1905.10437. [Google Scholar]
Figure 1. Typical deep-learning network structure for processing time series tasks [23,51,53,54].
Figure 1. Typical deep-learning network structure for processing time series tasks [23,51,53,54].
Applsci 12 02253 g001
Figure 2. The overall architecture of the Sample Convolution and Interaction Network (SCINet) [23].
Figure 2. The overall architecture of the Sample Convolution and Interaction Network (SCINet) [23].
Applsci 12 02253 g002
Figure 3. Shuffle swap split-sequence features.
Figure 3. Shuffle swap split-sequence features.
Applsci 12 02253 g003
Figure 4. Fusion with channel attention.
Figure 4. Fusion with channel attention.
Applsci 12 02253 g004
Figure 5. The prediction comparison between SFINet and SCINet for different datasets.
Figure 5. The prediction comparison between SFINet and SCINet for different datasets.
Applsci 12 02253 g005
Table 1. Illustration of the wind power data set including 12 variables.
Table 1. Illustration of the wind power data set including 12 variables.
IndexName1 January 2019 14:101 January 2019 14:201 January 2019 14:301 January 2019 14:401 January 2019 14:501 January 2019 15:00
1Wind Speed (m/s)3.373.00*3.403.133.46
2Generator Output Power (KW)0.08403.0268*21.673056.0637104.1866
3Pitch Angle (°)75.8127.10*40.70−0.5−0.5
4Nacelle Position (°)246.39165.04*162.67164.58158.25
5Vane Direction (°)−76.09−1.83*0.4770.564−2.687
6Cumulative Power Generation (KWh)7,856,8877,856,8877,856,8877,856,887.57,856,894.07,856,904.3
7Turbine Operation Mode9.4513.56613.132020
8Generator Speed (rpm)25.06575.92*464.831098.541099.91
9Temperature Outside Nacelle (°C)21.6021.60*21.6021.3020.80
10Nacelle Y-direction Vibration Displacement (m)−0.01098−0.01108*−0.01330−0.02163−0.01448
11Nacelle Z-direction Vibration Displacement (m)−0.01098−0.01108*−0.01331−0.02163−0.01448
12Turbine Brake Level27.62517.085*20.5700
Form description: * represents singular values.
Table 2. The overall information of the 5 datasets.
Table 2. The overall information of the 5 datasets.
DatasetsWPm1WPm2ETTh1ETTh2ETTm1
Variants1212777
Total time steps52,71052,71017,42017,42069,680
Time increment10 min10 min1 h1 h15 min
Start time1 January 20201 January 20201 July 20161 July 20161 July 2016
End time31 December 202031 December 202026 June 201826 June 201826 June 2018
Task typeMulti-stepMulti-stepMulti-stepMulti-stepMulti-step
Data partition5:2:33:1:1
Table 3. Illustration of the ETT data set including 7 variables.
Table 3. Illustration of the ETT data set including 7 variables.
IndexName1 April 2018 0:001 April 2018 0:151 April 2018 0:301 April 2018 0:451 April 2018 1:001 April 2018 1:15
1High Useful Load17.28116.07516.94615.60613.93217.281
2High Useless Load7.3016.9668.5066.7655.4257.301
3Middle Useful Load12.79312.15312.75711.8339.55912.864
4Middle Useless Load5.4374.6916.1484.7623.8735.472
5Low Useful Load4.3564.2644.2644.2034.2954.295
6Low Useless Load1.1271.0051.0661.0051.0971.066
7Oil Temperature9.0049.2159.2869.2159.2159.075
Table 4. Wind power forecasting task setting and number of samples.
Table 4. Wind power forecasting task setting and number of samples.
TaskWPm1WPm2
LengthHorizon61224486122448
look-back128256512512128256512512
Number of samplesTraining24,15022,81020,24820,03224,40023,41821,38821,220
Validation10,86210,72210,44210,39410,86210,72210,44210,394
Test15,03314,62513,80913,71315,04314,63513,81913,723
Table 5. ETT task setting and number of samples.
Table 5. ETT task setting and number of samples.
TaskETTh1ETTh2ETTm1
LengthHorizon24481683367202448168336720244896288672
look-back48963366721440489633667214409696384672672
Number of samplesTraining856984978137763371858569849781377633718534,44134,41734,08133,60133,217
validation285728332713254521612857283327132545216111,49711,47311,42511,23310,849
test285728332713254521612857283327132545216111,49711,47311,42511,23310,849
Table 6. Forecasting results evaluated in MAE on wind power data sets.
Table 6. Forecasting results evaluated in MAE on wind power data sets.
VariableMethodsWPm1WPm2
HorizonHorizon
61224486122448
MultivariateSCINet0.2870.3530.43590.4870.2770.32520.41610.4978
SFINet0.2760.3270.3920.4520.27090.31910.3840.452
RIP3.83%7.37%10.07%7.19%2.20%1.88%7.71%9.20%
UnivariateSCINet0.23510.30590.40320.50710.23190.30280.40590.4978
SFINet0.22910.29120.3790.47210.22410.28510.3780.4733
RIP2.55%4.81%6.00%6.90%3.36%5.85%6.87%4.92%
Table 7. Forecasting results evaluated in WMAPE on wind power data sets.
Table 7. Forecasting results evaluated in WMAPE on wind power data sets.
VariableMethodsWPm1WPm2
HorizonHorizon
61224486122448
MultivariateSCINet31.24%0.381449.95%55.38%30.06%39.69%53.52%65.83%
SFINet30.69%36.44%44.16%50.47%29.08%37.06%48.91%60.97%
AIP0.55%1.70%5.79%4.91%0.98%2.63%4.61%4.86%
UnivariateSCINet29.94%38.88%52.74%66.37%30.79%39.16%52.73%65.32%
SFINet28.84%37.32%49.58%61.44%29.08%37.06%48.91%60.97%
AIP1.10%1.56%3.16%4.93%1.71%2.10%3.82%4.35%
Table 8. Forecasting results evaluated by MAE on ETT datasets. The best results are in bold and the second-best results are underlined. RIP denotes the relative improvement of performance of the proposed method over the second-best results.
Table 8. Forecasting results evaluated by MAE on ETT datasets. The best results are in bold and the second-best results are underlined. RIP denotes the relative improvement of performance of the proposed method over the second-best results.
VariableMethodsETTh1ETTh2ETTm1
HorizonHorizonHorizon
24481683367202448168336720244896288672
MultivariateReformer0.7540.9061.1381.2801.5201.6131.7351.8461.6882.0150.6070.7770.9451.0941.232
LSTMa0.6240.6750.8670.9941.3220.8131.2211.6741.5491.7880.6290.9390.9131.1241.555
LogTrans0.6040.7570.8460.9521.2910.7501.0341.6811.7631.5520.4120.5830.7921.3201.461
Informer0.5490.6250.7520.8730.8960.6651.0011.5151.3401.4730.3690.5030.6140.7860.926
TCN0.5490.5290.6170.6820.7780.4780.6151.2661.3121.2760.2820.3600.3630.6461.371
SCINet0.3790.3950.4570.4970.5600.2880.3580.5040.5600.7610.2290.2740.2910.4150.604
SFINet0.3500.3700.4240.4650.5180.2680.3310.4090.5030.6570.2220.2560.2780.3680.461
RIP7.65%6.33%7.22%6.44%7.50%6.94%7.54%18.83%10.18%13.67%3.06%6.57%4.47%11.08%23.68%
UnivariateReformer0.3890.4451.1911.1241.4360.4370.5450.8791.2281.7210.2280.3900.7671.2451.528
LSTMa0.2750.3300.7631.8203.2530.3810.4621.0682.5434.6640.2900.3050.3960.5741.174
LogTrans0.2590.3280.3750.3980.4630.2550.3480.4220.4370.4930.2020.2200.3860.5720.702
Informer0.2470.3190.3460.3870.4350.2400.3140.3890.4170.4310.1370.2030.3720.5540.644
N-Beats0.1560.2000.2550.2840.4220.2100.2710.3930.4180.4320.1170.1680.2340.3110.370
SCINet0.1320.1730.2220.2420.3430.1940.2420.3110.3400.4030.0850.1340.1980.2660.328
SFINet0.1310.1630.2030.2400.3060.1830.2290.2970.3340.3490.0880.1400.1830.2500.306
RIP0.76%5.78%5.58%0.83%10.79%3.09%5.37%4.50%1.76%13.40%1.14%2.10%7.58%6.02%6.71%
Table 9. Forecasting results evaluated by WMAPE on ETT datasets.
Table 9. Forecasting results evaluated by WMAPE on ETT datasets.
VariableMethodsETTh1ETTh2ETTm1
HorizonHorizonHorizon
24481683367202448168336720244896288672
MultivariateSCINet47.10%49.92%57.15%61.57%70.97%20.83%26.19%36.79%41.73%56.88%17.51%19.98%23.43%30.45%43.96%
SFINet44.22%46.53%52.88%58.67%64.84%19.72%23.82%30.05%37.41%49.12%16.38%18.44%20.48%27.14%34.00%
AIP2.88%3.39%4.27%2.90%6.13%1.11%2.37%6.74%4.32%7.76%1.13%1.54%2.95%3.31%9.96%
UnivariateSCINet10.86%16.62%19.36%22.36%26.51%18.66%22.85%29.93%33.64%45.33%8.16%13.20%18.31%23.96%28.82%
SFINet9.82%12.34%15.09%17.61%22.18%17.19%20.65%26.64%32.83%30.79%7.73%12.74%16.60%22.54%27.35%
AIP1.04%4.28%4.27%4.75%4.33%1.47%2.20%3.29%0.81%14.54%0.43%0.46%1.71%1.42%1.47%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, X.; Xiao, C.; Zhang, T. SFINet: Shuffle–and–Fusion Interaction Networks for Wind Power Forecasting. Appl. Sci. 2022, 12, 2253. https://doi.org/10.3390/app12042253

AMA Style

Zhang X, Xiao C, Zhang T. SFINet: Shuffle–and–Fusion Interaction Networks for Wind Power Forecasting. Applied Sciences. 2022; 12(4):2253. https://doi.org/10.3390/app12042253

Chicago/Turabian Style

Zhang, Xu, Cheng Xiao, and Tieling Zhang. 2022. "SFINet: Shuffle–and–Fusion Interaction Networks for Wind Power Forecasting" Applied Sciences 12, no. 4: 2253. https://doi.org/10.3390/app12042253

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop