Next Article in Journal
Target Recognition in SAR Images by Deep Learning with Training Data Augmentation
Next Article in Special Issue
Secondary Interface Echoes Suppression for Immersion Ultrasonic Imaging Based on Phase Circular Statistics Vector
Previous Article in Journal
Construction Site Safety Management: A Computer Vision and Deep Learning Approach
Previous Article in Special Issue
Surface Decarburization Depth Detection in Rods of 60Si2Mn Steel with Magnetic Barkhausen Noise Technique
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Solar Power Prediction Using Dual Stream CNN-LSTM Architecture

1
Department of Electrical Engineering, Unaizah College of Engineering, Qassim University, Unaizah 56452, Saudi Arabia
2
Department of Information Technology, College of Computer, Qassim University, Buraydah 51452, Saudi Arabia
3
Department of Electrical Engineering, College of Engineering and Information Technology, Onaizah Colleges, Onaizah 56447, Saudi Arabia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(2), 945; https://doi.org/10.3390/s23020945
Submission received: 28 November 2022 / Revised: 4 January 2023 / Accepted: 6 January 2023 / Published: 13 January 2023
(This article belongs to the Special Issue Advanced Sensing and Evaluating Technology in Nondestructive Testing)

Abstract

:
The integration of solar energy with a power system brings great economic and environmental benefits. However, the high penetration of solar power is challenging due to the operation and planning of the existing power system owing to the intermittence and randomicity of solar power generation. Achieving accurate predictions for power generation is important to provide high-quality electric energy for end-users. Therefore, in this paper, we introduce a deep learning-based dual-stream convolutional neural network (CNN) and long short-term nemory (LSTM) network followed by a self-attention mechanism network (DSCLANet). Here, CNN is used to learn spatial patterns and LSTM is incorporated for temporal feature extraction. The output spatial and temporal feature vectors are then fused, followed by a self-attention mechanism to select optimal features for further processing. Finally, fully connected layers are incorporated for short-term solar power prediction. The performance of DSCLANet is evaluated on DKASC Alice Spring solar datasets, and it reduces the error rate up to 0.0136 MSE, 0.0304 MAE, and 0.0458 RMSE compared to recent state-of-the-art methods.

1. Introduction

Regarding solar energy generation, sustainable development and global climate change are the two main issues [1]. Each year energy consumption is increased by 2% globally, where the total energy production is significantly based on fossil fuels, such as natural gas, coal, and oil, which considerably increases anthropogenic greenhouse gas (GHG) emission [2,3]. Furthermore, power generation from fuels produces environmental risks and energy crises, such as energy resource reduction and an increase in environmental pollution, which is considered a major threat to lives [4,5,6]. These drawbacks of energy generation from fossil fuels force governments to explore the resources of renewable energies [6,7].
Solar power is considered the alternative when compared to fossil fuels due to various characteristics, such as being clean, green, and naturally replenished. Solar power generation, either as an islanded or grid-connected mode of operation, brings unstable uncertainty, which causes problems for the stability of the power systems, particularly for the integration of solar power in a large microgrid system [8,9]. To overcome these challenges a reliable solar power prediction is an effective way to decrease the uncertainty, which is important for the planning, management, and operation of energy systems [10]. Therefore, the researchers investigated several techniques for solar power prediction. These techniques are broadly categorized into statistical (ST), artificial intelligence (AI), and hybrid methods (HM) [11]. In ST-based methods, several algorithms are developed, including auto-regressive [12], Bayesian [13], Kalman [14], grey models [15,16], and the Markov chain model [17]. Additionally, MaatAllah et al. [18] and Reikard et al. [19] developed ST-based models for renewable power prediction. In contrast, statistical models rely on linear data for learning but are unable to learn complex data; therefore, ST-based methods are not recommended for problems requiring nonlinear predictions, such as those associated with solar power.
Due to their potential for extracting representative features and data mining, AI-based models have proven to be more successful than physical and statistical ones [20]. These AI-based methods developed in the literature for solar power generation include neural networks [21], SVR [22], the adaptive fuzzy approach [23], and ELM [24], etc. Unlike ST-based approaches, most of these AI-based approaches are used to manage nonlinear relationships between input and output. Additionally, in the literature on power generation prediction, some special AI-based models, such as those models based on CNNs and generative adversarial networks, were developed by [25], and it became evident that weather classification played a significant role in developing such an accurate model. Furthermore, a number of AI-based approaches, including RNN [26], LSTM [27], CNN [28], GRU [29], etc., have been developed by the researchers for solar power generation, where the details are given in a recent survey [30]. This survey [30] also concluded that due to balancing parameter stability with accuracy, and their pros and cons, hybrid models are effective for solar power prediction. These AI-based methods are constructed via shallow architecture, requiring handcrafted feature engineering and having limited generalization capabilities [31]. Furthermore, in AI-based methods, CNN and RNNs achieved better performance; however, using CNN, the feature is extracted in spatial dimensions [32,33,34], while the RNNS learns in temporal dimensions, while solar power generation includes both types of features. Therefore, an approach with the ability of spatial and temporal feature extraction is required for accurate solar power prediction.
Table 1. Summary of hybrid methods developed for power generation prediction.
Table 1. Summary of hybrid methods developed for power generation prediction.
Ref.MethodComparisonSummary
Agoua et al. [35]Spatiotemporal networkAuto-regression and decision treeA spatiotemporal network is developed for learning spatial and temporal information.
Gensler et al. [36]Auto-LSTMMLP, ANN, LSTM, DNN, DBNDeveloped an LSTM- and MLP-based hybrid model.
Sorkun et al. [37]LSTMLSTM, naive, GRU, RNN, and LSTMDeveloped an LSTM-based method for power generation forecasting.
Khan et al. [38]CNNESNLSTM, GRU, ESNA combined CNN- and ESN-based model is developed.
Dey et al. [39]SolarNetGaussian regression, SVR, ANNA CNN-based model for power generation prediction is developed.
Abdel et al. [26]LSTMRNNANN and regressionA RNN-LSTM-based hybrid model is developed.
Khan et al. [38]CNNESNSVR, decision tree, CNN, LSTMA combined CNN- and ESN-based model is developed.
Yan et al. [40]CNN-GRULSTM and GRUA combined inception and GRU model.
Dong et al. [41]chaotic hybrid CNN model CNN-based ablation studyThe performance of a CNN-based model was developed and improved their performance with the use of a chaotic hybrid model.
Khan et al. [7]ESN-CNNDetailed ablation studyIntegrated ESN and CNN for power generation prediction
In the light of current literature, hybrid models achieved state-of-the-art accuracy for solar power prediction [38]. These models include CNN-RNN [42], CNN-GRU [43], CNN-LSTM [44], CNNLSTM with autoencoder [45], convolutional LSTM (CLSTM) [46], CNN-GRU with preprocessing [45,47], and LSTM-CNN [48]. Some recent hybrid models for renewable power generation prediction are summarized in Table 1. Hybrid methods achieved improved prediction performance compared to other predictive modeling techniques. However, the current literature focuses on the stacked layers procedure to develop a hybrid model for solar power prediction where historical data of solar power have a limited number of features, which makes it difficult to learn spatial and temporal features using the stacked layers phenomena. Furthermore, prediction accuracy needs to be improved for reliable and accurate solar energy prediction. Therefore, in this work, we developed DSCLANet for solar power prediction with the ability to learn spatial and temporal features parallelly from actual solar power and weather data. The first stream of the proposed network utilizes CNN for spatial feature extraction, while the second stream is responsible for temporal feature extraction. Finally, the outcome of these streams is concatenated and passed to fully connected layers for solar energy prediction. The performance of the proposed model is evaluated on benchmark datasets and extensively decreases the error rates compared to state-of-the-art models. The following are the main contributions of this work:
  • To select the most suitable model for solar power prediction, an ablation study is conducted, where the main objective is to evaluate the performance of several techniques including CNN, LSTM, GRU, CNNLSTM, CNNGRU, and DSCLANet to select an accurate prediction model for solar power.
  • Our findings from this ablation study indicate that DSCLANet gives the best prediction accuracy comparatively, which has been confirmed experimentally by various comparisons. The DSCLANet process is the input via separate streams for spatial and temporal features which are then fused and passed to the attention for feature refinement. The refined features are then forwarded to a fully connected layer for final solar power prediction.
  • A number of benchmark datasets are utilized to assess the DSCLANet performance, and the results indicate a marginal reduction in error rates compared to other state-of-the-art methods.
  • The remainder of this article is organized as follows. Section 2 describes the internal architecture of DSCLANet, and Section 3 defines the datasets, evaluation metrics, and performance comparison of DSCLANet with ablation study and baseline methods. Finally, this article is concluded in Section 4, with possible future directions.

2. Materials and Methods

The main framework of the DSCLANet is shown in Figure 1. where the input data is parallelly processed using CNN and LSTM architecture to extract spatiotemporal information. The output of these two architectures is then fused and fed to the attention stage for feature refinement and, finally, to the fully-connected layers for prediction. The internal architecture of the proposed model is further described in the following subsection.

2.1. CNN-LSTM

Dual CNN-LSTM architecture integrates CNN and LSTM for solar energy prediction. The proposed model has the ability to store the irregular complex trend and can extract complex features from historical solar power generation data. The first stream is incorporated to extract spatial features via CNN from the input data, while the second stream is responsible for temporal features extraction using LSTM. The CNN is a well-known deep learning architecture consisting of four types of layers, namely convolutional, pooling, fully connected, and regression layers [49]. The convolutional layers include multiple convolution filters which perform convolutional operations between convolutional neuron weights and input volume connected regions which generate a feature map [50,51]. The LSTM architecture is responsible for storing time information about important characteristics of solar power data. It supplies a solution by maintaining log-term memory by merging memory units that can update the previous hidden state [52]. With this function, it will be easier to understand temporal relationships in a long-term sequence. In this case, gate units receive the output values from the preceding CNN layer. The LSTM network addresses vanishing and explosive gradient problems that can happen when learning basic RNNs. The three gates unit’s mechanism can be used for determining the state of each individual memory cell. The input, output, and forget gates represent the gate unit. The mathematical of an LSTM from input to output generation is given in Equations (1)–(6).
ƒ t = Φ   Ŵ f     h t 1 ,   x t + B f  
i t = Φ   Ŵ i     h t 1 ,   x t + B i
Ċ t = t a n h   Ŵ C     h t 1 ,   x t + B C
C t = f t   x   C t 1 + i t   x   Ċ t  
o t = Φ Ŵ o     h t 1 ,   x t + B o
h t = o t   x tanh ( Φ C t .
where x t is the input, hidden layer output is represented by h t , Φ is the sigmoid function, and C t is the cell state, while its state candidate is represented by Ŵ i , Ŵ o , Ŵ f , and Ŵ C , which are the input, output, forget gate, and memory cells weights, respectively, while B i , B o , B f , and B c are the bias terms for the input, output, forget gate, and cell, respectively. Finally, the output of CNN and LSTM streams are then fused with a concatenation layer and faded to attention layers for further processing.

2.2. Attention Mechanism

The final output of deep learning architectures named (CNN and LSTM) are integrated to obtain a single feature vector, and then fed the output streams to the self-attention SA mechanism to determine a representative feature vector for final forecasting. In addition, the invisible detail at different timestamps has a high impact on final results, but the CNN and LSTM streams are unable to predict forecasting accurately. To cope with these issues, our work is focused on integrating the SA architecture which has the capability to strengthen dominant and undermine trivial details by adaptively weighting the hidden features. In this paper, we utilized the SA architecture for the recognition of dominant features; in this regard, the combined feature vector of CNN and LSTM streams is used as an input to the SA network before forecasting. Moreover, the correlation of the proposed architecture at different timestamps among hidden features is investigated from every dimension. The calculation of the hidden features score, such as the k t h timestamp and N t h dimension, is based on Equation (7), as follows:
S J ,   d = f i ( w k ,   n h 1 ,   n ,   h 2 ,   n , h 3 ,   n , h n ,   k ,   N = 1 ,   2 , 3 n ,   k = 1 , 2 , 3 n i
where g k ,   n indicates the d t h dimension of the invisible state at k t h timestamp, whereas the weight matrix, such as w k ,   n , f i is a function applied using dense layers, and n and n i describe the number of timestamps and hidden feature dimensions, respectively.
The proposed network also contains dense layers, which are utilized to forecast power (PV) for a certain period of time, for instance an hour ahead of the PV power forecasting. The final output of the SA architecture is flattened to a Z i = z 1   ,   z 2   ,   z 3   . z n   feature vector, whereas i represents the output dimensions of the proposed model. The output of the S-AM architecture is fed to the fully connected layers as an input, where the mathematical form of these layers is presented as follows in Equation (9):
Z i l = j   w j i l 1 ( x X   i l 1 + b j l 1
where w j i l 1 indicates a weigh metric, x describes the activation function, namely the X   i l 1 input data in this equation, while B j l 1 represents the bias term.

2.3. DSCLANet Archatecture

The architecture of DSCLANet includes CNN, LSTM, attention, and fully connected layers. Optimal DSCLANet architecture is developed by adjusting various parameters, including the size of the filter for CNN, the size of the kernel, the size of the LSTM cell, etc. Several experiments are conducted to choose the optimal parameters for the model before finalizing its internal parameters. The two streams allow for the parallel extraction of spatiotemporal features from large data sets, which are inputs to both streams. The CNN stream includes three CNN layers, while the LSMT stream includes two LSTM layers for each type of feature extraction. A concatenation layer is then applied to the output of both streams, followed by a feature-attentional layer and fully connected layers. The internal architecture of DSCLANet in terms of number of parameters, filters, and kernels is given in Table 2. In the first stream, the hyper-parameters of CNN layer 1 are as follows: the filter size is set to 32, with a kernel size of 5, padding is set to the same, the stride is set to 1, with default valid padding, and we used ReLU as the activation function. In the second CNN layer, the filter size is set to 64 with a kernel size of 3 while other hyper-parameters are the same as CNN layer 1. Furthermore, in the third CNN layer, the filter size is set to 128 while the kernel size of 1 is used. Other hyper-parameters of CNN layer 3 are the same as CNN layer 1. In the second stream, two LSTM layers are used with the same cell size of 100. These streams are then concatenated with a fusion layer, and the output is forwarded to the attention layer. The combined feature vector from both streams of the network includes redundant information, making the network computationally expensive, leading to non-convergence of the network, and achieving limited performance. Thus, the attention layer is used to enable the network to remove the redundant information and to enable the network to focus on important information while ignoring the rest of the information, which leads to fast convergence of the network and achieves considerable performance. This optimal feature is then passed to a fully connected layer for the final prediction, where 3 fully connected layers of sizes 64, 32, and 12 are used in DSCLANet.

3. Results

This section delivers a comprehensive discussion about evaluation metrics, datasets, and experimental results. The experiments are conducted in the Keras framework with a backend TensorFlow, utilizing a GeForce RTX 2070 graphics card.

3.1. Evaluation Metrics

The performance of the DSCLANet is assessed on standard evaluation metrics, such as MAE, MBE, RMSE, and MSE. These are common metrics used in the literature to evaluate the forecasting performance of solar power prediction models. The MAE is the average absolute difference between actual and predicted values, and MBE indicates the average difference between these values. The MSE is the square difference between predicted and actual data, while RMSE is the square root of MSE. The mathematical equation of these metrics is given in Equations (9)–(12), as follows:
M A E = n = 1 m A n P n N
M B E = n = 1 m P n A n   N
M S E = n = 1 m A n P n 2 N
R M S E = n = 1 m A n P n 2 N
where A represents the actual and P represents the predicted values by the model.

3.2. Datasets

In this work, we utilized DKASC Alice Spring DKASC-AS datasets to evaluate the performance of the proposed and other models. Three datasets are selected from DKASC-AS, namely Trina 10.5 kW mono-Si Dual 2009 (Trina 1A), Trina 23.4 kW mono-Si Dual 2009 (Trina 1B), and eco-Kinetics 26.5 kW mono-Si Dual 2010 (Eco 2). These datasets include historical weather and solar power generation data with different generation capacities installed on different dates. Detailed information of the datasets, such as installation date, number of panels, type of panel, etc., are available of the DKASC website [53]. All the datasets are split into 70%, 20%, and 10% training, testing, and validation data, respectively. The proposed model and other ablation study models are evaluated using two-hour historical data as input to predict one hour ahead power generation.

3.3. Performance Evaluation of Deep Learning-Based Models

To substantiate the robustness of the proposed DSCLANet, we conducted experiments on several models based on deep learning. These models include LSTM, CNN, GRU, CNNGRU, CNNLSTM, and DCNN-BRLSTM. The results attained by each model for every dataset is demonstrated in Table 3. For instance, LSTM achieved 0.0804 MSE, 0.143 MAE, and 0.2836 RMSE over the Trina 1A dataset, while these values were 0.0767, 0.1473, and 0.2769, and 0.0416, 0.1069, and 0.2041 over the Trina 1B and Eco 2 datasets, respectively. The CNN achieved 0.0699 MSE, 0.1526 MAE, and 0.3108 RMSE over the Trina 1A dataset, 0.1196 MSE, 0.2041 MAE, and 0.3458 RMSE over the Trina 1B dataset, and 0.0433 MSE, 0.1288 MAE, and 0.2081 for the RMSE Eco 2 dataset. Furthermore, GRU attained 0.0848 MSE, 0.1518 MAE, and 0.2912 RMSE over the Trina 1A dataset, 0.065 MSE, 0.1196 MAE, and 0.2549 RMSE over the Trina 1B dataset, and 0.0384 MSE, 0.1011 MAE, and 0.196 RMSE over the Eco 2 dataset. Compared to the output of these models’ hybrid models, such as CNNGRU and CNNLSTM, DSCLANet achieved better prediction results due to learning both spatiotemporal information from historical data. For instance, CNNLSTM achieved 0.0679 MSE, 0.12 MAE, and 0.2606 RMSE over the Trina 1A dataset, 0.0648 MSE, 0.131 MAE, and 0.2546 RMSE over the Trina 1B dataset, and 0.0298 MSE, 0.088 MAE, and 0.1725 RMSE over the Eco 2 dataset. Similarly, CNNGRU achieved (0.0793, 0.01519, and 0.2817), (0.0641, 0.1365, and 0.2531), and (0.032, 0.0879, and 0.1789) values for the Trina 1A, Trina 1B, and Eco 2 datasets, respectively. The proposed DSCLANet further reduces the error metrics and achieved the lowest error rate as compared to the abovementioned models. The proposed DSCLANet achieved 0.0167 MSE, 0.0632 MAE, and 0.1291 RMSE over the Trina 1A dataset, 0.0279 MSE, 0.0889 MAE, and 0.167 RMSE over the Trina 1B dataset, and 0.0074 MSE, 0.0479 MAE, and 0.0858 RMSE over the Eco 2 dataset. Furthermore, the actual and predicted results of DSCLANet over each dataset are given in Figure 2.

3.4. Comparison with State-of-the-Art

In this section, we compared the performance of DSCLANet with other baselines. The performance of the proposed approach is compared with the wavelet packet decomposition (WPD-LSTM) [54], RCC-LSTM [55], HIMVO-SVM [56], ESN-CNN [7], CNN-LSTM [57], DenseNet [28], LSTM-CNN [48], ELM [58], graph-network [59], and SolarNet [60] models. The detailed performance of these models is given in Table 4, where the DSCLANet attained the smallest error rates comparatively. The DKASC Alice Spring sites include several solar power plants, and the researcher evaluated their model performance over one, two, or three sites’ data. Therefore, in this work, we compared the average performance of DSCLANet for three sites’ data, namely Trina 1A, Trina 1B, and Eco 2, with these methods. Comparatively, the DSCLANet achieved a better performance in all error metrics, as shown in Table 4.

4. Conclusions

It is important to forecast solar power generation accurately to avoid penalties from customers, build trust in the energy markets, and schedule power generation. In mainstream deep learning and traditional learning methods, features are based on simple phenomena, and they only take into account spatial or temporal features to get around the nonlinearities of solar power generation series. However, some studies combine different methods for spatial and temporal feature extraction via a stacked layers mechanism. Therefore, in this work, we developed a dual-stream CNN-LSTM network for solar power prediction. The performance of DSCLANet is evaluated for real solar power datasets collected from a photovoltaic system located in Alice Springs, Australia. Before selecting the proposed model, extensive experiments are performed over different deep learning-based models. Furthermore, we compared the performance of the DSCLANet with other baselines and found that the proposed model outperforms them in terms of error reduction. Alongside higher performance, the DSCLANet uses two architectures, namely LSTM and CNN, for spatial and temporal feature extraction. However, combining multiple methods for spatial and temporal feature extraction increases the model complexity. Therefore, in the near future, we intend to develop a solo architecture with the ability to extract both types of features. Furthermore, we also intend to investigate emerging technologies, such as probabilistic forecasting, incremental learning, active learning, and reinforcement learning for solar power prediction.

Author Contributions

Conceptualization, H.A. and S.H.; methodology, H.A. and M.I.; software, S.H. and M.I.; validation, S.H. and H.A.; formal analysis, M.I. and S.H.; investigation, S.H. and H. A; resources, S.H., H.A. and M.I.; data curation, M.I. writing—original draft preparation, H.A and S.H.; writing—review and editing, M.I.; visualization, H.A.; supervision, S.H and M.I.; project administration, S.H.; funding acquisition, H.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The researchers would like to thank the Deanship of Scientific Research, Qassim University, for funding the publication of this project.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nam, K.; Hwangbo, S.; Yoo, C. A deep learning-based forecasting model for renewable energy scenarios to guide sustainable energy policy: A case study of Korea. Renew. Sustain. Energy Rev. 2020, 122, 109725. [Google Scholar] [CrossRef]
  2. Foster, E.; Contestabile, M.; Blazquez, J.; Manzano, B.; Workman, M.; Shah, N. The unstudied barriers to widespread renewable energy deployment: Fossil fuel price responses. Energy Policy 2017, 103, 258–264. [Google Scholar] [CrossRef]
  3. Wang, H.; Lei, Z.; Zhang, X.; Zhou, B.; Peng, J. A review of deep learning for renewable energy forecasting. Energy Convers. Manag. 2019, 198, 111799. [Google Scholar] [CrossRef]
  4. Aladhadh, S.; Almatroodi, S.A.; Habib, S.; Alabdulatif, A.; Khattak, S.U.; Islam, M. An Efficient Lightweight Hybrid Model with Attention Mechanism for Enhancer Sequence Recognition. Biomolecules 2023, 13, 70. [Google Scholar] [CrossRef]
  5. Alsharekh, M.F.; Habib, S.; Dewi, D.A.; Albattah, W.; Islam, M.; Albahli, S. Improving the Efficiency of Multistep Short-Term Electricity Load Forecasting via R-CNN with ML-LSTM. Sensors 2022, 22, 6913. [Google Scholar] [CrossRef] [PubMed]
  6. Yar, H.; Imran, A.S.; Khan, Z.A.; Sajjad, M.; Kastrati, Z. Towards smart home automation using IoT-enabled edge-computing paradigm. Sensors 2021, 21, 4932. [Google Scholar] [CrossRef] [PubMed]
  7. Khan, Z.A.; Hussain, T.; Haq, I.U.; Ullah, F.U.M.; Baik, S.W. Towards efficient and effective renewable energy prediction via deep learning. Energy Rep. 2022, 8, 10230–10243. [Google Scholar] [CrossRef]
  8. Khan, Z.A.; Ullah, A.; Haq, I.U.; Hamdy, M.; Maurod, G.M.; Muhammad, K.; Hijji, M.; Baik, S.W. Efficient short-term electricity load forecasting for effective energy management. Sustain. Energy Technol. Assess. 2022, 53, 102337. [Google Scholar] [CrossRef]
  9. Khan, F.A.; Shees, M.M.; Alsharekh, M.F.; Alyahya, S.; Saleem, F.; Baghel, V.; Sarwar, A.; Islam, M.; Khan, S. Open-Circuit Fault Detection in a Multilevel Inverter Using Sub-Band Wavelet Energy. Electronics 2021, 11, 123. [Google Scholar] [CrossRef]
  10. Frías-Paredes, L.; Mallor, F.; Gastón-Romeo, M.; León, T. Assessing energy forecasting inaccuracy by simultaneously considering temporal and absolute errors. Energy Convers. Manag. 2017, 142, 533–546. [Google Scholar] [CrossRef]
  11. Hussain, T.; Min Ullah, F.U.; Muhammad, K.; Rho, S.; Ullah, A.; Hwang, E.; Moon, J.; Baik, S.W. Smart and intelligent energy monitoring systems: A comprehensive literature survey and future research guidelines. Int. J. Energy Res. 2021, 45, 3590–3614. [Google Scholar] [CrossRef]
  12. Habib, S.; Alyahya, S.; Islam, M.; Alnajim, A.M.; Alabdulatif, A.; Alabdulatif, A. Design and Implementation: An IoT-Framework-Based Automated Wastewater Irrigation System. Electronics 2023, 12, 28. [Google Scholar] [CrossRef]
  13. Zuhaib, M.; Shaikh, F.A.; Tanweer, W.; Alnajim, A.M.; Alyahya, S.; Khan, S.; Usman, M.; Islam, M.; Hasan, M.K. Faults Feature Extraction Using Discrete Wavelet Transform and Artificial Neural Network for Induction Motor Availability Monitoring—Internet of Things Enabled Environment. Energies 2022, 15, 7888. [Google Scholar] [CrossRef]
  14. Yang, D. On post-processing day-ahead NWP forecasts using Kalman filtering. Sol. Energy 2019, 182, 179–181. [Google Scholar] [CrossRef]
  15. Wu, L.; Gao, X.; Xiao, Y.; Yang, Y.; Chen, X. Using a novel multi-variable grey model to forecast the electricity consumption of Shandong Province in China. Energy 2018, 157, 327–335. [Google Scholar] [CrossRef]
  16. Muhammad, T.; Khan, A.U.; Chughtai, M.T.; Khan, R.A.; Abid, Y.; Islam, M.; Khan, S. An Adaptive Hybrid Control of Grid Tied Inverter for the Reduction of Total Harmonic Distortion and Improvement of Robustness against Grid Impedance Variation. Energies 2022, 15, 4724. [Google Scholar] [CrossRef]
  17. Wang, Y.; Wang, J.; Wei, X. A hybrid wind speed forecasting model based on phase space reconstruction theory and Markov model: A case study of wind farms in northwest China. Energy 2015, 91, 556–572. [Google Scholar] [CrossRef]
  18. Maatallah, O.A.; Achuthan, A.; Janoyan, K.; Marzocca, P. Recursive wind speed forecasting based on Hammerstein Auto-Regressive model. Appl. Energy 2015, 145, 191–197. [Google Scholar] [CrossRef]
  19. Khan, K.; Khan, R.U.; Albattah, W.; Nayab, D.; Qamar, A.M.; Habib, S.; Islam, M. Crowd Counting Using End-to-End Semantic Image Segmentation. Electronics 2021, 10, 1293. [Google Scholar] [CrossRef]
  20. Daut, M.A.M.; Hassan, M.Y.; Abdullah, H.; Rahman, H.A.; Abdullah, M.P.; Hussin, F. Building electrical energy consumption forecasting analysis using conventional and artificial intelligence methods: A review. Renew. Sustain. Energy Rev. 2017, 70, 1108–1118. [Google Scholar] [CrossRef]
  21. Wang, J.; Zhang, N.; Lu, H. A novel system based on neural networks with linear combination framework for wind speed forecasting. Energy Convers. Manag. 2019, 181, 425–442. [Google Scholar] [CrossRef]
  22. Deo, R.C.; Wen, X.; Qi, F. A wavelet-coupled support vector machine model for forecasting global incident solar radiation using limited meteorological dataset. Appl. Energy 2016, 168, 568–593. [Google Scholar] [CrossRef]
  23. Sharifian, A.; Ghadi, M.J.; Ghavidel, S.; Li, L.; Zhang, J. A new method based on Type-2 fuzzy neural network for accurate wind power forecasting under uncertain data. Renew. Energy 2018, 120, 220–230. [Google Scholar] [CrossRef]
  24. Ali, M.; Prasad, R. Significant wave height forecasting via an extreme learning machine model integrated with improved complete ensemble empirical mode decomposition. Renew. Sustain. Energy Rev. 2019, 104, 281–295. [Google Scholar] [CrossRef]
  25. Momin, A.M.; Ahmad, I.; Islam, M. Weed Classification Using Two Dimensional Weed Coverage Rate (2D-WCR) for Real-Time Selective Herbicide Applications. In Proceedings of the International Conference on Computing, Information and Systems Science and Engineering, Bangkok, Thailand, 29–31 January 2007. [Google Scholar]
  26. Abdel-Nasser, M.; Mahmoud, K. Accurate photovoltaic power forecasting models using deep LSTM-RNN. Neural Comput. Appl. 2019, 31, 2727–2740. [Google Scholar] [CrossRef]
  27. Zhang, J.; Chi, Y.; Xiao, L. Solar power generation forecast based on LSTM. In Proceedings of the 2018 IEEE 9th International Conference on Software Engineering and Service Science (ICSESS), Beijing, China, 23–25 November 2018; pp. 869–872. [Google Scholar]
  28. Zang, H.; Cheng, L.; Ding, T.; Cheung, K.W.; Wei, Z.; Sun, G. Day-ahead photovoltaic power forecasting approach based on deep convolutional neural networks and meta learning. Int. J. Electr. Power Energy Syst. 2020, 118, 105790. [Google Scholar] [CrossRef]
  29. Han, T.; Muhammad, K.; Hussain, T.; Lloret, J.; Baik, S.W. An efficient deep learning framework for intelligent energy management in IoT networks. IEEE Internet Things J. 2020, 8, 3170–3179. [Google Scholar] [CrossRef]
  30. Habib, S.; Alsanea, M.; Aloraini, M.; Al-Rawashdeh, H.S.; Islam, M.; Khan, S. An Efficient and Effective Deep Learning-Based Model for Real-Time Face Mask Detection. Sensors 2022, 22, 2602. [Google Scholar] [CrossRef]
  31. Yar, H.; Hussain, T.; Khan, Z.A.; Koundal, D.; Lee, M.Y.; Baik, S.W. Vision sensor-based real-time fire detection in resource-constrained IoT environments. Comput. Intell. Neurosci. 2021, 2021, 5195508. [Google Scholar] [CrossRef]
  32. Khan, Z.A.; Hussain, T.; Ullah, F.U.M.; Gupta, S.K.; Lee, M.Y.; Baik, S.W. Randomly Initialized CNN with Densely Connected Stacked Autoencoder for Efficient Fire Detection. Eng. Appl. Artif. Intell. 2022, 116, 105403. [Google Scholar] [CrossRef]
  33. Yar, H.; Hussain, T.; Agarwal, M.; Khan, Z.A.; Gupta, S.K.; Baik, S.W. Optimized Dual Fire Attention Network and Medium-Scale Fire Classification Benchmark. IEEE Trans. Image Process. 2022, 31, 6331–6343. [Google Scholar] [CrossRef] [PubMed]
  34. Yar, H.; Hussain, T.; Khan, Z.A.; Lee, M.Y.; Baik, S.W. Fire Detection via Effective Vision Transformers. J. Korean Inst. Next Gener. Comput. 2021, 17, 21–30. [Google Scholar]
  35. Albattah, W.; Habib, S.; Alsharekh, M.F.; Islam, M.; Albahli, S.; Dewi, D.A. An Overview of the Current Challenges, Trends, and Protocols in the Field of Vehicular Communication. Electronics 2022, 11, 3581. [Google Scholar] [CrossRef]
  36. Gensler, A.; Henze, J.; Sick, B.; Raabe, N. Deep Learning for solar power forecasting—An approach using AutoEncoder and LSTM Neural Networks. In Proceedings of the 2016 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Budapest, Hungary, 9–12 October 2016; pp. 002858–002865. [Google Scholar]
  37. Sorkun, M.C.; Paoli, C.; Incel, Ö.D. Time series forecasting on solar irradiation using deep learning. In Proceedings of the 2017 10th International Conference on Electrical and Electronics Engineering (ELECO), Bursa, Turkey, 30 November–2 December 2017; pp. 151–155. [Google Scholar]
  38. Khan, Z.A.; Hussain, T.; Baik, S.W. Boosting energy harvesting via deep learning-based renewable power generation prediction. J. King Saud Univ.-Sci. 2022, 34, 101815. [Google Scholar] [CrossRef]
  39. Dey, S.; Pratiher, S.; Banerjee, S.; Mukherjee, C.K. Solarisnet: A deep regression network for solar radiation prediction. arXiv 2017, arXiv:1711.08413. [Google Scholar]
  40. Yan, K.; Shen, H.; Wang, L.; Zhou, H.; Xu, M.; Mo, Y. Short-term solar irradiance forecasting based on a hybrid deep learning methodology. Information 2020, 11, 32. [Google Scholar] [CrossRef] [Green Version]
  41. Dong, N.; Chang, J.-F.; Wu, A.-G.; Gao, Z.-K. A novel convolutional neural network framework based solar irradiance prediction method. Int. J. Electr. Power Energy Syst. 2020, 114, 105411. [Google Scholar] [CrossRef]
  42. Kim, J.; Moon, J.; Hwang, E.; Kang, P. Recurrent inception convolution neural network for multi short-term load forecasting. Energy Build. 2019, 194, 328–341. [Google Scholar] [CrossRef]
  43. Sajjad, M.; Khan, Z.A.; Ullah, A.; Hussain, T.; Ullah, W.; Lee, M.Y.; Baik, S.W. A novel CNN-GRU-based hybrid approach for short-term residential load forecasting. IEEE Access 2020, 8, 143759–143768. [Google Scholar] [CrossRef]
  44. Qu, J.; Qian, Z.; Pei, Y. Day-ahead hourly photovoltaic power forecasting using attention-based CNN-LSTM neural network embedded with multiple relevant and target variables prediction pattern. Energy 2021, 232, 120996. [Google Scholar] [CrossRef]
  45. Khan, Z.A.; Hussain, T.; Ullah, A.; Rho, S.; Lee, M.; Baik, S.W. Towards efficient electricity forecasting in residential and commercial buildings: A novel hybrid CNN with a LSTM-AE based framework. Sensors 2020, 20, 1399. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Wang, F.; Yu, Y.; Zhang, Z.; Li, J.; Zhen, Z.; Li, K. Wavelet decomposition and convolutional LSTM networks based improved deep learning model for solar irradiance forecasting. Appl. Sci. 2018, 8, 1286. [Google Scholar] [CrossRef]
  47. Khan, Z.A.; Ullah, A.; Ullah, W.; Rho, S.; Lee, M.; Baik, S.W. Electrical energy prediction in residential buildings for short-term horizons using hybrid deep learning strategy. Appl. Sci. 2020, 10, 8634. [Google Scholar] [CrossRef]
  48. Wang, K.; Qi, X.; Liu, H. Photovoltaic power forecasting based LSTM-Convolutional Network. Energy 2019, 189, 116225. [Google Scholar] [CrossRef]
  49. Zhao, X.; Wei, H.; Wang, H.; Zhu, T.; Zhang, K. 3D-CNN-based feature extraction of ground-based cloud images for direct normal irradiance prediction. Sol. Energy 2019, 181, 510–518. [Google Scholar] [CrossRef]
  50. Wang, F.; Li, K.; Duić, N.; Mi, Z.; Hodge, B.-M.; Shafie-khah, M.; Catalão, J.P. Association rule mining based quantitative analysis approach of household characteristics impacts on residential electricity consumption patterns. Energy Convers. Manag. 2018, 171, 839–854. [Google Scholar] [CrossRef]
  51. Ullah, W.; Ullah, A.; Hussain, T.; Khan, Z.A.; Baik, S.W. An efficient anomaly recognition framework using an attention residual LSTM in surveillance videos. Sensors 2021, 21, 2811. [Google Scholar] [CrossRef]
  52. Ullah, W.; Hussain, T.; Khan, Z.A.; Haroon, U.; Baik, S.W. Intelligent dual stream CNN and echo state network for anomaly detection. Knowl.-Based Syst. 2022, 253, 109456. [Google Scholar] [CrossRef]
  53. Jia, X.; Han, Y.; Li, Y.; Sang, Y.; Zhang, G. Condition monitoring and performance forecasting of wind turbines based on denoising autoencoder and novel convolutional neural networks. Energy Rep. 2021, 7, 6354–6365. [Google Scholar] [CrossRef]
  54. Ding, Y.; Li, Y.; Cheng, L. Application of Internet of Things and virtual reality technology in college physical education. IEEE Access 2020, 8, 96065–96074. [Google Scholar] [CrossRef]
  55. Chen, B.; Lin, P.; Lai, Y.; Cheng, S.; Chen, Z.; Wu, L. Very-short-term power prediction for PV power plants using a simple and effective RCC-LSTM model based on short term multivariate historical datasets. Electronics 2020, 9, 289. [Google Scholar] [CrossRef] [Green Version]
  56. Li, L.-L.; Wen, S.-Y.; Tseng, M.-L.; Wang, C.-S. Renewable energy prediction: A novel short-term prediction model of photovoltaic output power. J. Clean. Prod. 2019, 228, 35is9–375. [Google Scholar] [CrossRef]
  57. Wang, K.; Qi, X.; Liu, H. A comparison of day-ahead photovoltaic power forecasting models based on deep learning neural network. Appl. Energy 2019, 251, 113315. [Google Scholar] [CrossRef]
  58. Zhou, Y.; Zhou, N.; Gong, L.; Jiang, M. Prediction of photovoltaic power output based on similar day analysis, genetic algorithm and extreme learning machine. Energy 2020, 204, 117894. [Google Scholar] [CrossRef]
  59. Cheng, L.; Zang, H.; Ding, T.; Wei, Z.; Sun, G. Multi-meteorological-factor-based graph modeling for photovoltaic power forecasting. IEEE Trans. Sustain. Energy 2021, 12, 1593–1603. [Google Scholar] [CrossRef]
  60. Korkmaz, D. SolarNet: A hybrid reliable model based on convolutional neural network and variational mode decomposition for hourly photovoltaic power forecasting. Appl. Energy 2021, 300, 117410. [Google Scholar] [CrossRef]
Figure 1. The proposed DSCLANet framework for solar power prediction.
Figure 1. The proposed DSCLANet framework for solar power prediction.
Sensors 23 00945 g001
Figure 2. Prediction performance of DSCLANet with (a) dataset1, (b) dataset2, and (c) dataset 3.
Figure 2. Prediction performance of DSCLANet with (a) dataset1, (b) dataset2, and (c) dataset 3.
Sensors 23 00945 g002
Table 2. Internal architectures of DSCLANet.
Table 2. Internal architectures of DSCLANet.
TypeNo. of FiltersKernel-SizeParams
Conv325992
Conv6436208
Conv128124,704
LSTM (100)--44,400
LSTM (100)--80,400
Fusion---
Attention--1089
Dense_64--4128
Dense_32--12,928
Dense_12--396
Table 3. Performance comparison of several models developed during the ablation study.
Table 3. Performance comparison of several models developed during the ablation study.
DatasetMethodMSEMAERMSE
Trina 1ACNN0.09660.15260.3108
LSTM0.08040.1430.2836
GRU0.08480.15180.2912
CNNLSTM0.06790.120.2606
CNNGRU0.07930.15190.2817
DSCLANet0.01670.06320.1291
Trina 1BCNN0.11960.20410.3458
LSTM0.07670.14730.2769
GRU0.0650.11960.2549
CNNLSTM0.06480.1310.2546
CNNGRU0.06410.13650.2531
DSCLANet0.02790.08890.167
Eco 2CNN0.04330.12880.2081
LSTM0.04160.10690.2041
GRU0.03840.10110.196
CNNLSTM0.02980.0880.1725
CNNGRU0.0320.08790.1789
DSCLANet0.00740.04790.0858
Table 4. Performance comparison of several models developed during the ablation study.
Table 4. Performance comparison of several models developed during the ablation study.
MethodMSEMAERMSE
WPD-LSTM [54]--0.2357
RCC-LSTM [55]-0.5870.94
HIMVO-SVM [56]--2805
ESN-CNN [7]0.03090.09710.1731
CNN-LSTM [57]-0.1260.343
DenseNet [28]0.0810.152-
LSTM-CNN [48]-0.2210.621
ELM [58]-0.2367-
Graph-network [59]-0.1170.336
SolarNet [60]-0.1750.309
DSCLANet0.01730.06670.1273
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alharkan, H.; Habib, S.; Islam, M. Solar Power Prediction Using Dual Stream CNN-LSTM Architecture. Sensors 2023, 23, 945. https://doi.org/10.3390/s23020945

AMA Style

Alharkan H, Habib S, Islam M. Solar Power Prediction Using Dual Stream CNN-LSTM Architecture. Sensors. 2023; 23(2):945. https://doi.org/10.3390/s23020945

Chicago/Turabian Style

Alharkan, Hamad, Shabana Habib, and Muhammad Islam. 2023. "Solar Power Prediction Using Dual Stream CNN-LSTM Architecture" Sensors 23, no. 2: 945. https://doi.org/10.3390/s23020945

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop