Next Article in Journal
Energy-Efficient Raptor-like LDPC Coding Scheme Design and Implementation for IoT Communication Systems
Previous Article in Journal
Merchant Energy Storage Investment Analysis Considering Multi-Energy Integration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Photovoltaic Power Forecasting Using Multiscale-Model-Based Machine Learning Techniques

1
Laboratory of Automatic Electrical Systems and Environment, National Engineering School of Monastir, University of Monastir, Monastir 5000, Tunisia
2
Research Unit Advanced Materials and Nanotechnologies, Higher Institute of Applied Sciences and Technology of Kasserine, Kairouan University, Kasserine 1200, Tunisia
3
Electrical and Computer Engineering Program, Texas A&M University at Qatar, Doha 23874, Qatar
*
Author to whom correspondence should be addressed.
Energies 2023, 16(12), 4696; https://doi.org/10.3390/en16124696
Submission received: 30 April 2023 / Revised: 15 May 2023 / Accepted: 19 May 2023 / Published: 14 June 2023
(This article belongs to the Section A1: Smart Grids and Microgrids)

Abstract

:
The majority of energy sources being used today are traditional types. These sources are limited in nature and quantity. Additionally, they are continuously diminishing as global energy consumption increases as a result of population growth and industrial expansion. Their compensation is made from clean energy and renewable energy. Renewable energy is strongly dependent on climatic conditions; therefore, an aspect of energy management is needed, which is essential in distribution systems, because it enables us to calculate the precise energy used by the load as well as by its many components. It also helps us understand how much energy is required and its origin. The energy management aspect contains two main phases: forecasting and optimization. In this study, we are focused on the forecasting level using intelligent machine learning (ML) techniques. To ensure better energy management, it is very important to predict the production of renewable energy over a wide time period. In our work, several cases are proposed in order to predict the temperature, the irradiance, and the power produced by a PV system. The proposed approach is validated by an experimental procedure and a real database for a PV system. The big data from the sensors are noisy, which pose a major problem for forecasting. To reduce the impact of noise, we applied the multiscale strategy. To evaluate this strategy, we used different performance criteria, such as mean error (ME), mean absolute error (MAE), root mean square error (RMSE), nRMSE and the coefficient of determination ( R 2 ). The obtained experimental results show good performance with lower error. Indeed, they achieved an error for nRMSE criteria between 0.01 and 0.37.

1. Introduction

Electricity is a significant factor in a country’s development in terms of economy and technology. The growing dependence on energy is resulting from a considerable increase in electricity production all over the world. The primary source of energy production in the world is conventional fossil fuels. These sources are on the point of extinction, and they are one of the key sources of gas emissions that result in extreme global warming [1]. As a result, in the last few years, attention has increasingly been switched from fossil fuels to renewable energy (RE) sources for electricity generation. Amongst these RE sources, the solar energy is considered as one of the promising sources [2]. Looking for new clean energy sources is not the only way to reduce the impact on energy consumption, but one of the most important ways to address this problem is to optimise energy use [3]. Therefore, the aim of designing a sustainable energy production system is to make smart energy use. Accordingly, the interest in energy management is growing. The energy management systems (EMSs) is a great choice for smart energy use and the improvement of company competitiveness, especially in the case of industry and building applications. In order to improve the generation and distribution of energy, modelling and forecasting of energy has become an extremely important challenge in EM.
In general, EM contains two principal steps, including the forecasting and optimization. In this paper, we are focused on the step of forecasting. The increased generation of renewable energy, optimal operation of storage devices, and more adaptable EMS techniques can all result in lower operating costs for power systems. In this regard, this article focuses on solar energy output and irradiance predictions for the best energy management (EM). The two primary categories for forecasting power models are typically physical and statistical models. The statistical approach is based on historical data and how those data statistically relate to forecasts made by meteorologists as well as SCADA (supervisory control and data acquisition) measurements [4]. Moreover, numerical weather prediction (NWP) data can be employed with the statistical forecasting methodology. Forecasts for weather variables such as wind speed, temperature, precipitation, humidity, pressure, and solar irradiation are introduced by NWP models. The physical method [5], however, focuses on including physical information in the model, such as topography data and energy source attributes. In addition, there are many power forecasting time horizon methodologies, which are often divided into several basic scales. A number of recent studies address the issue as interest in predicting solar power has grown. Many of these take into account predictions of the world’s radiation, which is effectively the same issue as forecasting solar power.
In [6], they developed a variety of state-of-the-art probabilistic models for forecasting solar irradiance. They investigated the use of post-hoc calibration techniques for ensuring well-calibrated probabilistic predictions. Four probabilistic models that output a probability distribution over the outcome space instead of a point prediction were discussed: a Gaussian process regression model, a neural network with uncertainty based on a dropout variation (dropout neural network), a neural network whose predictions parameterized a Gaussian distribution optimized to maximize likelihood (variational neural network), and [7] present natural gradient boosting (NGBoost) for probabilistic prediction. They used CRPS and maximum likelihood (MLE) as indicators to evaluate probabilistic prediction. In addition, an ANN inference system was used in the study by [8] to forecast electricity usage and guarantee an ideal EM. Therefore, the paper [9] gives a brief development of computational models based on machine learning (ML), support vector machine (SVM), and Gaussian process regression (GPR) techniques to estimate the PV panel power. The results showed that Matern 5/2 GPR offered the top performance among the offered ML algorithms, with RMSE, MAE, and R2 values of  7.967 5.302 , and  0.98 , respectively. This validates the Matern 5/2 GPR model’s excellent dependability and accuracy. Authors in [10] examined how several environmental factors, including irradiance, relative humidity, ambient temperature, wind speed, PV surface temperature, and collected dust, affected the output power of the PV panel. An internal PV system’s calibration of a number of sensors was reported. To anticipate the PV system’s hourly power output, a number of multiple regression models and ANN-based prediction models were developed and tested, and they yielded root mean square errors (RMSEs) of  2.1436 6.1555 , and  5.5351 , respectively. The ANN models, with all the features and features selected using correlation feature selection (CFS) and relief feature selection (ReliefF) approaches, were shown to accurately forecast PV output power. The goal of this study [11] is to compare the effectiveness of several machine learning models, such as artificial neural networks (ANNs), support vector regression (SVR), and regression trees (RTs), with varying hyper-parameters and variables, in forecasting the power output of PV systems. The ANN beat the other models in a comparison of ideally planned models, where it achieved the lowest prediction mean absolute percentage error (MAPE) and normalised root mean square error (nRMSE) of  0.6 %  and  0.76 % , respectively. Because the nRMSE findings were  1.13 %  and  1.33 % , respectively, the prediction capabilities of the SVR and RT models can be deemed similar in this scenario. Finally, all of the models outperformed the PM in terms of relative improvement, because the skill score (SS) ranged from 86 to 100.
In recent years, there has been an increase in interest in the subject of machine learning techniques for prediction [12]. Many engineering, natural science, and social science applications require precise forecasting models (FM), which are a key component. A forecast is created using data from prior efforts and research into potential future manifestations of identified characteristics [13]. The essential premise of predicting is that the future will in some way resemble the patterns or distribution of the past. The secret to making good forecasts is finding the patterns or hidden information in historical data. Energy forecasting has drawn interest from the development of artificial intelligence (AI) and machine learning (ML) approaches. For more than three decades, AI/ML algorithms have been used for energy forecasting [14,15]. Because of the development of computing technologies, the discipline of AI/ML has recently made more progress. For example, deep learning [16,17], reinforcement learning [18], and transfer learning [19] are some of the advanced AI/ML approaches that have been implemented in energy forecasting. Compared to the suggested forecasting method, this study represents the state-of-the-art in PV power forecasting techniques established over the recent years.
The aim of this paper is to predict the irradiance and PV power using stand-alone system to exploit the ability of each model and show the effectiveness of ML techniques under different operating cases using intervals of nine days (one minute) utilizing real data from the NOAA Surface Radiation (SURFRAD) network [6]. The techniques considered in this study are artificial neural network (ANN), Gaussian process regression, (GPR) and boosting trees. Many sources of failure can affect sensor data, including hardware noise, environmental influences, hardware imperfections, and hardware noise from external sources. In order to avoid this problem, we discuss a multiscale data representation to denoise data. The latter shows us its effectiveness and impact for data cleaning and improved predicting performance.
The mean error (ME), mean absolute error (MAE), root mean squared error (RMSE), nRMSE, and coefficient of determination ( R 2 ) were used to assess the results of each technique. The findings demonstrate that the prediction models have exceptional predicting accuracy and received very top status. The rest of paper is structured as follows: Section 2 presents the description of the Energy Management by giving its structure. A brief description of the proposed ML techniques, presenting a definition of the most well known performance criteria are given in Section 3. Section 4 presents and discusses the results, giving a statement of the source information. Section 5 concludes the paper with a discussion of potential new ideas.

2. Energy Management

The primary management goal is to maintain a balance between electricity production and demand by giving renewable energies more of a priority and taking into account system constraints.
The structure of a management system for renewable energy is shown in Figure 1. The energy management system (EMS) deals with energy control, management, maintenance, and consumption issues to help with the upkeep and repair of electrical equipment in a factory, farm, or even an entire city [20,21]. It may check on the equipment’s operational status and rapidly enhance management in general. An effective management technique can lower expenses and increase the life of electrical equipment. The EMS can also alert management staff to start looking for replacements for old, high-energy-demanding equipment. The system can instantly send out an alarm in the event of equipment malfunctions or other conditions, thereby allowing management employees to monitor and maintain the system and reduce losses to a minimum. In the renewable energy management system, the EMS can connect monitoring stations, management, and control centers distributed throughout the site using programmed control system technology, network communication technology, and database technology. This enables data collection, storage, processing, statistics, query and analysis, and even data monitoring and diagnosis.
In Figure 1, it is shown that, in order to accomplish the objectives of energy monitoring and efficient management, the EMS distributes renewable-energy-generated power in accordance with a power projections made using the forecasting model. Energy consumption per unit is decreased and economic and energy efficiency are considerably increased by centralized monitoring and the efficient administration of energy data. Hence, this paper focuses on the forecasting model, which is a vital component of the EMS.

3. Methodology and Techniques

This section describes information sources and the data clarification approach and shows how to build machine learning techniques to predict solar irradiance.
To predict the power of the field, we applied several machine learning techniques such as: ANN, GPR, LASSO, and boosting trees.

3.1. Artificial Neural Network

Artificial neural networks (ANNs) are a subset of machine learning and artificial intelligence (AI) that emulate human learning, memory, and relationship-finding abilities using computer programs. The technology known as artificial neural networks was developed after research into the brain and nervous system [22].
Artificial neural networks are mentioned as one of the time series prediction methods in [23] due to their high adaptability and ability to address difficult, nonlinear situations. The input layer, hidden layer, and output layer are the three layers that make up the ANN structure. Each layer is made up of a sets of nodes, where the hidden layer processes the information after it has been processed by the input layer, and the output layer then sends the network response. The input layer’s number of neurons is equal to the number of inputs. The fundamental equation for a neuron is
y = f w i x i + b
where y is the output, f is the activation function,  w i  and  x i  are the weight and input, respectively, b is the bias term, and  s u m  is the sum all inputs.
Ref. [24] asserts that a simple ANN architecture yields better predictions than a complicated ANN structure. Additionally, a signal with weight  w i j  connects every two neurons in a layer of increasing complexity. Each neuron sends the information to the neurons in the layer below after processing it using an activation function.
Figure 2 illustrates an ANN’s basic structure. It involves neural connections. Each connection is allocated a weight, which is a numerical value. Neuron i’s output in the hidden layer is  h i  [25].
h i = σ j = 1 N W i j x j + T i h i d
where
  • σ ( )  is the activation function.
  • N is the total number of input neurons.
  • W i j  are the weights.
  • x j  are the inputs to the input neurons.
  • T i h i d  are the threshold terms of the hidden neurons.
Figure 2. Architecture of ANN.
Figure 2. Architecture of ANN.
Energies 16 04696 g002
Given that it is a non-linear function that can be distinguished, the sigmoid activation function is the one that is employed the most commonly [26]. This function is a logistic function that has a range of 0 to 1 and has the following expression:
f = 1 1 + exp ( x )

3.2. Gaussian Process Regression (GPR)

Gaussian process regression (GPR) is a flexible nonparametric Bayesian model that allows a prior probability distribution to be constructed over functions directly [27,28]. It is commonly used for non linear problems. Due to their exceptional generalization capabilities, GPs represent one of the most significant Bayesian discriminative kernel learning approaches, where the interpolated values are modeled by a Gaussian process controlled by prior covariance [29]. The best linear unbiased prediction of the values is produced by the GPR by including the proper prior assumptions [30].
The GP offers accurate distributions for all possible functions using the training dataset. In light of this, the number of variables in a GP is limitless and grows as more training datasets are added. A Gaussian process of the mean function  n ( x )  and kernel function  t ( x , x )  is what is known as a GPR [9].
F ( x ) = G P ( n ( x ) , t ( x , x ) )
where  n ( x )  represents the central tendency of F. The test input x and test output Y values are related as follows:
Y = F ( x ) + ε
where  ε  denotes the independent noise term. It is encased by a distribution with a mean of 0 and a variance of  ε m  and is defined as:
ε = D ( 0 , σ m 2 )
The marginal likelihood for the sample dataset is provided by:
H ( y | f ) = D ( y | f , σ m 2 J )
here
Y = Y 1 , Y 2 , , Y n T
f = f ( x 1 ) , f ( x 2 ) , , f ( x m ) T
The distribution of the predicted dataset is provided by:
H ( Y s | x , x , Y ) = D ( μ s , σ s 2 )
μ s = k s M ( k M + σ m 2 J ) 1 Y
σ s 2 = k s s k s M ( k M + σ m 2 J ) 1 k M s
where
  • μ s  signifies the GP posterior mean’s average value.
  • σ s 2  displays the prediction’s covariance matrix.
  • k s M  is the covariance matrix that associated the training and the test data.
  • x  is the training data.
  • J represents the  M × M  matrix.
The GP prefers functions that correctly smooth and explain trained data. The smoothing functional property allows for excellent generalizations [31].

3.3. Least Absolute Shrinkage and Selection Operator (LASSO)

For computing effective model descriptions of nonlinear systems, the least absolute shrinkage and selection operator (LASSO) method is examined [32].
The least absolute shrinkage and selection operator (LASSO) [33] is a well-known high dimensional data analysis technique that may be used for biomarker data, because it can perform regularization and variable selection at the same time. This can increase the precision of predictions and the readability of the results [34]. This approach reduces the residual sum of squares for a linear regression model, provided that the total absolute value of the coefficients is smaller than a tuning parameter [33] Penalized binomial logistic regression has previously been discussed in detail [35]. For i = 1, …, N, let  x i  = ( x i 1 , …,  x i p ) T  signify p-predictors for N observations. Assume that the binary logistic regression model’s answers can have values G = 1, 2. Then,
Pr ( G = 1 | x ) = 1 1 + e β 0 + x i T β , Pr ( G = 2 | x ) = 1 1 + e β 0 + x i T β
where  β 0  is the intercept, and  β  = ( β 1 β 2 ,…, β p ) T  is a p-vector of regression parameters. This suggests
log Pr ( G = 1 | x ) Pr ( G = 2 | x ) = β 0 + x i T β .
Then, to identify parameter values to minimize, the LASSO method is applied, which is shown as follows:
1 N i = 1 N y i × β 0 + x i T β log ( 1 + e β 0 + x i T β ) + λ j = 1 p β j
where,  λ j = 1 p β j  denotes the LASSO’s penalty function.
It has been demonstrated that the LASSO approach does not always offer reliable variable selection. Even when the coefficients are big, the LASSO penalises all of them equally. In contrast, the AL penalises coefficients differently by using adaptive weights [36].
The AL employs a weighted fine
λ . j = 1 p w j . β j ,
where
w j = 1 β ^ j v , β ^ j
These represent the most likely estimate, and  v > 0 . The weighted penalty will make it possible for variables with larger coefficients to obtain smaller penalties, thereby potentially leading to a more reliable outcome.

3.4. Boosting Trees

Boosting was first used in the machine learning community after AdaBoost was released [37]. The fundamental concept is to average out the predictions from a number of weak classifiers (high PE) to create a strong classifier (low PE) [38]. Let us say we want to approximate the answer y by the function f(x) given the predictors x and the response y. Usually, to estimate  f ( x ) , we first give the functional form of f(x) along with a loss function  L ( y , f ( x ) ) ( x ) .
The linear model is defined with  f ( x ) = x β , where  β  is a matrix of parameters, and the squared error loss function is the most well-known form of f, i.e.,  L ( y , f ( x ) )  =  ( y f ( x ) ) 2 .

3.5. Performance Evaluation Metrics

The prediction performance of the suggested models was assessed using distinct statistical evaluation criteria. In this paper we are focused on the mean absolute error (MAE), which is computed by expressing the mean absolute deviation of the difference between the expected values and the actual values; the root mean square Error (RMSE), in which it is calculated by using the estimation errors’ standard deviation as its input; and finally the coefficient of determination, which it is determined by using a metric that represents the strength of the linear relationship between the predicted values of the modeling techniques and the actual values. Their expressions are respectively the following:
MAE = 1 N i = 1 N o i p i
RMSE = i = 1 N o i p i 2 N
R 2 = 1 i = 1 N o i p i 2 i = 1 N o i o 2
where
  • o i  is the observation’s actual value.
  • p i  is predicted value of of the observation.
  • o  is the average of the actual observation values.
  • N is the number of samples utilized for the statistical evaluation criteria.
These errors are frequently standardized, especially for the RMSE; as a standard, the mean value is typically employed,; however, various definitions can be obtained.
nRMSE = 1 N × i = 1 N o i p i 2 y ¯
where  y ¯  is the mean value of p. The index of agreement (d), which is normalized between 0 and 1, and the correlation coefficient R (Pearson Coefficient) are two more indices that can be employed.

4. Simulation Results

4.1. System Description

A single PV array block was composed of (2) parallel strings with (12) modules connected in series on each string. The stand-alone or off-grid PV was considered in this study, where the PV system is independent of the utility grid. We studied this type of system by supplying with a load based on the climatic for the purpose of extracting the power.
The “basic” single diode model typically serves as an illustration of how a PV cell operates. This simplified model is frequently employed for failure investigations, performance analysis, and stability analysis [39]. Figure 3 displays the equivalent circuit of a PV cell.
This model is composed of five parameters and it contains;
  • I p h : a current source which represents the irradiance received by the cell.
  • A diode for modeling the PN junction of the cell.
  • R s : the series resistance representing the resistivity of the material through which the cell is made. the cell is made.
  • R s h : the parallel resistance, which represents all the paths crossed by the leakage current, either in parallel with the leakage current, in parallel with the cell, or at the edge of the cell.
The current supplied by the cell is given by the following equation:
I = I p h I d I s h u n t
where
I s h u n t = V + R s I R s h
As we said before, this model is also called “implicit model with five parameters parameters” which are: ( I p h I s , a R s R s h ).
Where
  • I p h  is the photo current, which is proportional to the irradiance received by the cell.
  • I s  is the saturation current of the diode.
  • a is the ideality factor of the diode at (1 to 2).
  • V t  =  N s K b T c / q  is the thermal voltage as a function of the number of cells in series in the PV module, the cell temperature, Boltzmann’s constant, and the charge of the electron.
The photo current depends linearly on the global irradiance G and the temperature T, and it is found in the following equation:
I p h = G G n I p h , n + K i Δ T
Δ T = T T n
T c = T α + 0.2 G G n
where
  • T T c  + 273.15
  • I p h , n  is the photo current at the standard test condition, which corresponds to the short-circuit current of the PV module given by the manufacturer.
  • K i  is the temperature coefficient of the short circuit current.
  • T is the temperature of the PV cell in kelvin.
  • T n  is the cell temperature at standard test condition [25 °C].
  • G is the solar radiation received by the PV cell [W/m2].
  • G n  is the solar radiation at standard test conditions [1000 W/m2].
  • T a  is the ambient temperature in °C.
  • T c  is the temperature of the PV cell in °C.
The well-known equation for the saturation current of the diode is given by:
I s = I s , n T T n 3 exp q E g a K b 1 T n 1 T
E g  is the the band energy of the semiconductor.  I s , n  is the the nominal saturation current at (STC).
An improved equation to describe the saturation current that considers the variation of temperature is given by:
I s = I s c , n + K i × Δ T exp V s c , n + K v × Δ T / a V t 1
where the constant  K v  is the temperature coefficient of the open circuit voltage.
These equations were implemented on the Simulink that followed the architecture in Figure 4. In order to evaluate the model (Figure 4), simulation tests were performed using the manufacturer settings [39].

4.2. Data Collection

MATLAB software (R2021b) was used in this work to train machine learning techniques. A case study was conducted using a dataset obtained from the NOAA’s Surface Radiation (SURFRAD) network, which is composed of seven stations from the U.S. [6] to assess the effectiveness of the proposed method, which includes nine days of data (from 1 July 2017 to 9 July 2017) from the Goodwin Creek, Mississippi (GWN) station. The initial photoelectric data included 9 × 660 = 5940 observations from 8 am to 6 pm; the choice of this time interval was made at the time that there was an energy production and a power generation, which were gathered at 1-min intervals. A collection of data inputs was employed in this experiment. In fact, we fed the studied system with the irradiance and the temperature, which were processed in the database detailed above to extract the power. Then, for each input pair (irradiance, temperature), we determined the power produced. Therefore, from a real database, we determined a database for the power supplied by the PV system. Figure 5 shows us the curves of the temperature, irradiance, and the power measured on one day. The following figure shows us, moreover, that the variation of the irradiance was not very big compared to the variation of the power, and the variation of the temperature was light. We can conclude that if the temperature is not very important, the power will be the image of the irradiance. However, the variation in the power was multiplied to the value of the irradiance (between 120 and 800); that is why we predicted the latter and we made a prediction of the power, since it affects the robustness of the techniques. This expresses the importance of the power compared to the irradiance.
To accomplish the objectives of energy monitoring and efficient management, the EMS distributes renewable energy produced power in response to a power projection made using the forecasting model. The energy consumption per unit is decreased, and economic and energy efficiency are considerably increased by the centralised monitoring and efficient administration of energy data. As a result, the forecasting model is a key component of the EMS. Because of this, the EMS can predict future energy output with accuracy.
The best option for quick and effective energy reduction is energy management. In order to maximise energy conservation, our major goal is to analyse the techniques of consuming energy and to improve generation and utilization. For optimal energy usage, both the generating and utilisation sides must be considered. The most interesting goal for this paper is to ensure a better management, to reach an optimal energy, to reduce the energy consumption, and to obtain higher accuracy. We proposed four different cases by adding time intervals in each case to obtain better management.
In the current work, different predictions horizon lengths were used to improve the energy management as well as to evaluate the forecasting performance of the proposed techniques. To achieve these goals, four cases were addressed as follows:
  • Case 1: During this case, one day was examined. A total of 8 hours (h) and 48 minutes (min) were used for training to predict 2 h and 12 min.
  • Case 2: In this case, two days were considered (1320 observations) for the training phase to predict the next day.
  • Case 3: Four days (2640 observations) were examined for training to predict the following two days.
  • Case 4: We continued increasing the training samples to obtain larger prediction horizon lengths. In such a way, six days (3960 observations) were intended to perform the training phase to predict three successive days.
The selection of the forecasting horizon length was based on the significant variation of the irradiance, either during the hours of one day or through successive days, in which the irradiance underwent several peaks with totally different behavior. These considerable variations are clearly shown in the following curves illustrated in Figure 6, Figure 7, Figure 8 and Figure 9, especially in the three latter.
For example, in Figure 9, which represents Case 4, we see the variation in irradiance and the power over the time, such as in (Figure 9a). The three successive predicted days did have not the same variation in irradiance; every day had its behavior. We can see a big variation in the first forecasted day with a significant irradiance. On the second predicted day, we can see that the values of irradiance decreased compared to the first day; then, they started to increase a little with a totally different variation in the irradiance in comparison with the previous day. We can say here that the irradiance was a bit weak with another weak behavior. For the third day, we found another different pattern than the two previous ones, with weak variation peaks.
As a result, the error was practically low, but we attempted to diminish it and enhance the obtained results to show better energy management; we used the multiscale model to denoise the raw data. By using this model, the cases used show us that the errors were decreased even more and that the proposed techniques remained generally robust.

4.3. Discussion

The previous section briefly described the proposed approaches, as well as the use and the need of the system and the studied cases. According to the latter, we can discuss the following results. Figure 6 displays the predicted values of the proposed ANN model that was used to estimate the irradiance, as well as a time graph showing the actual values over the first day (1 July 2017). Additionally, it was reported that the actual and predicted values indicated almost the same trend. In addition, this figure shows the average daily irradiance and power, where the error of the irradiance and power was higher, as shown in Figure 6a,b.
Figure 7, Figure 8 and Figure 9 show us the prediction of the irradiance and the power for the three other cases. It can be concluded here that the error was lower than the first case, so we conclude that these cases have good prediction properties.
In addition, the effectiveness of the proposed ANN model was evaluated against four trained and tested machine learning methods, including LASSO, GPR, and boosting trees for the time period of 1 July through 9 July. The ME, MAE, RMSE, NRMSE, and  R 2  values were used to compare the effectiveness of these strategies. The following table displays the findings. The mentioned evaluation indicators were used to evaluate the performance of the NN and other models in order to assess the accuracy of various prediction models. The performance of the persistence models for raw data and multiscale model, including NN, GPR, LASSO, and boosting rrees, is shown in Table 1 and Table 2 for four distinct prediction scales. For Case 1, Table 1 show us that the different criteria of GPR, which are ME, RMSE, nRMSE, MAE, and  R 2 , were, respectively,  18.40 19.10 0.11 18.40 , and  1.00  in raw data, as shown in the same table that the predicted values in multiscale were improved with blue 43.70, 43.40, 45.45, 43.70, and  0 %  for the five different criteria to achieve the following respective values of  10.36 10.81 0.06 10.36 , and  1.00 . This approach had the good performance. For the second case, the GPR model had the best values where, in the raw data, it achieved  7.85 13.30 0.06 10.53 , and  1.00  for the different evaluation criteria ME, RMSE, nRMSE, MAE, and  R 2 , respectively. Correspondingly, these values were enhanced with 87.64, 83.83, 83.33, 85.94, and  0 %  when we used the multiscale, so the obtained values were  0.97 2.15 0.01 1.48 , and  1.00 , respectively.
For Case 3, when we increased the data more than the previous case, we can see that the GPR technique had the best values of performance criteria for all different performances, which were  1.75 22.87 0.05 18.98 , and  1.00  in the raw data. When using the multiscale, these values were 56.57, 81.99, 80, 83.14, and  0 %  to obtain  0.76 4.12 0.01 3.20 , and  1.00 , respectively, for the ME, RMSE, nRMSE, MAE, and  R 2 .
By increasing the data in the last case to achieve six days for the training and three days for testing to show the performance of our models, we can see that the ME, RMSE, nRMSE, MAE, and  R 2  of the GPR model were, respectively  0.55 21.87 0.06 17.87 , and  1.00 ; after using the multiscale, this technique achieved  0.13 3.71 0.01 2.92 , and  1.00  respectively. This model was enhanced by a percentage of 76.36, 83.04, 83.33, 83.66 and  0 % , respectively. We can say that the multiscale had an especially good impact for this technique. For the four prediction irradiance scales of Case 1, Case 2, Case 3, and Case 4, these findings demonstrate that multiscale representation has good impact for enhancing the performance criteria of the different models.
We selected four prediction scales for Cases 1, 2, 3, and 4 to test the performance of the power prediction model. As shown in Table 2, the LASSO model had the best values for raw data, and the multiscale model performed best for the first case. In the second case, we can see that the GPR technique had the minimum error values for the two applications. Moving to the third case, we can see that the multiscale model had a good impact for enhancing the NN method. The last case showed us that the NN technique outperformed other prediction models for the raw data and for using the multiscale model.

5. Conclusions

In this study, the use of multiscale machine learning techniques (i.e., ANN, GPR, LASSO, and boosting trees) to process irrandiance and power forecasting has been demonstrated using historical real data from the Goodwin Creek, Mississippi (GWN) station in the United States. Given that measured process data are generally tainted by errors (noise) that hide the significant features in data and degrade the effectiveness of the forecasting approach, the multiscale representation data has been proven to be an effective tool of data analysis and feature extraction owing to its capacity to present useful separations of features. Accordingly, the goal of the developed approach was to apply the multiscale representation data to further improve the efficiency of applied ML irradiance and power foretasting techniques. In order to evaluate the proposed approach, four cases were studied that varied from one day to nine (1 July 2017 to 9 July 2017) using the mean error (ME), mean absolute error (MAE), root mean square error (RMSE), nRMSE, and coefficient of determination  R 2  criteria. The obtained results proved the efficiency of the proposed approach with nRMSE values (for exemple) between  0.01  and  0.37  for predicted irradiance, and nRMSE values between  0.02  and  0.71  for predicted power.

Author Contributions

Methodology, M.M. (Majdi Mansouri); software, M.M. (Manel Marweni); validation, M.M. (Manel Marweni); investigation, M.H. and M.F.M.; writing—original draft preparation, M.M. (Manel Marweni); writing—review and editing, M.H. and M.M. (Majdi Mansouri); supervision, M.H., M.M. (Majdi Mansouri) and M.F.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Qatar National Library through the Qatar National Research Fund (QNRF) Research Grant.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author, MM, upon reasonable request.

Acknowledgments

Open access funding provided by the Qatar National Library.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kumari, P.; Toshniwal, D. Deep learning models for solar irradiance forecasting: A comprehensive review. J. Clean. Prod. 2021, 318, 128566. [Google Scholar] [CrossRef]
  2. Feng, J.; Xu, S.X. Integrated technical paradigm based novel approach towards photovoltaic power generation technology. Energy Strategy Rev. 2021, 34, 100613. [Google Scholar] [CrossRef]
  3. Cárdenas, J.J.; Romeral, L.; Garcia, A.; Andrade, F. Load forecasting framework of electricity consumptions for an Intelligent Energy Management System in the user-side. Expert Syst. Appl. 2012, 39, 5557–5565. [Google Scholar] [CrossRef]
  4. Svensson, M. Short-Term Wind Power Forecasting Using Artificial Neural Networks. Master’s Thesis, KTH–School of Computer Science and Communication Location, Kista, Sweden, 2015. [Google Scholar]
  5. Landberg, L. Short-term prediction of the power production from wind farms. J. Wind. Eng. Ind. Aerodyn. 1999, 80, 207–220. [Google Scholar] [CrossRef]
  6. Zelikman, E.; Zhou, S.; Irvin, J.; Raterink, C.; Sheng, H.; Avati, A.; Kelly, J.; Rajagopal, R.; Ng, A.Y.; Gagne, D. Short-term solar irradiance forecasting using calibrated probabilistic models. arXiv 2020, arXiv:2010.04715. [Google Scholar]
  7. Duan, T.; Anand, A.; Ding, D.Y.; Thai, K.K.; Basu, S.; Ng, A.; Schuler, A. Ngboost: Natural Gradient Boosting for Probabilistic Prediction. In Proceedings of the International Conference on Machine Learning ( PMLR), Online, 13–18 July 2020; pp. 2690–2700. [Google Scholar]
  8. Li, Z.; Ye, L.; Zhao, Y.; Song, X.; Teng, J.; Jin, J. Short-term wind power prediction based on extreme learning machine with error correction. Prot. Control. Mod. Power Syst. 2016, 1, 1. [Google Scholar] [CrossRef] [Green Version]
  9. Zazoum, B. Solar photovoltaic power prediction using different machine learning methods. Energy Rep. 2022, 8, 19–25. [Google Scholar] [CrossRef]
  10. Khandakar, A.; Chowdhury, M.E.H.; Khoda Kazi, M.; Benhmed, K.; Touati, F.; Al-Hitmi, M.; Gonzales, A.S.P., Jr. Machine learning based photovoltaics (PV) power prediction using different environmental parameters of Qatar. Energies 2019, 12, 2782. [Google Scholar] [CrossRef] [Green Version]
  11. Theocharides, S.; Makrides, G.; Georghiou, G.E.; Kyprianou, A. Machine Learning Algorithms for Photovoltaic System Power Output Prediction. In Proceedings of the 2018 IEEE International Energy Conference (ENERGYCON), Limassol, Cyprus, 3–7 June 2018; pp. 1–6. [Google Scholar]
  12. Ensafi, Y.; Amin, S.H.; Zhang, G.; Shah, B. Time-series forecasting of seasonal items sales using machine learning—A comparative analysis. Int. J. Inf. Manag. Data Insights 2022, 2, 100058. [Google Scholar] [CrossRef]
  13. Green, F. Machine-learning Sales Forecasting: A Review. Sage Sci. Rev. Appl. Mach. Learn. 2022, 5, 1–21. Available online: https://journals.sagescience.org/index.php/ssraml/article/view/3 (accessed on 19 April 2023).
  14. Chen, B.J.; Chang, M.W.; Lin, C.J. Load forecasting using support vector machines: A study on EUNITE competition 2001. IEEE Trans. Power Syst. 2004, 19, 1821–1830. [Google Scholar] [CrossRef] [Green Version]
  15. Kariniotakis, G.; Stavrakakis, G.; Nogaret, E. Wind power forecasting using advanced neural networks models. IEEE Trans. Energy Convers. 1996, 11, 762–767. [Google Scholar] [CrossRef]
  16. Shi, H.; Xu, M.; Li, R. Deep learning for household load forecasting—A novel pooling deep RNN. IEEE Trans. Smart Grid 2017, 9, 5271–5280. [Google Scholar] [CrossRef]
  17. Wang, L.; Zhang, Z.; Chen, J. Short-term electricity price forecasting with stacked denoising autoencoders. IEEE Trans. Power Syst. 2016, 32, 2673–2681. [Google Scholar] [CrossRef]
  18. Feng, C.; Sun, M.; Zhang, J. Reinforced deterministic and probabilistic load forecasting via Q-learning dynamic model selection. IEEE Trans. Smart Grid 2019, 11, 1377–1386. [Google Scholar] [CrossRef]
  19. Cai, L.; Gu, J.; Jin, Z. Two-layer transfer-learning-based architecture for short-term load forecasting. IEEE Trans. Ind. Inform. 2019, 16, 1722–1732. [Google Scholar] [CrossRef]
  20. Huang, C.J.; Kuo, P.H. A short-term wind speed forecasting model by using artificial neural networks with stochastic optimization for renewable energy systems. Energies 2018, 11, 2777. [Google Scholar] [CrossRef] [Green Version]
  21. Fathima, A.H.; Palanisamy, K. Energy storage systems for energy management of renewables in distributed generation systems. In Energy Management of Distributed Generation Systems; InTech: Rijeka, Croatia, 2016; pp. 157–181. [Google Scholar]
  22. Dhibi, K.; Mansouri, M.; Bouzrara, K.; Nounou, H.; Nounou, M. Reduced neural network based ensemble approach for fault detection and diagnosis of wind energy converter systems. Renew. Energy 2022, 194, 778–787. [Google Scholar] [CrossRef]
  23. Rojat, T.; Puget, R.; Filliat, D.; Del Ser, J.; Gelin, R.; Díaz-Rodríguez, N. Explainable artificial intelligence (xai) on timeseries data: A survey. arXiv 2021, arXiv:2104.00950. [Google Scholar]
  24. Hassoum, M. Fundamentals of Artificial Neural Networks; MIT Press: Cambridge, MA, USA, 1995. [Google Scholar]
  25. Hichri, A.; Hajji, M.; Mansouri, M.; Abodayeh, K.; Bouzrara, K.; Nounou, H.; Nounou, M. Genetic-Algorithm-Based Neural Network for Fault Detection and Diagnosis: Application to Grid-Connected Photovoltaic Systems. Sustainability 2022, 14, 10518. [Google Scholar] [CrossRef]
  26. Jamii, J.; Mansouri, M.; Trabelsi, M.; Fouazi Mimouni, M.; Shatanawi, W. Effective ANN based on Wind Power Generation and Load Demand Forecasting for Optimum Energy Management. Front. Energy Res. 2022, 10, 898413. [Google Scholar] [CrossRef]
  27. Bishop, C.M.; Nasrabadi, N.M. Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006; Volume 4. [Google Scholar]
  28. Nabney, I. NETLAB: Algorithms for Pattern Recognition; Springer Science & Business Media: Berlin, Germany, 2002. [Google Scholar]
  29. Raghavendra, N.S.; Deka, P.C. Multistep ahead Groundwater Level Time-Series Forecasting Using Gaussian Process Regression and ANFIS. In Advanced Computing and Systems for Security; Springer: Berlin/Heidelberg, Germany, 2016; pp. 289–302. [Google Scholar]
  30. Rasmussen, C. Gaussian Processes in Machine Learning. In Advanced Lectures on Machine Learning; Springer: Berlin/Heidelberg, Germany, 2014; Volume 3176, pp. 63–71. [Google Scholar]
  31. Heyns, T.; De Villiers, J.P.; Heyns, P.S. Consistent haul road condition monitoring by means of vehicle response normalisation with Gaussian processes. Eng. Appl. Artif. Intell. 2012, 25, 1752–1760. [Google Scholar] [CrossRef] [Green Version]
  32. Kukreja, S.L.; Löfberg, J.; Brenner, M.J. A least absolute shrinkage and selection operator (LASSO) for nonlinear system identification. IFAC Proc. Vol. 2006, 39, 814–819. [Google Scholar] [CrossRef] [Green Version]
  33. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  34. Vasquez, M.M.; Hu, C.; Roe, D.J.; Chen, Z.; Halonen, M.; Guerra, S. Least absolute shrinkage and selection operator type methods for the identification of serum biomarkers of overweight and obesity: Simulation and application. BMC Med. Res. Methodol. 2016, 16, 154. [Google Scholar] [CrossRef] [Green Version]
  35. Friedman, J.; Hastie, T.; Tibshirani, R. Regularization paths for generalized linear models via coordinate descent. J. Stat. Softw. 2010, 33, 1. [Google Scholar] [CrossRef] [Green Version]
  36. Zou, H. The adaptive lasso and its oracle properties. J. Am. Stat. Assoc. 2006, 101, 1418–1429. [Google Scholar] [CrossRef] [Green Version]
  37. Efron, B.; Hastie, T.; Johnstone, I.; Tibshirani, R. Least angle regression. Ann. Statist. 2004, 32, 407–499. [Google Scholar] [CrossRef] [Green Version]
  38. De’Ath, G. Boosted trees for ecological modeling and prediction. Ecology 2007, 88, 243–251. [Google Scholar] [CrossRef]
  39. Mansouri, M.; Hajji, M.; Trabelsi, M.; Harkat, M.F.; Al-khazraji, A.; Livera, A.; Nounou, H.; Nounou, M. An effective statistical fault detection technique for grid connected photovoltaic systems based on an improved generalized likelihood ratio test. Energy 2018, 159, 842–856. [Google Scholar] [CrossRef]
Figure 1. Structure of an energy management system (EMS).
Figure 1. Structure of an energy management system (EMS).
Energies 16 04696 g001
Figure 3. Equivalent circuits of one-diode model for a PV cell.
Figure 3. Equivalent circuits of one-diode model for a PV cell.
Energies 16 04696 g003
Figure 4. The block diagram achieved in Simulink for the current, output voltage, and power generation.
Figure 4. The block diagram achieved in Simulink for the current, output voltage, and power generation.
Energies 16 04696 g004
Figure 5. Temperature, irradiance, and power on the first day.
Figure 5. Temperature, irradiance, and power on the first day.
Energies 16 04696 g005
Figure 6. Irradiance and power prediction for Case 1.
Figure 6. Irradiance and power prediction for Case 1.
Energies 16 04696 g006
Figure 7. Irradiance and power prediction for Case 2.
Figure 7. Irradiance and power prediction for Case 2.
Energies 16 04696 g007
Figure 8. Irradiance and power prediction for Case 3.
Figure 8. Irradiance and power prediction for Case 3.
Energies 16 04696 g008
Figure 9. Irradiance and power prediction for Case 4.
Figure 9. Irradiance and power prediction for Case 4.
Energies 16 04696 g009
Table 1. Performance evaluation of Predicted Irradiance under different cases.
Table 1. Performance evaluation of Predicted Irradiance under different cases.
CasesMethodsGlobal Performances
Raw DataMultiscale
MERMSEnRMSEMAE R 2 MERMSEnRMSEMAE R 2
Case 1NN44.6053.390.2644.600.9819.0735.980.2028.460.91
GPR18.4019.100.1118.401.0010.3610.810.0610.361.00
LASSO24.3926.930.2024.391.0011.7112.930.0911.711.00
Boosting Trees63.8280.190.3665.070.8663.3081.650.3765.680.84
Case 2NN7.3214.500.0611.251.003.8727.200.124.390.99
GPR7.8513.300.0610.531.000.972.150.011.481.00
LASSO33.6942.230.2233.691.0019.8724.900.1219.871.00
Boosting trees9.2921.140.0915.150.991.4312.830.069.241.00
Case 3NN1.8823.400.0519.231.000.0710.530.025.231.00
GPR1.7522.870.0518.981.000.764.120.013.201.00
LASSO77.2885.960.2477.281.0034.6638.560.0134.661.00
Boosting trees0.5633.970.0822.160.990.9815.820.0411.991.00
Case 4NN0.3921.640.0617.431.000.494.580.012.871.00
GPR0.5521.870.0617.871.000.133.710.012.921.00
LASSO64.3774.960.2464.371.0035.6141.470.1235.611.00
Boosting trees0.0230.600.0821.420.991.7912.560.038.711.00
Table 2. Performance evaluation of Predicted Power under different cases.
Table 2. Performance evaluation of Predicted Power under different cases.
CasesMethodsGlobal Performances
Raw DataMultiscale
MERMSEnRMSEMAE R 2 MERMSEnRMSEMAE R 2
Case 1NN41.7548.010.5741.750.9539.0458.440.7139.040.32
GPR11.9211.920.2211.921.006.826.860.146.821.00
LASSO7.319.750.277.311.004.165.540.144.161.00
Boosting trees30.4035.690.4830.770.9225.9033.300.4827.920.90
Case 2NN10.7426.490.2215.040.992.044.520.042.871.00
GPR10.2921.010.1814.400.991.443.790.032.301.00
LASSO23.3642.550.5023.361.0012.1322.100.2312.131.00
Boosting trees11.6628.620.2418.030.991.1414.520.138.921.00
Case 3NN0.1839.170.1229.300.990.176.900.024.161.00
GPR0.2736.900.1129.221.000.716.220.024.021.00
LASSO76.5297.800.3976.521.0025.8333.020.1125.831.00
Boosting trees1.5445.140.1431.990.991.1423.960.0715.531.00
Case 4NN0.4137.480.1326.101.000.586.990.024.351.00
GPR0.9339.970.1427.351.000.677.000.024.121.00
LASSO61.3587.550.4061.351.0028.7541.020.1628.751.00
Boosting trees1.2649.130.1828.410.990.1217.240.0610.881.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Marweni, M.; Hajji, M.; Mansouri, M.; Mimouni, M.F. Photovoltaic Power Forecasting Using Multiscale-Model-Based Machine Learning Techniques. Energies 2023, 16, 4696. https://doi.org/10.3390/en16124696

AMA Style

Marweni M, Hajji M, Mansouri M, Mimouni MF. Photovoltaic Power Forecasting Using Multiscale-Model-Based Machine Learning Techniques. Energies. 2023; 16(12):4696. https://doi.org/10.3390/en16124696

Chicago/Turabian Style

Marweni, Manel, Mansour Hajji, Majdi Mansouri, and Mohamed Fouazi Mimouni. 2023. "Photovoltaic Power Forecasting Using Multiscale-Model-Based Machine Learning Techniques" Energies 16, no. 12: 4696. https://doi.org/10.3390/en16124696

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop