Next Article in Journal
Approximated Backscattered Wave Models of a Lossy Concentric Dielectric Sphere for Fruit Characterization
Previous Article in Journal
Efficient Discovery of Partial Periodic Patterns in Large Temporal Databases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Data-Driven Model to Forecast Multi-Step Ahead Time Series of Turkish Daily Electricity Load

by
Kamil Demirberk Ünlü
Department of Mathematics, Atilim University, Ankara 06830, Turkey
Electronics 2022, 11(10), 1524; https://doi.org/10.3390/electronics11101524
Submission received: 14 April 2022 / Revised: 6 May 2022 / Accepted: 9 May 2022 / Published: 10 May 2022
(This article belongs to the Section Computer Science & Engineering)

Abstract

:
It is critical to maintain a balance between the supply and the demand for electricity because of its non-storable feature. For power-producing facilities and traders, an electrical load is a piece of fundamental and vital information to have, particularly in terms of production planning, daily operations, and unit obligations, among other things. This study offers a deep learning methodology to model and forecast multistep daily Turkish electricity loads using the data between 5 January 2015, and 26 December 2021. One major reason for the growing popularity of deep learning is the creation of new and creative deep neural network topologies and significant computational advancements. Long Short-Term Memory (LSTM), Gated Recurrent Network, and Convolutional Neural Network are trained and compared to forecast 1 day to 7 days ahead of daily electricity load. Three different performance metrics including coefficient of determination (R2), root mean squared error, and mean absolute error were used to evaluate the performance of the proposed algorithms. The forecasting results on the test set showed that the best performance is achieved by LSTM. The algorithm has an R2 of 0.94 for 1 day ahead forecast, and the metric decreases to 0.73 in 7 days ahead forecast.

1. Introduction

Accurate electricity load models and forecasts are critical for electric power system planning and operation. Many important decisions about how to run the power system and trade energy are easily made when you know how much load you will have. Load forecasts are used to make a variety of operational decisions, including generation allocation, security assessment, and maintenance management. It has been changed since the early 1990s when a deregulation structure was added and competitive markets were set up. Market rules such as spot and derivative contracts are being considered by a large number of individuals as a method to trade energy [1].
Electricity, a need for most people, is a limited resource. Economic efficiency, or making the greatest use of limited resources, is at the heart of economic theory. Consumer and producer wellbeing may be seen as a single pie that can be maximized via economic efficiency. Complex relationships among the players in the electrical business need varying levels of government action. Instantaneous adjustment is required in the system for generating and transferring power. When there is not enough electricity to go around, there are power outages. Although power outages have decreased in recent years, they continue to occur in Turkey [2].
Stochastic characteristics of the electrical load make it difficult to precisely forecast power production and consumption on a normal day. Therefore, the ability to predict electricity loads is critical to the planning of both demand and supply. It is a common yet tough time-series forecasting subject researched by both academics and practitioners alike. The forecasting period is an essential aspect of any time series forecasting, along with the load data’s input–output linkages, stationarity, and periodicity. The forecasting period is usually divided into three categories. For short-term load forecasting, the range is from one hour to one week. Medium and long-term load forecasting, on the other hand, covers a range of periods from a few weeks to several months and from a year to several years in the future [3,4].
This study aims to forecast multiple steps of daily electricity loads in Turkey by employing and comparing recurrent neural network (RNN) algorithms and convolution neural network (CNN) for the periods between 5 January 2015 and 26 December 2021. The suggested model forecasts both short-term and mid-term timeframes. Moreover, we propose a model that uses its own lag, such as univariate times series methodology. In many time series applications, this sort of model has been employed since it does not make any assumptions about the time series and is particularly good at mimicking the nonlinear structure that occurs in the time series. Although the strongest aspect of this model seems to be that it only employs its lags, it does not need the basic assumptions in traditional time series analysis such as linearity, normality, and zero error covariance. Some of the efficient studies and areas that RNN models used in time series models other than electricity load are air pollutants [5], COVID-19 mutation rate [6], the stock price [7], natural gas demand [8], workload of cloud data center [9], and solar power system assessment [10].
To the best of our knowledge, this will be the first study that employs and compares RNN algorithms of gated recurrent unit (GRU), long-short term memory (LSTM), and CNN to forecast and model multistep daily Turkish electricity load. The proposed models are compared to forecast 1 to 7 days’ electricity loads. An attempt was made to propose a single model that can be used in short- to mid-term forecasting without using any exogenous variables. In light of the context presented in the section on the literature review, the following contributions of this research to the literature might be listed:
  • Using a deep learning technology, Turkey’s daily electricity loads can be modeled and forecasted with an impressive degree of accuracy up to 7 days ahead. The proposed model can be used for short- and mid-term forecasting.
  • The suggested model is of the univariate type, which means that it simply makes use of information collected just from the time series. Since it creates a solid prediction using just its own lags, it is cost and time effective. As a result, it is both efficient and powerful.
  • LSTM is demonstrated to be superior to GRU and CNN type methods when compared. It is, thus, a viable option for forecasting many stages of electricity loads without considering other factors.
  • A thorough evaluation of forecast models is carried out.
The rest of this paper is structured as follows. Section 2 summarizes the available literature, while Section 3 provides the theoretical foundation for the methods used. Section 4 is dedicated to data and analysis. Finally, Section 5 brings this study to a close.

2. Recent Literature

The literature review is organized into four sections. In the first part, studies on short-term forecasting are introduced. It is followed by mid-term and long-term forecasting. Lastly, some recent studies on Turkey are summarized. Since the literature on the electricity load model is intense, we cannot mention many valuable studies. We refer interested readers for a comprehensive review to [11,12,13].
Pai and Hong [14] compared Support Vector Machine (SVR) hybridized by simulated annealing (SA) algorithms by autoregressive integrated moving average (ARIMA) to forecast 1 year ahead electricity load of Taiwan. The authors used yearly data sets between the periods 1945 and 2003. The empirical results show that the suggested model provides a viable option for use in electricity load modeling. Zhang et al. [15] developed a hybrid model for predicting short-term power demand based on improved empirical mode decomposition (IEMD), wavelet neural network (WNN), ARIMA, and optimization via fruit fly optimization algorithm (FOA). The suggested model’s performance is shown using electrical load data from the Australian and New York energy markets, and the results showed that the proposed model outperforms the compared models.
The short-term electricity load of Macedonia was forecasted by [16] using a deep belief network and the results indicate that the proposed model is superior to traditional methods. An Ensemble Kalman Filter (EnKF) was combined with multiple regression and shrinkage methods proposed by [17] to model the short-term electricity load of Tokyo. When compared to current state-of-the-art models, the authors discovered that their predictions were far more accurate, and this method also provides rich analytical data. A new feature selection algorithm with a hybrid deep learning methodology based on Elman neural network (ENN) and ridgelet neural network (rNN) was suggested by [18] to model electricity loads of Australia, North America, and Pennsylvania–New Jersey–Maryland. It has been determined that the suggested approach of the study is effective based on the findings obtained.
Khwaja et al. [19] improved short-term electricity demand forecasting by using artificial neural networks (ANN) based on machine learning. In contrast to earlier strategies, the suggested solution combines bagging and boosting to teach bagged–boosted (BB-ANN). According to the authors, the proposed method reduces bias and variance when compared to a single ANN, boosted ANN, and bagged ANN using actual data. In addition, the authors demonstrate that it minimizes predicting errors when compared to current approaches.
In [20], an AS-GCLSSVM hybrid model that combines autocorrelation function (ACF) and least squares support vector machines (LSSVM) is constructed to forecast the electricity demand of Australia. ACF selects interesting input variables and LSSVM predicts. LSSVM parameters are tuned by Grey Wolf Optimization Algorithm (GWO) and cross-validation. The suggested model forecasts the next week’s half-hour power load, and when compared to benchmark models, the experimental findings reveal it to be a very successful strategy. Deep learning-based forecasting is used in [21] to predict power demand. As a result, an improved support vector machine (ISVM) and extreme learning machine (ELM) are used for classification and forecasting, as well as feature selection utilizing the hybrid feature selector and feature extraction. A meta-heuristic method is used to adjust ELM hyper-parameters. The simulation findings show that the novel techniques outperform those deemed state of the art. Indonesian electricity load is modeled by [22]. The article proposes a hybrid method that consists of singular spectrum analysis (SSA), linear recurrent formula (LRF), weighted fuzzy time series (WFTS), and ANN. Empirical analysis showed that the SSA-LRF-NN approach, which is based on the RMSE and MAPE, is the most suited method for predicting the future values of electrical load series.
A strong deep learning model, the N-BEATS neural network, is used in [23] to illustrate how well it performs in midterm load forecasting over 35 European nations. Based on 35 monthly European power demand time series, the methodology is compared against 10 baseline methodologies including machine learning, traditional statistical methods, and hybrid approaches. According to the results of the empirical investigation, the suggested neural network surpasses all rivals in terms of accuracy and prediction bias.
The dynamic and fuzzy time series (D-FTS) methodology is hybridized by [24] to model the midterm electricity loads of Seoul. The hybrid approach is applied to the household, public, service, and industrial sectors independently in order to allow various reactions from each load sector. Researchers found that the suggested model is more accurate in its prediction and less than 3% off the mark when it comes to the actual monthly power load for each sector.
Dudek and Pelka [25] trained pattern similarity-based machine learning algorithms to forecast mid-term electricity loads of 35 European countries. There are four models considered by the authors: fuzzy neighborhood, nearest-neighbor, general regression neural network, and kernel regression. Three alternative approaches were offered. A fundamental one- and two-hybrid solution based on similarity and statistical methodologies. The suggested models surpass both conventional statistical and machine learning models in terms of optimization ease, simplicity, and accuracy. The most accurate strategy was a combination of similarity-based algorithms and exponential smoothing. Using phase space reconstruction (PSR) and SVM approaches, Li and Roozitalab [26] provided a multistep forecasting strategy. The model can forecast configurable stages of future load without divergence of inaccuracy and so has significant engineering application relevance. Applied to the European Network on Intelligent Technologies (ENIT) dataset, the findings suggest that the multistep implemented model is more accurate and resilient than earlier techniques. The approach is very simple to use and may be used with other sophisticated methods to improve performance. Baek [27] offered an RNN-based forecasting approach for mid-term daily peak demand. A recurrent-type ANN application and input data substitution for special days are proposed in the research as a solution to these challenges in mid-term load forecasting. During heat waves, the suggested RNN performs well in terms of predicting rapid and nonlinear demand increases. The suggested RNN’s performance and efficacy are shown via case studies using South Korean load data.
Li et al. [28] offered a mid-term load forecasting approach based on manifold learning (ML) that can identify the underlying components of load changes to assist enhance forecasting accuracy and greatly cut computation time. In comparison to linear dimensionality reduction techniques, ML has more nonlinear feature extraction capabilities and is better suited for load data with nonlinear features. In the low-dimensional space formed by manifold learning, LSTM neural networks are also used to develop forecasting models. The suggested approach is evaluated using New England datasets, and load forecasting is performed on different ranges of time intervals. The numerical findings demonstrate that the suggested strategy outperforms numerous mature solutions in the mid-term time scale.
Ahn et al. [29] proposed a 12-month SARIMA-based forecasting technique for 16 South Korean regions. Mohammed and Al-Bazi [30] improved an ANN model with an ABPA for forecasting long-term power load demand. To account for behavioral differences between training and future input datasets, ABPA incorporates unique forecasting formulations. The proposed innovation is based on the Multi-Layer Perceptron (MLP) model architecture and its standard Backpropagation Algorithm (BPA). Adjustment variables are used to smooth out behavior variations across the training and test datasets. The proposed ABPA, including the adjustment factor, enables current ANN techniques to anticipate long-term energy demands.
In [31], machine learning strategies such as artificial ANN, MLR, ANFIS, and SVM were used to figure out how much electricity Cyprus needs and what criteria should be used for power generation. Long-term and short-term data were used to analyze power use in 2016 and 2017. Long-term and short-term research revealed that SVM and ANN outperformed other ML approaches in terms of producing more accurate and dependable results for Cyprus’s time series forecasting criterion for electricity production. A brief representation of the above-mentioned literature can be found in Table 1.
Specifically, this paragraph is dedicated to studies on the Turkish electrical market, the majority of which are forecasts for the short-term. Bozkurt et al. [32] compared ANN and seasonal autoregressive integrated moving average (SARIMA). Model performances were observed over an average of 12 test weeks, and ANN generated 1.80% mean absolute percentage error (MAPE), outperforming SARIMA, which produced 2.60% MAPE. The authors conclude that the ANN model is more appropriate for the Turkish market than the SARIMA model. SARIMA, on the other hand, outperforms ANN in certain situations, particularly when it comes to predictions following holidays. Çevik and Çunkaş [3] proposed to forecast short-term electricity loads by utilizing fuzzy logic and an adaptive neuro-fuzzy inference system (ANFIS). In the study, historical data were evaluated, and weekdays are classified based on their load characteristics. Then, as inputs, historical load, temperature differential, and season are used, and the hourly load projection is conducted over one year. Using extremely large test data sets over one year, this research demonstrates that fuzzy logic may provide excellent outcomes.
The artificial neural networks (ANNs) are used in [33] to estimate the short-term load in Düzce, Turkey. The data from April were used as a baseline, and the estimations were created based on the input results from that month. As a consequence of this research, it has been discovered that ANN is capable of accurately forecasting load consumption while dealing with nonlinear data. Yukseltan et al. [34] used a linear model to build a technique for estimating hourly demand on yearly, weekly, and daily timescales utilizing harmonics and seasonal modulation of diurnal periodic oscillations. There is no use of meteorological or economic data in the suggested model, which is exclusively based on sinusoidal fluctuations and anticipates hourly changes. Data from the Turkish electricity market between 2012 and 2014 were used to model demand across the daily and weekly timeframes.
Another study based on Fourier transforms is performed by [2] to model Turkish electricity load. The study compares ARIMA and harmonic regression. The results showed that the model’s predicting ability for Turkish electricity consumption seems to be superior to that of the classic time series model. For long-term, mid-term, and short-term load forecasting in the Turkish electricity distribution network, Nalcaci et al. [1] proposed three models based on multivariate adaptive regression splines (MARS), ANN, and LR. Model predictions are based on wind, humidity, day of the year (holiday, summer, weekday), and temperature data. The MARS model outperforms the ANN and LR models in terms of accuracy and stability. Four distinct ANN models were constructed in [35], and the best one was chosen to simulate the impacts of seasonality and the trend of monthly Turkish electricity load. Furthermore, the chosen ANN model was compared to the SARIMA model to improve the ANN model’s acceptance and dependability. The ANN model, which can produce effective and high-accuracy forecasts based on performance metrics, was used to forecast Turkey’s monthly power consumption between 2015 and 2018.
Nature-inspired approaches are employed in [36] to assist fuzzy models in forecasting the quantity of Turkey required in the future. Ant colony optimization (ACO) and a genetic algorithm were used to improve the suggested models (GA). The scientists utilized historical hourly load consumption and temperature data acquired between 2011 and 2014 to train and test the new systems. The authors discovered that the suggested models may increase the accuracy of hourly short-term load predictions during the experiments. Based on least square SVM and ARIMA, a hybrid model is proposed in [37]. Results from this hybrid technique are compared to multiple linear regression (MLR), ARIMA, government predictions, and comparable research in the literature. Moreover, it is used to anticipate Turkey’s projected net power consumption until 2022. Findings show that the suggested model may provide more accurate and dependable predictions. It also reacts better to certain unexpected responses in the time series.
Yukseltan et al. [38] offered a feedback-based forecasting system that uses the current hour’s inaccuracy to update the estimate for the following hour. In the Turkish electricity market from 2012 to 2017, the suggested technique offers a strong tool to forecast demand on an hourly, daily, and annual basis using only historical demand data. The hourly forecasting errors in demand are 0.87 percent, 2.90 percent, and 3.54 percent, respectively, in the MAPE norm. To improve the accuracy of the Fourier series expansion predictions, an autoregressive (AR) model is utilized. A summary of the Turkish electricity load forecasting can be found in Table 2.
As Table 2 suggests there are many valuable studies in forecasting the electricity loads of Turkey. These studies are mainly based on short-term forecasting. The proposed algorithms are ANN, SARIMA, LR-FS, HR, and simple LR. Neither of them proposed to use deep neural networks in a univariate sense. These studies are a precious part of the literature. We tried to extend and utilize new algorithms that are powerful in both short-term forecasting and mid-term forecasting. Shortly said, we offer a novel technique for forecasting Turkish power demand over a multi-step time horizon. Our research, to the best of our knowledge, is the first to look at and compare different forecasting algorithms for both short and mid-term load forecasting of Turkey. In addition, it should be noted that the suggested model is based on a univariate case. To make use of this characteristic, it only requires data received directly from the examined time series itself. As a result, it may be utilized in any place and not only in the research location.

3. Theoretical Background

Deep learning (DL) is a subset of machine learning (ML) that is inspired by brain structure. It attempts to imitate the network of neurons found in the human brain. The human brain has billions of neurons. Neurons are in charge of transmitting electrical and chemical messages. ANN is a mathematical model that simulates a neural network. Each neuron is in charge of weighting and summing the incoming information and pulsing it to other neurons through a non-linear function (activation function). The input layer, the hidden layer, and the output layer are the three layers of neurons in a basic ANN. The input values (features) are taken by the input layers, and these values are passed through to the hidden levels through synapsis. The inputs are weighted by synapsis. All of the weighted inputs are summed in the hidden layers, and then an activation function is applied. The altered weighted total of the inputs is then pulsed to additional neurons. Finally, the output layers provide a value. In time series analysis, ANN offers several benefits. Unlike traditional autoregressive moving average (ARIMA) models, it does not need analyzed data to be stationary. It may also employ non-linear activation functions to better simulate complicated non-linear systems. The authors of [39] provide further theoretical context.
By training on examples, artificial neural networks provide a feasible method for forecasting a vector-valued, real-valued, or binary output. It is categorized as supervised learning. In supervised learning, the algorithm is fed inputs with labels. The algorithm then forecasts the proper outcome using the cases provided by the user. The network may be utilized for regression with real-valued target functions as well as classification with binary goal functions. ANN is made up of layers that are linked together. The layers comprise artificial neurons, which are also known as nodes or units. An input layer, an output layer, and multiple hidden layers comprise multilayer feedforward neural networks. If the network does not include hidden layers, it is termed as a Perceptron and is used to anticipate linearly behaved situations, while a feed-forward neural network (FFNN) is employed in many nonlinear forecast problems. RNNs are referred to as such when they include feedback connections in the model.
For example, RNNs are utilized for time-series data, text, and picture classification. RNNs are a form of neural network. In networks with loops, information may be preserved and re-used throughout time. Grid-based data processing is its specialty. Inputs can be used as outputs, but hidden states may be maintained. To put it another way, the network has feedback loops that may be utilized for predicting purposes. One way to conceive of an RNN is that it is made up of many identical networks that all communicate with one another by sending messages to each other. RNN suffers from the issue of vanishing gradients. There is a fresh approach to this problem proposed by [40]. Long Short-Term Memory is an RNN extension. Essentially, it is an RNN variant capable of learning about long-term associations. An LSTM representation is shown in Figure 1.
In the case of LSTMs, the problem of long-term dependency is purposely avoided. Long-term memory is a natural state of things for them, and they do not have to exert any efforts to maintain it. In addition, as seen in Figure 1, LSTM introduces a new parameter ct, which denotes the memory cell and is utilized to encode information up to the time t. The activity of a memory cell is governed by three gates: ft, it, and it, which are referred to as the input gate, forget gate, and output gate, respectively, in the circuit diagram. The equations for the three gates are as follows.
f t = sigmoid V f h t 1 + U f x t + b f ,
i t = sigmoid V i h t 1 + U i x t + b i ,
o t = sigmoid V o h t 1 + U o x t + b o .
The rest of the updating equations are as follows.
h t = o t tanh c t ,
c t = f t c t 1 + i t d t ,
Component-wise multiplication is denoted by *. To add information to the cell, the input gate adds it, the forget gate removes it, and the output gate chooses information from the cell to be utilized as input in the prior step. The first forget gate acquires information at epoch t as a function of the input xt and the previous hidden layer ht−1. If the forget gate’s value is close to one, the last memory cell ct−1 will be retained. Otherwise, the data are deleted. Second, the new information is combined with the old concealed state to generate the input gate i t . It is turned into a memory cell to create a new c t . Finally, the output gate determines which information will be utilized to create the next concealed state. More information on the algorithm’s architecture may be found in [41].
GRU is also presented as a solution for the vanishing gradient issue, similarly to how LSTM works. Sherstinsky [42] presented an extension to the LSTM. The system’s recurrent units can handle long-term dependencies across a broad range of periods. In the GRU algorithm, the input and forgotten gates of the LSTM are coupled with a single update gate, which serves as both the input and forgotten gates. Furthermore, the cell states and the hidden states are combined in the method developed by [43]. The representation of a GRU cell can be seen in Figure 2.
The architecture has been enhanced by the addition of two additional gates. The two sorts of gates are reset gates and update gates. The gates are used to store information and transfer it ahead as needed. The following is the model for GRU that may be written utilizing the new gates.
r t = sigmoid U r h t 1 x t ,
z t = sigmoid U z h t 1 x t ,
h ˜ t = tanh U t r t h t 1 x t
h t = 1 z t h t 1 + z t h t
GRU’s performance is boosted by the reset and the update, which also saves time [44]. It is up to the reset gate and hidden layer to decide whether or not the knowledge gleaned from the prior state will be lost. Data parsing has had a significant impact on the model’s overall performance and speed. Please refer to [42] for further in-depth details.
CNN are specific types of networks that function very well when dealing with data that possess a grid-type architecture, such as time-series data, images, and streaming videos. The mathematical process that gave origin to the network’s name is referred to as “convolution.” CNN performs convolution. Then, pooling, normalizing, and completely connected layers follow, each with the main purpose of multiplication, dot product, or ReLU. The first layer in CNN is the convolutional layer. Convolutional layers convolve the input and transmit the output to the next layer. This is analogous to a neuron’s reaction to a particular stimulus in the visual cortex. Each convolutional neuron only processes information for its receptive field. Although fully linked feedforward neural networks may be used to learn features and categorize data, they are often unfeasible for bigger inputs such as high-resolution photos. The second layer is the pooling layer. Pooling layers reduce the size of data by combining the outputs of neuron clusters at one layer into a single neuron at the next layer. This makes the data smaller. Local pooling brings together small groups of people. Global pooling affects all the neurons in the feature map, which means it affects all of them. Two types of pooling are used a lot: max and average. In the feature map, max-pooling takes the maximum value from each cluster of neurons. Average pooling only takes the average value from each cluster. The third layer is the flattening layer. It consists of taking the pooled feature map that was created during the pooling stage and converting it into a one-dimensional vector using a one-dimensional vector transform. This is performed in order to be able to feed them as inputs to the thick layer later on. The last layer is the fully connected layer. When all neurons in one layer are connected to all neurons in another layer, they work together to make sense of things. A multilayer perceptron neural network is the same as one that has a lot of different layers. To classify images, the flattened matrix proceeds through a layer that is fully connected. Figure 3 represents a CNN.
The majority of the time, this form of network is employed in image processing. Images are seen as a two-dimensional grid of pixels by the system. When applied to time-series data, this method is highly successful. As a result, it regards time-series data as a one-dimensional space of space intervals. For a more in-depth study of CNN, we recommend the book [45].

4. Data and Analysis

The dataset of this study is obtained from the publicly available website [46]. The data set contains the total electricity production of Turkey and is measured in MWh. It aggregates the electricity production from natural gas, lignite, river, import coal, wind, solar, fuel oil, geothermal, asphaltite coal, black coal, biomass, naphtha, LNG, import, and waste heat. The data set represents the real-time production of Turkey. We obtained the daily total electricity loads of Turkey for the periods between 5 January 2015, and 26 December 2021. The data set starts on Monday and ends on Saturday. The data set contains 2548 observations that correspond to 364 weeks. The electrical demands for all days in the forthcoming week are projected based on measurements from which complete weeks of lags are included in the prediction. We use the sliding window technique to forecast one to seven days ahead of observations. In each case, different lag lengths and numbers of nodes are used. The analysis is perfombed by using Python libraries Keras and Sklearn on the compiler Atom. Figure 4 summarizes the analysis procedure of the manuscript.
Our methodology consists of five steps as Figure 4 indicates, and these steps are summarized as follows:
  • Since high valued values may yield high weight values, high weight values are often unstable, resulting in poor learning performance and input sensitivity, which leads to greater generalization error. To overcome instability, the entire data set is transformed by using min–max scaler transformation:
    X n = X X m i n X m a x X m i n ,
    where X represents the observation, and X m i n and X m a x represent the lowest and highest values of the data set, respectively. After the transformation, the data set is divided into train and test sets. In the train set, the hyper parameter of the algorithm is decided, and on the test set, the algorithm runs on observations that are not used in training the data. The train set contains 312 weeks of observations while the test set contains 52 weeks of observation. The first six years act as the training data while the last year serves as the test set. Approximately, the training set contains 85% of the data while the test set contains 15%.
  • In the second step, LSTM, GRU, and CNN algorithms are trained on the training set. A different number of lag lengths (sliding window length) and the number of nodes were tested to achieve the highest performance. The sliding window lengths are as follows: 1 week, 2 weeks, 3 weeks, 4 weeks, half a year, a year, one and a half year, and two years. Because the proposed deep learning design is data-driven, it is not possible to talk about a separate architecture. As a consequence, different numbers of nodes are chosen to find the best structure because the number of nodes that is employed is determined by the size of the inputs. For each sliding window, we employed a total of 100 nodes. In each model, Adam was used as the optimizer, and the mean square error was used as the loss function.
  • In the third step, the lag lengths and number of nodes with the highest performance are run on the test set.
  • In the fourth step, forecasted values are back-transformed to the observed range.
  • In the last step, performance metrics are calculated to compare the performances of the algorithms. Three different performance metrics are utilized. These are mean absolute error (MAE), root mean squared error (RMSE), and coefficient of determination (R2). The formulas for each metric are given as follows:
    R 2 = 1 S S r e s S S t o t ,
    R M S E = M S E = 1 n i = 1 n y i f ^ i 2 = 1 n i = 1 n e i 2
    M A E = 1 n i = 1 n y i f ^ i = 1 n i = 1 n e i ,
    where S S r e s = i = 1 n y i f i ^ 2 = i = 1 n e i 2 ; it is known as the residual sum of squares and f i ^ is the anticipated output. S S t o t = i = 1 n y i y ¯ 2 is the total sum of the square, and y ¯ is the mean of the observed data. Small values of RMSE and MAE indicate good performance while a value near 1 for R 2 represents a good fit.
The computer’s operating system is Windows 10, with a CPU Intel(R) Core(TM) i7-10510U and 8 GB of RAM, on which the algorithms are trained. We used the Keras library with an Adam optimizer with a learning rate of 0.001 and an epoch size of 100 to train the models mentioned above. An early stop mechanism is also employed to obtain the best possible outcome on the test set. MSE is used as the loss function while the activation function is set as ReLU. In LSTM, the following layers are utilized: LSTM layer with shape of (1,100) while the dense layer shape has shape (1,7). The computation time of LSTM is 3 s for each step. We employed a fixed window size for CNN layers. Columns are assigned to features, while rows are assigned to lagged values. The computation time is 2 s for each step. The architecture of the CNN is convolution layer of size (None, 1,64), MaxPooling layer of size (None, 1,64), flatten layer of size (None, 64), dense layer (None, 100), and dense layer (None, 7). Finally, the computation time for GRU is 3 s for each step. GRU consists of GRU layer of size (1,100) and dense layer of (1,7). The performance metrics for LSTM for different lag lengths on the test set are given in Table 3.
Here, y t ^ represents the next-day forecast, y ^ t + 1 represents the two-day-ahead forecast, and in the same manner, y ^ t + 6 represents the seven days ahead forecast. Each panel represents the performance metrics when different lag lengths are used. For example, if we consider Panel A, we may summarize the forecasting procedure as x t 7 ; x t 6 , x t 1 are used to forecast y ^ t   ^ , y ^ t + 1 , y ^ t + 6 where x t represents the input values.
According to the results given in Table 3, the best performance is achieved when lag length is determined as 364. This case is given in Panel F. R2 for the day ahead forecast is calculated as 0.94, and it decreased to 0.73 for seven days ahead forecasts. Moreover, in this case, the algorithm achieves its lowest MAE and RMSE for each forecasted value. Table 4 represents the performance metrics of GRU on the test set.
According to the results given in Table 4, the best performance is achieved when lag length is determined as 364. The case is given in Panel F. The R2 for the day ahead forecast is calculated as 0.91, and it decreased to 0.58 for seven days ahead forecasts. Moreover, in this case, the algorithm achieves its lowest MAE and RMSE for each forecasted value. Table 4 represents the performance metrics of GRU. The difference between GRU and LSTM occurs in the mid-term forecast. As both tables indicate the performance of LSTM is better than GRU. The R2 of GRU in the mid-term decreases faster than in the LSTM case. Lastly, Table 5 represents the performance metrics of CNN.
In this case, according to the performance metrics, the best length is 546, which represents a year and a half. R2 for the next day forecast is calculated as 0.92, while for seven days ahead, it is calculated as 0.66. As in the case of GRU, R2 decreases as the length of the forecasting horizon increases. When CNN is compared to LSTM and GRU, it has the second-best performance according to the calculated performance metrics.

5. Discussion

The aim of this study is to forecast short-term to mid-term electrical usage utilizing deep learning algorithms such as LSTM, GRU, and CNN. These algorithms were selected for this investigation because they have been utilized effectively in various time-series studies. Moreover, the proposed models can handle entire data sequences as well as single data points. Although there are many powerful RNN algorithms, in this study, we employed LSTM because the LSTM cell increases long-term memory capacity in an even more efficient manner since it allows learning even more parameters. Moreover, it has the capacity of handling a large amount of non-linear data [47]. This makes it the most effective method of forecasting, particularly when there is a longer-term trend in the data set. We trained LSTM and others algorithms in the same way that we would estimate a time series model of Box–Jenkins. Algorithms, as a time series model, use the lags of the time series data that we are analyzing. The suggested methodology exclusively employs data derived solely from time-series itself. As a result, it is effective, straightforward, and forceful. The univariate structure of the methodology leads it to be utilized globally as well as locally
It should also be noted that we represented the loads in a univariate manner; thus, the information is entirely generated from the data itself, which increases the model’s efficacy while minimizing its overall complexity. To the best of our knowledge, this is the first research study that compares Turkey’s short-term and mid-term algorithms. Because of this, the suggested model is not limited to the Turkish market but may also be applied in any other market. It does not need the use of any exogenous variables or other information. It is worth mentioning that this is the first research to use deep learning algorithms to simulate the short-term Turkish power market. Table 6 summarizes the performance metrics of the proposed model for the best cases of each.
LSTM and GRU had the best performance when lag lengths are set as 364 while the best performance of CNN is achieved with 546 lag lengths. According to the empirical results provided in Table 6, the best algorithm when compared to CNN and GRU is found to be LSTM. It has an R2 of 0.94 in the short-term and 0.73 in the mid-term. Moreover, it is interesting to see that GRU and CNN have high R2 in the short-term forecasting but decrease gradually in mid-term forecasts. Thus, LSTM with a one-year lag can be used efficiently to model and forecast the daily Turkish electricity load of Turkey. The proposed model only needs its own lagged values; when compared to the other studied in the literature, it is more efficient and powerful. In some cases, we obtained very good forecasting results without using any exogenous variables such as temperature, precipitation, and other influencing factors.
The power of the proposed model not only comes from its univariate case but it can also handle multi-step forecasting with low computation costs. There are many valuable studies that attempt to forecast electricity loads of Turkey or stations located at Turkey. The next two paragraphs compare the results of our study with the literature on Turkish case, which uses ML, DL, or Fuzzy time series analyses. The first paragraph devoted to multivariate case, while the last paragraph is about the univariate case.
In their studies, Tosun et al. [33] proposed to forecast short-term electricity loads of Düzce, Turkey with ANN. The model used hour, temperature, previous temperature, and previous consumption as the features. The proposed model is a multivariate methodology, and the results showed that, on average, the best R2 for hourly electricity consumption ranged in 0.927 and 0.978. Bozkurt et al. [32] also preferred multivariate modeling for short-term electricity loads. The feature set of ANN contains calendar date, previous load estimation plan, electricity price, weather, and currency. Each feature set also contains a different number of features. The total number of features to train ANN was 19. The performance metrics of MAPE ranged from 0.98 to 3.26. Luy et al. [36] used temperature as a feature of the proposed algorithms to forecast short-term electricity loads of Turkey. The other features are the last day of consumption, the last week of consumption, the weekly load trend, and the weekly air temperature trends. In the best case, MAPE is calculated as 3.389. Kaytez [37] used multivariate methodology to forecast net electricity consumption of Turkey and MAPE of the best case ranging from 0.971 to 1445. Several environmental variables are used by [1] to forecast loads of Turkey in short- and mid-term periods. In the best case, the performance metric of R2 is calculated at 0.907.
Hamzaçebi et al. [35] used ANN in a univariate case to forecast monthly electricity loads of Turkey. Different ANN combinations are compared, and the best model had RMSE ranged from 438 to 572.95.
In light of the above two paragraphs, our study is as powerful as the multivariate cases according to Table 6. Moreover, we introduced a model in the univariate case that can be used in both short- and mid-term forecasting. To the best of our understanding, there is only one univariate time series methodology that utilizes ANN to forecast monthly electricity loads of Turkey. We would like to emphasize once again that the proposed model is based on a univariate case and is capable of forecasting daily electricity loads of Turkey in multiple steps. Thus, it can be used to forecast and model Turkish electricity loads to have better projections and planning.

6. Conclusions

Electricity is a critical indicator of human life and the health of a country’s economic structure. When developing economic planning, it is vital to have precise projections of power consumption levels. Accurate energy demand forecasting is crucial for decision-makers and power-generating firms when it comes to policy creation and power generation planning. Several approaches have been explored in the past to increase peak load forecasting accuracy. The data set used in this study is a daily data collection used to assess electrical demands. In classical time series analysis, despite the fact that the methodology’s predicting ability has been shown, some assumptions must be satisfied, such as the assumption of stationarity. It should be emphasized that the Box–Jenkins type models are linear time series models, as opposed to the other types of deep learning and machine learning algorithms. In the literature, there are many valuable works that compare ARIMA-type models with the others. For example, Akdi et al. [2] showed that HR is more powerful than AR or Tokgöz, and Ünal [48] compared ARIMA with deep neural networks and showed that ARIMA had the lowest MAPE. Instead of using the Box–Jenkins approach, we used a very customized model of deep learning in our study. The proposed model has no assumptions, such as the stationarity of the investigated model or the residuals terms of the model that should be normally distributed and are uncorrelated. This feature can be shown as the strength of the proposed model.
Although the suggested model’s primary strength is its univariate structure, further research into the relationships between electrical demand and meteorological parameters, such as those described in [1], will be possible via the use of deep learning algorithms. Additionally, as mentioned in [49], the combination of time series approaches and deep learning algorithms to forecast electricity consumption may be of interest; on the other hand, as in [47], the effects of wavelet transformation can be investigated. To anticipate power demands in Turkey, it is also possible to study the hybrid approaches of [5] and the machine learning methods of [50], which may be used in both uni- and multivariate contexts. Moreover, the proposed models are utilized and investigated as standalone. It is also possible to use them together as hybrid models or in an ensemble manner as in [51]. Since there is no exact rule to decide the hyper parameters of the proposed algorithm. This can be shown as the weakness of the models and the optimization of the proposed algorithms can be performed by using the metaheuristic algorithms of [52]. We leave these ideas as future research opportunities.
Weather and seasonal impacts have a direct influence not only on load demand but also on the utilization of certain renewable energy sources in Turkey’s power grid; hence, the load forest must be considered as a collection of factors including people’s daily routines [53]. It may also be interesting to investigate the influence of weather-related time series and people’s daily habits on power demand forecasting as a challenge. The holiday weekdays and weekends affect power use in diverse ways, as has been shown in several research studies [54]. It is obvious that the forecasting ability of the models will be improved by including these elements as the features of the algorithms. In addition, the impact of these variables on the forecasting capacity of the model should be investigated with different data pre-processing techniques [55].
In this study, the main aim was to train and test different deep neural networks to forecast short-term to mid-term forecasting of Turkish electricity load in a univariate sense. The proposed models are investigated to forecast 1 to 7 data points simultaneously and it was observed that, overall, LSTM has the best performance compared to CNN and GRU. Long-term forecasting is more challenging than short- and mid-term forecasting. There are three categories of issues in long-term power demand forecasting: what technical and economic aspects to include, what regional and temporal scales to pick, and to what degree long-term and short-term uncertainty should be taken into consideration [12]. It might be more challenging to model long-term electricity loads by univariate time series methodologies. The other influencing variables should be considered in this task.
The development of effective energy forecasting models is critical in the development of energy policy, which may involve planning, production, pricing, and consumption. As illustrated in this research study, determining the appropriate lag length and plugging in models improve the accuracy of forecasts and predictions. In conclusion, LSTM should be viewed as a potent tool for electrical load forecasting in the short- and mid-term for both short- and mid-term forecasting. Because this model more closely matches the data than GRU and CNN models, it will be more helpful in developing policies based on energy demand. In this context, the findings of this study’s methodology provide valuable evidence for policymakers on how to interfere in electricity markets in a manner that legitimizes evidence-based policymaking, which is critical in today’s world.

Funding

This research received no external funding.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: https://seffaflik.epias.com.tr/transparency/ (accessed on 10 January 2022).

Acknowledgments

We appreciate the associate editor’s and four anonymous reviewers’ helpful comments and revisions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nalcaci, G.; Özmen, A.; Weber, G.W. Long-term load forecasting: Models based on MARS, ANN and LR methods. Cent. Eur. J. Oper. Res. 2019, 27, 1033–1049. [Google Scholar] [CrossRef]
  2. Akdi, Y.; Gölveren, E.; Okkaoğlu, Y. Daily electrical energy consumption: Periodicity, harmonic regression method and forecasting. Energy 2020, 191, 116524. [Google Scholar] [CrossRef]
  3. Çevik, H.H.; Çunkaş, M. Short-term load forecasting using fuzzy logic and ANFIS. Neural Comput. Appl. 2015, 26, 1355–1367. [Google Scholar] [CrossRef]
  4. Yazici, I.; Beyca, O.F.; Delen, D. Deep-learning-based short-term electricity load forecasting: A real case application. Eng. Appl. Artif. Intell. 2022, 109, 104645. [Google Scholar] [CrossRef]
  5. Akbal, Y.; Ünlü, K.D. A deep learning approach to model daily particular matter of Ankara: Key features and forecasting. Int. J. Environ. Sci. Technol. 2021, 1–17. [Google Scholar] [CrossRef]
  6. Pathan, R.K.; Biswas, M.; Khandaker, M.U. Time series prediction of COVID-19 by mutation rate analysis using recurrent neural network-based LSTM model. Chaos Solitons Fractals 2020, 138, 110018. [Google Scholar] [CrossRef]
  7. Berradi, Z.; Lazaar, M. Integration of principal component analysis and recurrent neural network to forecast the stock price of casablanca stock exchange. Procedia Comput. Sci. 2019, 148, 55–61. [Google Scholar] [CrossRef]
  8. Su, H.; Zio, E.; Zhang, J.; Xu, M.; Li, X.; Zhang, Z. A hybrid hourly natural gas demand forecasting method based on the integration of wavelet transform and enhanced Deep-RNN model. Energy 2019, 178, 585–597. [Google Scholar] [CrossRef] [Green Version]
  9. Kumar, J.; Goomer, R.; Singh, A.K. Long short term memory recurrent neural network (LSTM-RNN) based workload forecasting model for cloud datacenters. Procedia Comput. Sci. 2018, 125, 676–682. [Google Scholar] [CrossRef]
  10. Haseeb, I.; Armghan, A.; Khan, W.; Alenezi, F.; Alnaim, N.; Ali, F.; Muhammad, F.; Albogamy, F.R.; Ullah, N. Solar Power System Assessments Using ANN and Hybrid Boost Converter Based MPPT Algorithm. Appl. Sci. 2021, 11, 11332. [Google Scholar] [CrossRef]
  11. Fallah, S.N.; Ganjkhani, M.; Shamshirband, S.; Chau, K.W. Computational intelligence on short-term load forecasting: A methodological overview. Energies 2019, 12, 393. [Google Scholar] [CrossRef] [Green Version]
  12. Lindberg, K.B.; Seljom, P.; Madsen, H.; Fischer, D.; Korpås, M. Long-term electricity load forecasting: Current and future trends. Util. Policy 2019, 58, 102–119. [Google Scholar] [CrossRef]
  13. Al Mamun, A.; Sohel, M.; Mohammad, N.; Sunny, M.S.H.; Dipta, D.R.; Hossain, E. A comprehensive review of the load forecasting techniques using single and hybrid predictive models. IEEE Access 2020, 8, 134911–134939. [Google Scholar] [CrossRef]
  14. Pai, P.F.; Hong, W.C. Support vector machines with simulated annealing algorithms in electricity load forecasting. Energy Convers. Manag. 2005, 46, 2669–2688. [Google Scholar] [CrossRef]
  15. Zhang, J.; Wei, Y.M.; Li, D.; Tan, Z.; Zhou, J. Short term electricity load forecasting using a hybrid model. Energy 2018, 158, 774–781. [Google Scholar] [CrossRef]
  16. Dedinec, A.; Filiposka, S.; Dedinec, A.; Kocarev, L. Deep belief network based electricity load forecasting: An analysis of Macedonian case. Energy 2016, 115, 1688–1700. [Google Scholar] [CrossRef]
  17. Takeda, H.; Tamura, Y.; Sato, S. Using the ensemble Kalman filter for electricity load forecasting and analysis. Energy 2016, 104, 184–198. [Google Scholar] [CrossRef]
  18. Ghadimi, N.; Akbarimajd, A.; Shayeghi, H.; Abedinia, O. Two stage forecast engine with feature selection technique and improved meta-heuristic algorithm for electricity load forecasting. Energy 2018, 161, 130–142. [Google Scholar] [CrossRef]
  19. Khwaja, A.S.; Anpalagan, A.; Naeem, M.; Venkatesh, B. Joint bagged-boosted artificial neural networks: Using ensemble machine learning to improve short-term electricity load forecasting. Electr. Power Syst. Res. 2020, 179, 106080. [Google Scholar] [CrossRef]
  20. Yang, A.; Li, W.; Yang, X. Short-term electricity load forecasting based on feature selection and Least Squares Support Vector Machines. Knowl. Based Syst. 2019, 163, 159–173. [Google Scholar] [CrossRef]
  21. Ahmad, W.; Ayub, N.; Ali, T.; Irfan, M.; Awais, M.; Shiraz, M.; Glowacz, A. Towards short term electricity load forecasting using improved support vector machine and extreme learning machine. Energies 2020, 13, 2907. [Google Scholar] [CrossRef]
  22. Sulandari, W.; Lee, M.H.; Rodrigues, P.C. Indonesian electricity load forecasting using singular spectrum analysis, fuzzy systems and neural networks. Energy 2020, 190, 116408. [Google Scholar] [CrossRef]
  23. Oreshkin, B.N.; Dudek, G.; Pełka, P.; Turkina, E. N-BEATS neural network for mid-term electricity load forecasting. Appl. Energy 2021, 293, 116918. [Google Scholar] [CrossRef]
  24. Lee, W.J.; Hong, J. A hybrid dynamic and fuzzy time series model for mid-term power load forecasting. Int. J. Electr. Power Energy Syst. 2015, 64, 1057–1062. [Google Scholar] [CrossRef]
  25. Dudek, G.; Pełka, P. Pattern similarity-based machine learning methods for mid-term load forecasting: A comparative study. Appl. Soft Comput. 2021, 104, 107223. [Google Scholar] [CrossRef]
  26. Li, G.; Li, Y.; Roozitalab, F. Midterm load forecasting: A multistep approach based on phase space reconstruction and support vector machine. IEEE Syst. J. 2020, 14, 4967–4977. [Google Scholar] [CrossRef]
  27. Baek, S.M. Mid-term load pattern forecasting with recurrent artificial neural network. IEEE Access 2019, 7, 172830–172838. [Google Scholar] [CrossRef]
  28. Li, J.; Wei, S.; Dai, W. Combination of Manifold Learning and Deep Learning Algorithms for Mid-Term Electrical Load Forecasting. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–10. [Google Scholar] [CrossRef] [PubMed]
  29. Ahn, B.H.; Choi, H.R.; Lee, H.C. Regional Long-term/Mid-term Load Forecasting using SARIMA in South Korea. J. Korea Acad. Ind. Coop. Soc. 2015, 16, 8576–8584. [Google Scholar] [CrossRef] [Green Version]
  30. Mohammed, N.A.; Al-Bazi, A. An adaptive backpropagation algorithm for long-term electricity load forecasting. Neural Comput. Appl. 2022, 34, 477–491. [Google Scholar] [CrossRef] [PubMed]
  31. Solyali, D. A comparative analysis of machine learning approaches for short-/long-term electricity load forecasting in Cyprus. Sustainability 2020, 12, 3612. [Google Scholar] [CrossRef]
  32. Bozkurt, Ö.Ö.; Biricik, G.; Tayşi, Z.C. Artificial neural network and SARIMA based models for power load forecasting in Turkish electricity market. PLoS ONE 2017, 12, e0175915. [Google Scholar] [CrossRef] [Green Version]
  33. Öztürk, A.; Taşpinar, F. Short term load forecasting for Turkey energy distribution system with artificial neural networks. Tehnički Vjesnik 2019, 26, 1545–1553. [Google Scholar] [CrossRef]
  34. Yukseltan, E.; Yucekaya, A.; Bilge, A.H. Forecasting electricity demand for Turkey: Modeling periodic variations and demand segregation. Appl. Energy 2017, 193, 287–296. [Google Scholar] [CrossRef]
  35. Hamzaçebi, C.; Es, H.A.; Çakmak, R. Forecasting of Turkey’s monthly electricity demand by seasonal artificial neural network. Neural Comput. Appl. 2019, 31, 2217–2231. [Google Scholar] [CrossRef]
  36. Luy, M.; Ates, V.; Barisci, N.; Polat, H.; Cam, E. Short-term fuzzy load forecasting model using genetic–fuzzy and ant colony–fuzzy knowledge base optimization. Appl. Sci. 2018, 8, 864. [Google Scholar] [CrossRef] [Green Version]
  37. Kaytez, F. A hybrid approach based on autoregressive integrated moving average and least-square support vector machine for long-term forecasting of net electricity consumption. Energy 2020, 197, 117200. [Google Scholar] [CrossRef]
  38. Yukseltan, E.; Yucekaya, A.; Bilge, A.H. Hourly electricity demand forecasting using Fourier analysis with feedback. Energy Strategy Rev. 2020, 31, 100524. [Google Scholar] [CrossRef]
  39. Alpaydin, E. Introduction to Machine Learning; The MIT Press: Cambridge, MA, USA, 2020. [Google Scholar]
  40. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  41. Sherstinsky, A. Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar] [CrossRef] [Green Version]
  42. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  43. Dutta, A.; Kumar, S.; Basu, M. A gated recurrent unit approach to bitcoin price prediction. J. Risk Financ. Manag. 2020, 13, 23. [Google Scholar] [CrossRef] [Green Version]
  44. Zhang, Y.G.; Tang, J.; He, Z.Y.; Tan, J.; Li, C. A novel displacement prediction method using gated recurrent unit model with time series analysis in the Erdaohe landslide. Nat. Hazards 2021, 105, 783–813. [Google Scholar] [CrossRef]
  45. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; Available online: http://www.deeplearningbook.org (accessed on 5 January 2022).
  46. EPIAS. Seffaflik Platformu. 2021. Available online: https://seffaflik.epias.com.tr/transparency/uretim/gerceklesen-uretim/gercek-zamanli-uretim.xhtml (accessed on 10 January 2022).
  47. Stefenon, S.F.; Kasburg, C.; Nied, A.; Klaar, A.C.R.; Ferreira, F.C.S.; Branco, N.W. Hybrid deep learning for power generation forecasting in active solar trackers. IET Gener. Transm. Distrib. 2020, 14, 5667–5674. [Google Scholar] [CrossRef]
  48. Tokgöz, A.; Ünal, G. A RNN based time series approach for forecasting Turkish electricity load. In Proceedings of the 2018 26th Signal Processing and Communications Applications Conference (SIU), Izmir, Turkey, 2–5 May 2018. [Google Scholar]
  49. Manowska, A.; Rybak, A.; Dylong, A.; Pielot, J. Forecasting of Natural Gas Consumption in Poland Based on ARIMA-LSTM Hybrid Model. Energies 2021, 14, 8597. [Google Scholar] [CrossRef]
  50. Lee, M.H.L.; Ser, Y.C.; Selvachandran, G.; Thong, P.H.; Cuong, L.; Son, L.H.; Tuan, N.T.; Gerogiannis, V.C. A Comparative Study of Forecasting Electricity Consumption Using Machine Learning Models. Mathematics 2022, 10, 1329. [Google Scholar] [CrossRef]
  51. Stefenon, S.F.; Ribeiro, M.H.D.M.; Nied, A.; Yow, K.C.; Mariani, V.C.; dos Santos Coelho, L.; Seman, L.O. Time series forecasting using ensemble learning methods for emergency prevention in hydroelectric power plants with dam. Electr. Power Syst. Res. 2022, 202, 107584. [Google Scholar] [CrossRef]
  52. Stefenon, S.F.; Furtado Neto, C.S.; Coelho, T.S.; Nied, A.; Yamaguchi, C.K.; Yow, K.C. Particle swarm optimization for design of insulators of distribution power system based on finite element method. Electr. Eng. 2022, 104, 615–622. [Google Scholar] [CrossRef]
  53. Hoori, A.O.; Al Kazzaz, A.; Khimani, R.; Motai, Y.; Aved, A.J. Electric load forecasting model using a multicolumn deep neural networks. IEEE Trans. Ind. Electron. 2019, 67, 6473–6482. [Google Scholar] [CrossRef]
  54. Yadav, S.; Jain, A.; Sharma, K.C.; Bhakar, R. Load Forecasting for Rare Events using LSTM. In Proceedings of the 2021 9th IEEE International Conference on Power Systems (ICPS), Kharagpur, India, 16–18 December 2021. [Google Scholar]
  55. Kong, W.; Dong, Z.Y.; Wang, B.; Zhao, J.; Huang, J. A practical solution for non-intrusive type II load monitoring based on deep learning and post-processing. IEEE Trans. Smart Grid 2019, 11, 148–160. [Google Scholar] [CrossRef]
Figure 1. Representation of an LSTM cell.
Figure 1. Representation of an LSTM cell.
Electronics 11 01524 g001
Figure 2. Representation of a GRU cell.
Figure 2. Representation of a GRU cell.
Electronics 11 01524 g002
Figure 3. Representation of a CNN.
Figure 3. Representation of a CNN.
Electronics 11 01524 g003
Figure 4. Summary of the forecasting procedure.
Figure 4. Summary of the forecasting procedure.
Electronics 11 01524 g004
Table 1. Brief representation of some recent studies in forecasting electricity load.
Table 1. Brief representation of some recent studies in forecasting electricity load.
MethodDatasetReferenceForecast Horizon
SVM-SATaiwan[14]Short-term
IEMD, ARIMA, WNN, FOAAustralia and New York[15]
DBNMacedonia[16]
rNN, ENNAustralia, North America, Pennsylvania-New Jersey-Maryland[18]
EnKFTokyo[17]
BB-ANNNew England[19]
AS-GCLSSVMAustralia[20]
ISVM, ELMNew England[21]
SSA-LRF, WFTS, ANNIndonesia[22]
N-BEATS35 European Countries[23]Mid-term
D-FTSSeoul[24]
ML methods35 European Countries[25]
PSR, SVMENIT[26]
RNNSouth Korea[27]
ML, LSTMNew England[28]
SARIMASouth Korea[29]Long-term
ANNIraq[30]
ANN, MLR, ANFIS, SVMCyprus[31]
Table 2. Brief representation of some recent studies in forecasting the electricity load of Turkey.
Table 2. Brief representation of some recent studies in forecasting the electricity load of Turkey.
MethodForecast HorizonReferenceDataset
ANN, SARIMAShort-term[32]Turkey
ANFIS[3]
ANN[33]
LR-FS[34]
HR[2]
ANN, SARIMA[35]
Fuzzy-ACO-GA[36]
SVM-ARIMA[37]
LR-FS[38]
LR-FSMid-term[34]
MARS, ANN, LRLong-term[1]
Table 3. Performance metrics of LSTM on train set for different lag size.
Table 3. Performance metrics of LSTM on train set for different lag size.
Estimated Values
y ^ t y ^ t + 1 y ^ t + 2 y ^ t + 3 y ^ t + 4 y ^ t + 5 y ^ t + 6
Panel A: Lag length = 7
MAE15,699.0524,122.2329,272.7932,555.0434,794.4636,880.2638,159.13
RMSE24,521.7238,779.3945,610.3050,527.0453,893.3456,125.7657,531.93
R20.920.800.730.670.620.590.57
Panel B: Lag length = 14
MAE15,581.2022,976.7627.896.9931,732.3434,604.0336,783.6437,949.07
RMSE23,701.2136,435.1044,048.8250,052.4354,024.8056,539.065793.76
R20.930.830.750.670.620.590.56
Panel C: Lag length = 21
MAE19,410.9029,598.3736,034.6840,335.4541,303.4643,352.6743,856.13
RMSE31,150.2247,841.7660,509.5968,103.5468,850.8165,431.7863,913.05
R20.870.700.520.400.390.450.47
Panel D: Lag length = 28
MAE17,153.5226,952.2432,011.7535,466.3138,208.3838,528.4039,586.90
RMSE26,279.8140,563.7448,472.5153,813.5257,249.4458,689.0359,796.33
R20.910.790.690.620.570.550.54
Panel E: Lag length = 182
MAE17,451.3125,541.3128,991.6131,138.2232,718.4633,835.2636,008.39
RMSE2509.9838182.1643,601.3847,018.5948,844.4449,449.2351,886.17
R20.920.810.750.710.690.680.65
Panel F: Lag length = 364
MAE15,502.6123,570.2427,577.3030,641.9332,041.7832,760.7234,045.72
RMSE21,561.5732,538.2137,346.1640,974.1042,506.32243,478.0645,782.79
R20.940.860.820.780.770.760.73
Panel G: Lag length = 546
MAE16,262.1324,255.6830,396.8132,241.5632,610.8935,494.7136,846.01
RMSE22,835.5833,540.9640,476.9643,986.3645,137.0947,090.8648,918.94
R20.930.850.790.750.740.710.69
Panel H: Lag Length = 728
MAE16,976.5924,364.7628,113.4831,015.4334,180.1536,634.8438,668.64
RMSE23,308.7433,971.5039,104.1142,854.7146,421.8549,947.7852,931.25
R20.930.850.800.760.720.680.64
Table 4. Performance metrics of GRU on train set for different lag size.
Table 4. Performance metrics of GRU on train set for different lag size.
Estimated Values
y ^ t y ^ t + 1 y ^ t + 2 y ^ t + 3 y ^ t + 4 y ^ t + 5 y ^ t + 6
Panel A: Lag length = 7
MAE19,394.8329,989.5537,397.9643,087.9047,997.7949,682.0950,972.89
RMSE29,685.1244,974.2554,601.3160,463.1664,933.4167,078.7768,605.49
R20.880.730.610.520.450.420.39
Panel B: Lag length = 14
MAE18,442.0830,479.9039,077.0647,732.6050,860.6751,966.7451,205.42
RMSE26,251.8044,369.6357,378.4367,591.3271,879.1973,737.6973,698.62
R20.910.740.570.410.330.300.29
Panel C: Lag length = 21
MAE19,200.6531,087.7440,138.7147,851.9754,597.7257,569.0558,608.46
RMSE27,639.0743,832.2054,643.5664,115.1472,562.2077,055.5278,916.40
R20.900.750.610.460.320.230.19
Panel D: Lag length = 28
MAE18,268.7330,688.0739,384.8647,825.3555,541.0960,740.5066,4387.58
RMSE25,493.6643,038.8855,662.8865,935.2874,158.9578,728.5283,011.19
R20.910.760.600.430.290.200.10
Panel E: Lag length = 182
MAE20,759.1930,960.1330,643.2732,815.7436,172.2537,763.6740,939.88
RMSE28,232.9942,474.9844,552.4647,638.7649,405.7548,905.5952,472.01
R20.900.760.740.700.680.690.64
Panel F: Lag length = 364
MAE20,281.4030,132.4638,391.2737,534.923866.2340,951.2744,929.31
RMSE26,900.1639,285.1848,824.4548,195.7051,052.0952,933.2056,855.68
R20.910.800.690.690.660.640.58
Panel G: Lag length = 546
MAE22,134.4334,811.6045,676.5652,662.0649,250.8048,575.7346,582.51
RMSE28,952.5944,469.7857,428.5066,416.1064,495.2264,762.9863,057.94
R20.890.740.570.460.460.460.48
Panel H: Lag Length = 728
MAE19,724.9939,265.0649,729.8949,852.4152,810.0045,557.8064,353.77
RMSE27,133.7550,171.9861,195.2461,860.4664,526.6057,388.5881,414.38
R20.900.670.510.500.460.570.14
Table 5. Performance metrics of CNN on train set for different lag size.
Table 5. Performance metrics of CNN on train set for different lag size.
Estimated Values
y ^ t y ^ t + 1 y ^ t + 2 y ^ t + 3 y ^ t + 4 y ^ t + 5 y ^ t + 6
Panel A: Lag length = 7
MAE16,356.0625,708.0731,697.2235,591.2437,120.3737,861.6139,063.72
RMSE25,765.2740,912.0049,607.6455,626.9458,202.8458,216.2459,015.31
R20.910.780.680.600.560.560.55
Panel B: Lag length = 14
MAE16,445.0324,700.0030,280.4735,430.5737,406.2538,861.4537,997.71
RMSE25,034.4039,139.9948980.2655,134.0657,333.2859,121.0858,697.51
R20.920.800.690.600.570.550.55
Panel C: Lag length = 21
MAE16,724.6326,348.332,906.1837,370.3738,553.6138,771.6839,881.52
RMSE2622.3740,865.1450,672.0156,788.2158,362.0857,924.4759,204.25
R20.910.780.660.580.560.570.54
Panel D: Lag length = 28
MAE18,348.5027,976.5133,386.0737,748.1341,008.7040,982.0540,650.68
RMSE26,852.1441,811.1751,001.6258,105.2061,217.1960,627.5860,271.08
R20.900.770.660.560.510.520.52
Panel E: Lag length = 182
MAE21,458.6529,987.7132,860.2837,032.2641,267.4339,066.8239,647.72
RMSE31,305.8544,841.7848,393.3354,752.0959,953.7957,048.4457,289.49
R20.870.740.690.610.530.580.57
Panel F: Lag length = 364
MAE19,893.0128,803.5234,274.1137,084.6139,495.8837,885.7436,249.55
RMSE29,312.5240,326.4847,292.9250,441.6454,319.9152,452.0950,747.24
R20.890.790.710.670.610.640.66
Panel G: Lag length = 546
MAE17,694.6127,589.2333,409.7837,503.4938,201.5439,553.4637,239.67
RMSE25,212.9539,388.3846,514.9352,304.8352,018.6153,394.2452,067.60
R20.920.800.720.640.650.630.65
Panel H: Lag Length = 728
MAE17,505.8927,472.7129,315.0232,728.2935,896.0835,859.1734,807.66
RMSE24,884.3238,493.6441,143.2346,708.9750,173.8750,275.1449,692.07
R20.920.810.780.720.670.670.67
Table 6. Summary of the performance metrics of the investigated models on train set.
Table 6. Summary of the performance metrics of the investigated models on train set.
Estimated Values
y ^ t y ^ t + 1 y ^ t + 2 y ^ t + 3 y ^ t + 4 y ^ t + 5 y ^ t + 6
Panel A: LSTM (Lag Length = 364)
MAE15,502.6123,570.2427,577.3030,641.9332,041.7832,760.7234,045.72
RMSE21,561.5732,538.2137,346.1640,974.1042,506.32243,478.0645,782.79
R20.940.860.820.780.770.760.73
Panel B: GRU (Lag length = 364)
MAE20,281.4030,132.4638,391.2737,534.923866.2340,951.2744,929.31
RMSE26,900.1639,285.1848,824.4548,195.7051,052.0952,933.2056,855.68
R20.910.800.690.690.660.640.58
Panel C: CNN (Lag length = 546)
MAE17,694.6127,589.2333,409.7837,503.4938,201.5439,553.4637,239.67
RMSE25,212.9539,388.3846,514.9352,304.8352,018.6153,394.2452,067.60
R20.920.800.720.640.650.630.65
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ünlü, K.D. A Data-Driven Model to Forecast Multi-Step Ahead Time Series of Turkish Daily Electricity Load. Electronics 2022, 11, 1524. https://doi.org/10.3390/electronics11101524

AMA Style

Ünlü KD. A Data-Driven Model to Forecast Multi-Step Ahead Time Series of Turkish Daily Electricity Load. Electronics. 2022; 11(10):1524. https://doi.org/10.3390/electronics11101524

Chicago/Turabian Style

Ünlü, Kamil Demirberk. 2022. "A Data-Driven Model to Forecast Multi-Step Ahead Time Series of Turkish Daily Electricity Load" Electronics 11, no. 10: 1524. https://doi.org/10.3390/electronics11101524

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop