Next Article in Journal
Communication, Computing, and Caching Trade-Off in VR Networks
Previous Article in Journal
Leading Role of E-Learning and Blockchain towards Privacy and Security Management: A Study of Electronics Manufacturing Firms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards Convergence in Federated Learning via Non-IID Analysis in a Distributed Solar Energy Grid

Department of Computer Education, Sunchon National University, Suncheon 57922, Republic of Korea
Electronics 2023, 12(7), 1580; https://doi.org/10.3390/electronics12071580
Submission received: 23 February 2023 / Revised: 13 March 2023 / Accepted: 15 March 2023 / Published: 27 March 2023

Abstract

:
Federated Learning (FL) is an effective framework for a distributed system that constructs a powerful global deep learning model, which diminishes the local bias and accommodates the successful aggregation of locally trained models with heterogeneous datasets. However, when local datasets have the non-IID attribute, the optimization metric tends to diverge or show unstable convergence in the trajectory space. This paper delves into building a global model for the distributed Smart Grid environment, with regionally cumulated three solar energy datasets from January 2017 to August 2021 in a decentralized power grid in South Korea via FL. This distributed energy network involves local properties and physical distance between the regions, which raises a fundamental question of “Will time-serially curated non-IID local features be effective in constructing a global regression model?”. This paper probes this question by leveraging FL and conducts the theoretically viable non-IID case-by-case convergence analysis, providing the interpretation of the embedded temporal non-IID features and application on real-world data. Moreover, most of the FL studies predetermine the global update period, which lacks applicability when adapting FL in actual practice. As FL is a cumulative-basis structure, the update term is a crucial factor that needs to be carefully selected. This paper articulates this problem and explores the effective update period via multiple experiments on the 4.5 years of solar energy dataset, and to the best of my knowledge, this is the first literature that presents the optimal update period in the FL regression in an energy domain.

1. Introduction

The distributed environment is a practical setting that independently curates local datasets, which reflect the intrinsic attributes and properties that may vary in each local environment. In such settings, Federated Learning (FL) [1] is known to be a suitable architecture that accommodates the successful aggregation of heterogeneously trained local models that mirrors its local environments. It reduces regional bias and enhances generalizability without transmitting all the local datasets into one primary server, which guarantees local data privacy. However, when periodically accumulated local datasets are not independent and identically distributed (non-IID), the global model in FL tends to diverge without obtaining a smooth convergence trajectory in the learning space. This condition typically occurs in the real-world system and has been suggested as a significant issue that needs to be solved for further implementation of FL to achieve successful utilization.
One of the distributed system domains that show an interest in utilizing FL is Smart Grid. Smart Grid is a decentralized-oriented power grid network architecture that adaptively accommodates the best utilization of electricity resources that consist of multiple agents (prosumers), synchronously producing and consuming energy by forecasting the future energy amounts. As a vast dataset volume is being generated in the Smart Grid environment in real-time (e.g., multimodal sensors), and the conventional process such as transmitting raw dataset, combining, storing, and preprocessing into an ideal format, then training the model inside a global server is a heavy process, and is not systematically suitable in such decentralized environment. Moreover, the energy dataset involves valuable information, such as correlation to national security or companies’ and users’ privacy; its security must be preserved. FL provides a breakthrough to those problems, offering solutions to efficiently operate the existing processes and to preserve data privacy. Smart Grid is a paradigm that operates in a parallel fashion; it optimizes the allocation of available resources (i.e., Distributed Energy Resources; DER) and energy inside the discrete geographical area (e.g., Micro-Grid), which is an ideal environment for applying FL, as shown in Figure 1.
However, to effectively adapt FL in a Smart Grid, it encounters several realistic hindrances. Establishing a high-performance deep learning forecasting model is crucial, which requires data engineering and extracting salient features that directly and deviously affect the target variable. The primary issues of FL in a dispersed energy grid to predict the DER can be summarized as follows.
Non-deterministic attribute—Obfuscates the future strategy: The FL with an objective function being classification has been widely studied concerning convergence analyses [2] and applications [3,4] since it assumes that the existing label types are predefined, which makes comparatively a straightforward setting. However, regression is a relatively arduous task compared to classification since the expected output vector is mostly continuous within the time domain, which must be accompanied by an assumption that the local datasets share similar distribution-wise statistical features to extract information. Such stochastic attribute increases the efforts of fine-tuning the single global model in FL to build a successful management system and hinders the stable convergence of the iteratively aggregated global model.
Data-level limitation—non-IID in real practice: In actual practice, most datasets frequently encompass the skewed, biased distribution with non-IID attributes. The target distribution range and observed time domain properties may vary in each local dataset. Additionally, data loss may incur triggered by equipment malfunctions, low bandwidth, etc., which will unbalance the dataset volume at the local level. These unexpected anomalies enhance the difficulty of maintaining the identical setting in the system, which will dilute the similarity of the input features.
Algorithmic-level limitation—Unstable convergence: FL is essentially an iteratively cumulated process that non-linearly trains the parameters and linearly aggregates the selected locals. Thus, a poorly trained local model and ineffective selection among the network will trigger the global model to suffer unstable convergence that generates a biased model, or it may incur divergence (i.e., client drift [5]) in the worst case.
This paper adopts the FL framework in a non-IID scenario that accommodates the effective convergence optimization in the non-IID local dataset while predicting the quantitative energy utilization (e.g., energy production, consumption to target) for constructing future response strategies. This work conducts a theoretical convergence analysis while inserting a biased dataset and discovering the preeminent periodical global updates. The presented empirical experiment results validate this claim via training a real-setting solar power energy generation dataset from January 2017 to August 2021 in three disparate locations in South Korea [6]. Furthermore, we explore the periodical training effect and update the central global model in the real-world environment, presenting the optimal update period that converges, returning the lowest loss value among the suggested candidates of the update scheme. To the best of my knowledge, this is the initial study that offers the effective update period selection of the regression-based task in FL, also investigating the time-series non-IID property analysis with practically viable cases. The summary of contributions of this research is as follows:
  • This work defines the viable case-by-case scenarios of the locally collected non-IID dataset. It suggests a quantification scheme of the non-IID degree in a regression-based FL network with both local and global perspectives. To the best of my knowledge, this is the first study that covers the domain of local non-IID attributes in the FL regression task.
  • This work suggests theoretical convergence analysis in FL regression optimization based on the quantitative degree of IID and non-IID attributes at the local training dataset, proposing a practical training approach to enhance the convergence rate of the global model.
  • This work validates the suggested analysis through multiple experiments in the FL model with the distributed solar energy generation dataset in specific regions of South Korea within the period of January 2017 to August 2021 to empirically vindicate the usability and performance of the suggested FL regression and update periods in future applications in the Smart Grid system.
The contents covered in this paper are organized as follows: Section 2 illustrates the milestones concerning FL in energy prediction in the Power Grid environment and previous footprints of non-IID-related FL studies. Section 3 covers the preliminaries of FL and explicates the non-IID and IID attributes in FL regression, which defines the feasible non-IID cases suited for actual practice. Section 4 illustrates the estimated dataset’s non-IID degree and performs convergence analysis of the FL regression with multiple experiments. Finally, Section 5 concludes this study by offering insights and core elements that need to be covered in future studies.

2. Related Works

FL is a diversely implemented architecture for optimizing the power grid system for pragmatic energy usage [7,8] via collaboration with geographically dispersed AI models. Diverse studies were suggested to design a cost-effective structure in such an energy-grid system, analyzing the challenges and issues from multiple perspectives, such as resource allocation, convergence analysis [3,4], and network cost [9]. Zhai et al. [10] applied the FL algorithm in dynamic network bandwidth settings in edge computing in the power grid. Liu et al. [11] brought the concept of horizontal and vertical FL and implemented those concepts into predicting power consumption. They merged the encryption algorithm while transferring information between the local and the server to ensure privacy. Since privacy concerns are a significant issue in the Smart Grid, studies were conducted to mitigate the vulnerability through FL. Wen et al. [12] proposed the energy theft anomaly detection model in the FL network and compared its detection performance with other models. Zhao et al. [13] designed the anomaly detection scheme in the solar farm data to prevent the injection attack. FL-based blockchain technologies were also studied in the smart environment [14]. Al-Quraan et al. [15] presented a cost-efficient measure concerning communication and computation by processing the collected data in a Smart Grid environment. Additionally, Wang et al. [16] analyzed the consumer’s characteristic dataset, leveraged privacy-enhanced PCA to diminish the features, and applied it to the ANN for target attribute identification. Zhao et al. [17] applied differential privacy into FL to enhance the security of local-level LSTM that predicts the energy consumption of individual households’ datasets. Similarly, privacy-preserving FL was recently studied in energy grid environments, most of them to achieve the secure aggregation of local information at the global level [18,19]. Petrangeli et al. [20] performed energy forecasting on the distributed residential household areas with an energy dataset measured from multimeters, and they empirically showed the trade-off between privacy and performance. Likewise, previous studies that dealt with FL in specific energy grid settings were mainly on the rudimentary steps of adapting the FL toward achieving classification-based tasks.
The primary objective of FL in a Smart Grid is to aggregate the local information and collaboratively produce the constructive measure by predicting the target unit, such as energy demand (i.e., base load and peak load) and energy production (i.e., renewable energy). To investigate related works, Taïk et al. [21] conducted individual short-term load forecasting by FL in distantly proximate houses in the state of Texas, USA, and applied a real-world dataset to build a robust prediction model in FL. Zhang et al. [22] introduced an efficient data-management FL system trained through deep reinforcement learning in IIoT settings. Moreover, FL was widely implemented in energy demand learning (EDL) domains, and Saputra et al. [23] adopted EDL in the distributed data of electric vehicles, predicting the energy demand in Dundee city, UK. Huang et al. [24] propounded FL in parking service estimation and resource allocation via LSTM and applied Stackelberg game theory and reinforcement learning to enhance the utility of the parking lot operations. FL scheme is also used in non-intrusive load monitoring (NILM) in the power grid and technology in a distributed environment to analyze the power consumption data for constructive usage [25,26]. These initial approaches proposed the novelty and proof that FL is indeed an effective structure to be applied in the power system. On the other hand, most of these studies do not involve practical usage, as the FedAvg tends to be vulnerable to classification problems while non-IID data are fed into the model.
When it comes to non-IID, although in a different field compared to this study, multiple literature were presented to minimize the effect of a non-IID using the clustering algorithm. Nightingale et al. [27] suggested energy forecasting schemes using multiple clusters to categorize the clients with similar attributes using K-means and scored lower RMSE and MAPE compared to conventional FL, which offers that sorted models with spatial information perform better. Similarly, Brigg et al. [28] leveraged the clustering method in FL (FL + HC; hierarchical clustering) and built an energy forecasting model with a non-IID dataset. This manuscript lies in a similar context in dealing with a non-IID dataset. Although studies [29,30] exist that explore the classification task in non-IID settings, existing works did not thoroughly investigate the intrinsic properties of the serially curated non-IID dataset, which fundamentally limits the utilization of further regression-oriented application domains. This work offers a theoretical analysis of possible case-by-case non-IID scenarios and applies the analysis in a real-life setting to validate the claim. Moreover, this work provides intuitive explanations by exploring interior features that it involves non-IID attributes to essentially comprehend the time-serial statistics.
FL is a cumulative process that iteratively updates the global model after local-level aggregation, and the optimal update period is considered critical to building a high-performance global model. This period agenda has recently been intensely studied by Yan et al. [31] with various criteria, including batch size, weight decay, learning rate annealing, etc. Most of the conventional FL studies pre-set the update period, which is a significant matter when designing an applicable system in the Smart Grid scenario. This work articulates this issue and utilizes 4.5 years-period dataset to demonstrate the training impact of a multiple global update period. Furthermore, this work offers the optimal updating term based on empirical analysis that realizes the best performance.

3. FL Regression: Trained with Non-IID and IID

3.1. Federated Learning

FL has become a successful framework as the locally curated dataset volume has vastly increased [32], and transmitting, preprocessing, managing, and training those datasets only in a centralized system undisputedly requires high cost. Moreover, more distributed systems are adopting a decentralized collaboration since locally acquired heterogeneous information is valuable for discovering local and global knowledge and reducing bias. Furthermore, FL preserves the privacy and security of the edge dataset. It iteratively builds a powerful global model by aggregating the local clients [33], which is generalizable, locally and time-wise agnostic, not overfitted to a particular distribution. In general, when the target incorporates robust spatiotemporal features, this is a well-suited task in FL, where it productively aggregates the embedded features. However, when the target domain must embrace diversity, FL may not be the best alternative, and preserving those heterogeneities may be appropriate. Nevertheless, one particular setting will plausibly trigger the overfitting; thus, contemplating the locals that share similar features is helpful. Therefore, examining the target environment is critical to evaluate the potential effect of FL. Our Smart Grid is considered befitting since clients share high-level features, and acquiring those subtly different distributions helps avoid overfitting and construct a global model. The basic formula of the global model in FL (FedAvg [1]) is indicated as follows Equations (1)–(5).
G ( T ) W ( T ) : = 1 n ( R ) i = R [ 0 ] n ( R ) W i ( T , t )
W i ( T , t ) : = { j w i , j ( T , t ) , j b i , j ( T , t ) }
R : = { l | 1 l   n ( i i ) } s . t . R   i i    
W i ( T , t + 1 ) = W i ( T , t ) η L ( h i ( T , t ) , y i ) W i ( T , t ) s . t .      lim t 𝕥 ( min i L ( f ( D i ( T ) , W i ( T , t ) ) , y i ( T ) ) ) : W i ( T , 1 ) W i ( T , 𝕥 )
f ( T ) ( u j ( t ) x j ( t ) + w j ( t ) h j ( t ) + b j ( t ) ) h ( T , t ) d
Note that the variables and parameters used in this study are listed in Notations. The individual model  M i ( )  are identical regression model types, and  n ( k ( i , k ) ) = c  since (1) is a pairwise and elementwise computation inside a matrix form, where  n ( W i ) = c ^  and when  ( n ( W i ) n ( W i ) )   o r   ( n ( k ( i , k ) ) n ( k ( i , k ) ) ) lim T T ( lim t 𝕥 ( min i L ( h i ( T , t ) , y ) ) )  [2,5].

3.2. Non-IID Cases in Regression-Based FL

This section defines the non-IID cases in the regression-based FL environment, which globally modifies the global model with two perspectives. First, this section describes the FL cases with the data-structural view in Section 3.2.1 and construes the FL-structural standpoint in Section 3.2.2.

3.2.1. Non-IID Cases with a Structural View

The feasible cases are as follows:
  • Case A.  n ( D ( i , t r a i n ) ( T , t ) ) n ( D ( i , t r a i n ) ( T , t ) )   s . t .   i i   a n d   D i = ( j x j , j y j ) .
  • Case B.  ψ ( D ( j y ( i , j ) ) , D ( j y ( i , j ) ) ) 0   s . t .   i i   a n d   n ( j y ( i , j ) ) = n ( j y ( i , j ) )  in (5).
  • Case C.  1 j 1 j ( δ ( y ( i , j ) , y ( i , j ) ) ) 0 .
Case A indicates the different amount of data in each local while identical epoch (i.e., global/local). Assorted environmental factors such as low bandwidth and faulty sensors may incur data loss in identical sampling rate and data acquisition frequency. Case B illustrates the different distribution of  j y j  in  i  and  i , representing the condition of  y i . This value can be utilized to measure the quantitative quality of a particular local dataset to serve as the standard for selecting  M l R . It implements the Kullback–Leibler divergence in (6) to compute the density difference of  y ( i , i ) , and determine as non-IID when it satisfies Case B. Depending on the environmental characteristic,  D ( j y ( i , j ) )  may have unique distribution, such as  y i ~ N ( μ ( y i ) , σ 2 ( y i ) ) y i ~ G a m m a ( α , β ) , etc., and when  σ ( i D ( j y ( i , j ) ) ) 0 , the most frequent  D ( j y ( i , j ) )  becomes the ground truth. The left term in Case C shows the time-dependent distribution difference  δ ( y i , y i )  in (7). On the other hand, Case B does not consider the time-series attribute, which may incur the risk of unstable convergence while training the local models. Typically, time-series datasets embed the temporal-domain features that must be utilized to obtain (4). Thus, this is designated as another metric for the non-IID dataset.
ψ ( D ( j y ( i , j ) ) , D ( j y ( i , j ) ) ) : = j D ( j y ( i , j ) ) log D ( j y ( i , j ) ) D ( j y ( i , j ) )
δ ( y i , y i ) : = y i y i   i n   e l e m e n t w i s e   and   p a i r w i s e   f a s h i o n
s . t .     y i = j y ( i , j ) , y i = j y ( i , j )

3.2.2. Non-IID Cases in FL-Structural View

The training in FL has three primary levels: individual local client-level training with  t  epoch, aggregation of every local client in a single global  T  communication round, and training with  T T  global round. The categorization of the cases via three levels is as follows:  ( i , T ) ( i , T ) ( i , T ) .
  • Case 1.  ( M i ( T ) , M i ( T ) ) c a s e   A   o r   B   o r   C .
  • Case 2.  ( M i ( T ) , M i ( T ) ) c a s e   A   o r   B   o r   C .
  • Case 3.  ( M i ( T ) , M i ( T ) ) c a s e   A   o r   B   o r   C .
  • Case 4.  ( M i ( T ) , M i ( T ) ) c a s e   A   o r   B   o r   C .
  • Case 5.  ( M i ( T ) , M ) c a s e   A   o r   B   o r   C .
  • Case 6.  ( M i ( T ) , M ) c a s e   A   o r   B   o r   C .
The inputs were formulated by  ( 2 ~ 4 ) : M i ( T = 1 , t = 1 ) M i ( T = T , t = 𝕥 ) W i ( T , t )  in Equations (1)–(3), where  ( j x ( i , j ) , y i ) D i . Case 1 and 2, respectively, exhibit the non-IID and IID cases, where both are adjusted at the local client level. In the IID case, we know that  n ( D i T ) = n ( D i T ) ψ ( D i ( T ) , D i ( T ) ) 1 j 1 j ( δ ( y i , y i ) ) 0 , which  M i ( T ) M i ( T ) . When the two inputs satisfy any cases of A, B, and C, this designates it as non-IID. Likewise, Cases 3 and 4 indicate the non-IID and IID case in the single global communication level, which incorporates all existing local clients (i.e., or only the selected few, depending on the FL aggregation algorithm). Case 4 shows that  M i ( T ) M i ( T ) . In Cases 5 and 6,  M lim T T ( lim t 𝕥 ( min i μ ( L ( M i ( T , t ) ,   M i ( T , t ) ) ) ) )  is the robust model trained with an ideal local dataset that guarantees convergence. Moreover,  M i ( T , t )   M i ( T , t )  in Case 6, which is the supreme model presenting the highest performance.

3.3. IID Dataset in FL Regression

A locally extracted IID dataset assumes a particular distribution, and such an attribute is beneficial. When the target system shares the analogous features that hold the IID attribute with specific distribution, this neglects the limitations that hinder the convergence, such as the geographical distance among the system becoming meaningless. With a unique convergence that trains the local model independently,  M i W i  conducts (4), (5) and satisfies  lim t 𝕥 ( 4 ) : min t = 1 L ( h i ( t ) , y i ( t ) ) min t = 𝕥 L ( h i ( t ) , y i ( t ) )  in parallel in  i . Then the server performs (1), where  1 i 1 i M i : = G ( T ) . This case holds the following assumption.
Assumption 1.
D ( x ( i , f ) , y i ) D ( x ( i , f ) , y i )  and  1 j 1 j ( δ ( x ( i , f ) , x ( i , f ) ) ) 0 ,  which does not satisfy cases A, B, and C.
This assumption will preserve the properties of the independently trained local models, which share similar features at a high-level while (3) and (4) achieve  min i L ( h i ( t ) , y i ( t ) ) , thus  ( Δ W i ( t ) Δ W i ( t ) ) d . Furthermore, additive measures such as regularization schemes (e.g., batch normalization) will enhance the similarity in the convergence trajectory such that  E [ ( W ^ i ( t ) μ ( t W ^ i ( t ) ) ) ( W ^ i ( t ) μ ( t W ^ i ( t ) ) ) ] 1 σ ( t W ^ i ( t ) ) σ ( t W ^ i ( t ) ) > 0 , where  W ^ i ( t ) = Δ W i ( t ) μ ( t Δ W i ( t ) ) σ ( t Δ W i ( t ) ) . Let  D i = f x f , such that  x f = { x ( f , t 1 ) , x ( f , t 2 ) , , x ( f , t n ) } , which indicates a time-series dataset aligned with  t n . In  D i ( T , t ) D i ( T , t ) D i ( T , t )  where  ( t t ) 𝕥 , when (8), where  ϕ ( 𝓉 , q )  refers to a multivariate function, and  q  is a set of variables, then (9) must be met if  n ( D i ( T , t ) )  and  Δ 𝓉 s ( T ) = | 𝓉 s ( T ) 𝓉 s ( T + 1 ) |  are not deterministic. In FL,  M i  gradually updates (3) in a synchronous fashion; thus, (10) is computed where  τ  indicates the sampling rate per second and  ϵ  is the error term. Based on (10), this sets the optimal  T  in (11) that ensures high-performance training, and Section 4 empirically validates this claim.
T D i ( T , t ) = ϕ ( 𝓉 , q ) β sin ( α t ) d d 𝓉 ϕ ( 𝓉 , q )
min n ( D i ( T , t ) ) > 2 α π   s . t .   α
| | Δ 𝓉 ( T ) | | τ = n ( D i ( T , t ) ) + ϵ
( T = T + 1 ) | | Δ 𝓉 ( T ) τ | | n ( D i ( T , t ) ) > 2 α π

4. Property and Performance Analysis of FL Regression

As the renewable energy source is an unstabilized power source due to inconstant generative output patterns that vary throughout time, this section implements FL experiments to predict the solar energy generation amount in solar power plants [6] in three regions of South Korea. By conducting multiple experiments on such non-IID sundry environments, this literature focuses on analyzing the primary properties of regression-based FL in each subsection.

4.1. Dataset Analysis

The acquired dataset [6] was imported from the public portal provided by the Korean government, which shows the solar power generation amount in the solar power plants located in the three provinces (Sejong-si, Youngnam-gun, Ansan-si;  L 1 , L 2 , L 3 ). The cumulation period was from 1 January 2017 to 31 August 2021 with hourly measurements. Such a time-domain dataset has two main repeated periodical data time-unit frames: day (24 h) and year. Monthly, weekly, and seasonally do not show a cyclic pattern, which was not considered. The generation amount repeatedly displays a similar trend in those two units, such that  n ( D i ( T ) ) = 24  or  8760 , and  δ ( D i ( T ) , D i ( T )   s . t .   n ( D i ( T ) ) = 24   or   8760 ) δ ( D i ( T ) , D i ( T )   s . t .   n ( D i ( T ) ) 24   or   8760 ) .
The following figures show the ideal sample and the defective sample of daily energy collection. Figure 2a displays a smooth curve starting around 7:00 a.m., reaches its peak at 13:00 p.m., and halts the generation around 20:00 p.m. Let this ideal form be (12). In (12),  max   S ( 𝓉 h ) d d 𝓉 S ( 𝓉 h ) = 0 , and  12 argmax   S ( 𝓉 h ) 14 . Therefore,  d d 𝓉 S ( argmax   S ( 𝓉 h ) ± c ) < d d 𝓉 S ( 𝓉 h argmax   S ( 𝓉 h ) ± c ) 2 ( 1 + 2 c 24 ) n ( 𝓉 h ( 𝓉 h ( S ( 𝓉 h ) 0 ) ) ) 24 , where  c  indicates the proximal hours surrounding  argmax   S ( 𝓉 h ) , and this explains Figure 2c,d. In contrast, Figure 2a,b consists of unstable peaks without intuitive periodical variation, which has the form of (13) and the most distinctive large density at  S ( 𝓉 h ) 0 .
S ( 𝓉 h ) = { i f   6 < h < 20 ,     S ( 𝓉 h ) e 1 2 ( 𝓉 h μ ( 𝓉 h ) σ ( 𝓉 h ) ) σ ( 𝓉 h ) 2 π + ϵ o t h e r w i s e ,     S ( 𝓉 h ) 0
S ( 𝓉 h ) = { i f   6 < h < 20 ,     S ( 𝓉 h ) e 1 2 ( 𝓉 h μ ( 𝓉 h ) σ ( 𝓉 h ) ) σ ( 𝓉 h ) 2 π + ϵ o t h e r w i s e ,     S ( 𝓉 h ) 0
The non-IID property resembles risk while collaborating to aggregate the individual local models. It was noted that the predefined non-IID case C to quantify its internal non-IID degree, where  1 j 1 j ( δ ( y i , y i ) ) 0 . Case C is computed through  y e a r s 1 24 i = 1 t = 24 ( | L ^ i L ^ i | ) ( 7 ) : = δ ^ ( L i , L i ) , and  L ^ ( i , j ) = j L ^ ( i , j ) ,   s . t .   L ^ ( i , j ) = L ( i , j ) max   j L ( i , j ) max   j L ( i , j ) min   j L ( i , j ) . Table 1 displays the  μ ( δ ^ ( L i , L i ) ) ± σ ( δ ^ ( L i , L i ) )  in each year and the overall values combined for all the periods. Based on this, assuming that the dataset shares similar features in the synchronous time domain, it is certain that the degree of the non-IID attribute is:  δ ^ ( L 2 , L 3 ) > δ ^ ( L 1 , L 3 ) > δ ^ ( L 1 , L 2 ) . Additionally, although  L i  shares a synchronous direct source, the geographical distance between the entities should be pondered. The relationship of the Euclidean distance is  ( L 2 , L 3 ) 330 > ( L 1 , L 3 ) 230 > ( L 1 , L 2 ) 130 ,  where the unit is kilometers, which is identical to the degree of non-IID. This can be roughly interpreted as the proximal distance is likely to share analogous features, and utilizing those factors can be beneficial; this claim is validated in Section 4.3.

4.2. Training RNN-Based Models in Local Client

Based on the collected dataset ( D i L i ), the experiment initially implements the three RNN-based models: LSTM, GRU, and BiLSTM, to train the solar energy generation amount. The training dataset is nominated with the other two  D i  ( x f = D i ) in the initial experiment setting, and the training label is target  D i  (e.g.,  i = 1 ,   2 ,   i = 3 ). The second experimental setting inserts all the  D i x f = D i  in the training dataset to train the single model (e.g.,  i = 1 ,   2 ,   3 ,   i = 1   o r   2   o r   3 ). The third experiment independently trains three RNN-based models ( M i ) with their unique  D i . Every experiment had an identical model structure with 1000 epochs, RNN-based model (64)-Dense (64)-Dense (32)-Dense (1), 32 batch size, Adam optimizer, ReLU activation function, mean-absolute-error loss function, 20% validation split from the training dataset, training dataset were January 2017–December 2020, and test dataset were January 2021–August 2021. The results of experiments 1 and 2 are shown in Figure 3, and experiment 3 is shown in Figure 4.
Compared to other experiments, the MAE of experiment 1 reveals to be lower than expected, with an average value of 58,202.96 (Local 1), 62,592.67 (Local 2), and 635,606.56 (Local 3) in LSTM. Local 3 shows the worst performance, approximately ten times larger than the other two local models. However, the average data value for all locals is 250,251.28, which the unit itself is considered large. This implies that a training dataset that incorporates past data is the compelling feature that must be fused. Assuming from the general knowledge that a more quality dataset volume enhances the performance, experiment 2 spends more resources than experiment 1 with additional features (i.e., an identical unit of target labels with a different period). The training was successful, lowering the average of MAE to 9.66 (Local 1), 71.93 (Local 2), and 75.47 (Local 3), which is a far better achievement than experiment 1. However, experiment 3 surpasses this with 1.28 (Local 1), 12.20 (Local 2), and 27.50 (Local 3). This implies that FL performance is unlikely to be superior to the locally trained model when forecasting its individual target value. The following subsection explicates this in detail. The overall results are shown in Table 2.

4.3. Merging Heterogeneous Local Models in FL

The MAE of the individual local models in Section 4.2 results shows that  M A E ( L 1 ) < M A E ( L 2 ) < M A E ( L 3 )  in experiments 1–3. This relation corresponds to  δ ^ ( L 2 , L 3 ) > δ ^ ( L 1 , L 3 ) > δ ^ ( L 1 , L 2 )  non-IID degree:  L 3 > L 2 > L 1 . Furthermore, Figure 5 shows the heatmap that displays the Pearson correlation of the yearly period dataset.  P C C ( L ( i , y e a r ) , L ( i , y e a r ) ) > 0 0.71 P C C ( L ( i , y e a r ) , L ( i , y e a r ) ) 0.79 , and  0.87 P C C ( L ( i , y e a r ) , L ( i , y e a r ) ) 1 , where  P C C ( x , y )  denotes the Pearson Correlation Coefficient of input  x , y ,  and  y e a r y e a r i i . This implies that the unmatched period encompasses a relatively low statistical feature than the equivalent year dataset, which is likely to be a non-IID dataset in terms of the period (i.e., dates rotate with the years).
Based on the locally trained models with a monthly dataset with 70 epochs, it merges them to create  G  and iteratively propagates the  lim T 48 G ( T ) L i ( T )  (i.e., 12 months, 4 years), where the overall training epoch is  70 12 4 = 3360 . The result is shown in Figure 5, where the loss values are trained from  G ( t r a i n = L i ) ( T 1 , t = 70 ) f ( D i ( T ) ) min L ( f ( D i ( T ) , W i ( T , t ) ) , y i ) W i ( t + 1 ) = Δ W i ( t ) + W i ( t ) , and  L ( f ( D i ) , y i ) ; training and validation datasets are shown in Figure 6a,b. At every global epoch, a surge in loss is observed. Although  lim T T = 48 ( lim t 𝕥 = 20 L ( f ( D i ( T ) , W i ( T , t ) ) , y i ) ) 0 , which satisfies the objective function,  1 i 1 i W i ( T ) = G ( T ) L ( f ( D i ( T ) , W i ( T , t = 𝕥 ) ) , y i ) < L ( f ( D i ( T ) , W i ( T + 1 , t = 1 ) ) , y i ) . Figure 6c–f shows the loss of initial and final local epoch  L ( f ( D i ( T ) , W i ( T , t = 1 ) ) ,   L ( f ( D i ( T ) , W i ( T , t = 70 ) ) , and results show that  L ( f ( D i ( T ) , W i ( T > 2 , t ) )  tend to converge. Note that the LSTM layer was used in local models.
This denotes that after the merge in global communication, further training must be accompanied for optimal convergence, and the following experiment proves this. After 1000 local epochs in the LSTM model, the experiment merged the locals, and the outcomes were unsatisfactory while predicting the three test datasets, generating the MAE average shown in Table 3. Furthermore, experiments were conducted on the different combinations of locals while aggregation, such as combining only local models (1, 2), (2, 3), and (1, 3). The results are shown in Table 3, along with the performance mentioned above.
Lemma 1.
Loss in FL regression (i.e.,  L ( y , f ( x ) ) : = 1 n i ( y i f ( x i ) ) 2
) converges such that  lim T T ( lim t 𝕥 L ( f ( D i ( T ) , W i ( T , t ) ) , y i ) ) 0  in the  d  loss space where  ( T , 𝕥 , d )
, while  1 i 1 i W i ( T ) = G ( T ) L ( f ( D i , W i ( T , t = 𝕥 ) ) , y i ) < L ( f ( D i , W i ( T + 1 , t = 1 ) ) , y i )
Theorem 1.
Based on Lemma, additional local training is required s.t.  ( T , 𝕥 ) 2 ,  after  1 i 1 i W i ( T ) = G ( T ) lim T T ( lim t 𝕥 ( min i L ( f ( D i ( T ) , W i ( T , t ) ) , y i ) ) )  in FL regression.
Proof of Theorem 1.
Empirical experiments—Experiments results in Table 2 and Figure 5 show that the theorem is true. Note that if  ( T , 𝕥 ) = 1 Δ W i ( t ) = , and  ( T , 𝕥 ) > 1 Δ W i ( t ) , and  Δ W i ( t ) > 0 , it satisfies (3). □

4.4. Federated Learning on Periodical Update

This section focuses on exploring the impact of global periodical collaboration, which is considered a vital factor while achieving convergence in FL. While total epochs are identical, the experiment considers three following periodical aggregation cases:
  • Case  A . Daily update period, such that  𝕥 = 24 , and  T = c 𝕥  in (3).
  • Case  . Monthly period, such that  𝕥 = 24 D a y s , and  T = c 𝕥 D a y s  in (3).
  • Case  C . Yearly period, such that  𝕥 = 24 D a y s M o n t h s , and  T = c 𝕥 D a y s M o n t h s  in (3).
All the cases have identical total epochs ( = 𝕥 T = 3384 ) and equivalent layer structures using LSTM and hyperparameters to former experiments in this Section 4. Case  A  results are shown in Figure 7a,b, with unstable convergence and relatively high loss values. Case   was already conducted and delivered in Figure 6a,b, and case  C  results are displayed in Figure 7c,d. In general, case  C  had the best performance, with a much smaller loss (i.e., when observing the scale of the suggested Figure 7a–d and stable trajectory compared to other cases). This implies that more local database volume will incorporate better convergence when training FL.

5. Conclusions

FL is a collaborative framework in a distributed system that accommodates a synergy effect among locally trained models with independently collected datasets from its unique environment. Aggregating the independently trained heterogeneous models construct a powerful global model by absorbing various features that enhance generalization and reduce bias. FL regression is widely impacted by the spatiotemporal features, where FL operates in a decentralized system that subsumes the geographical distance between the local units, and such curated datasets tend to be non-IID, which must take account of the timely global update with period while training the FL. The learning trajectory of FL is most likely to diverge when the adjustments are asynchronous or the assemblage does not mitigate the heterogeneous properties. This implies that combining diverse and unique local models trained with non-IID datasets in FL must be effectively aggregated in a way that directs the global model to return a high prediction rate.
This paper explores regression models in the FL, defining the frequent non-IID cases in real-life regression problems and discovering the temporal non-IID combination method via theoretical analysis and empirical experiments. This work utilized the solar energy generation dataset acquired from the three regions in South Korea and trained three types of RNN-based models, and showed that the improvement of conglomerating local regression models in the FL network correlates to the non-IID degree of the local dataset. This paper quantitively assessed the local non-IID degree by examining the dataset and observing the FL performance, which validated that the non-IID attribute hinders convergence during iterative aggregation in regression. This paper also suggested and proved that the global update period is a decisive factor that determines the performance of the FL through multiple experiments with a 4.5-year-long dataset, which is a novel approach and pragmatic matter that must be optimized when applied to a real-practice Smart Grid environment.
This study offers a collaboration scheme to build an efficient FL-based regression model that can be utilized in the decentralized power grid system, which encompasses the geographical features that are primarily nonadjacent and independently operate to each neighbor’s power source. Through theoretical analysis and experiments suggested in this study, the author hopes this FL research provides the practical approach and perspective to a successful operation in a diversified disseminated energy grid system such as Smart Grid and Micro Grid via FL.

Funding

This research was funded by the National Research Foundation of KOREA (NRF) grant funded by the Korean government (MSIT) (No. 2020R1A2C1012363).

Data Availability Statement

The data described in this article are openly available in the public data portal provided by Korean government at https://www.data.go.kr/data/15003553/fileData.do, accessed on 14 March 2023.

Conflicts of Interest

The authors declare no conflict of interest.

Notations

NotationDefinitionNotationDefinition
  G Global model   t Local epoch
  W Set of global parameters   T Global epoch
  W Tensor shape of parameters   𝕥 Number of local epoch
  w , b 2d vector of weights and bias   T Number of global epoch
  i , i Local client index ( i i )   L ( ) Loss function of input
  j , j i ’s parameter index ( j j )   u , h Set of hidden units
  c , c ^ Constant   , k Layer, Layer index
  D Dataset   ϵ Error term
  R Required number of i   d Loss space with  d  dimensions
  D ( ) Distribution of input   η Learning rate
  ψ ( a , b ) Kullback Leibler divergence of  a , b     M Ideal benchmark dataset
  M Local model   𝓉 s   o r   h Time unit s: sec, h: hour
  σ ( ) Standard deviation of the input   f Set of feature;  f f  
  μ ( ) Average of the input   x x Input
  τ sampling rate per second   y y Ground truth
  ϕ ( ) multivariate function   L 1 Sejong-si
  q multivariate function   L 2 Youngnam-gun
  L 3 Yansan-si

References

  1. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Artificial Intelligence and Statistics; PMLR: Cambridge, MA, USA, 2017; pp. 1273–1282. [Google Scholar]
  2. Li, X.; Huang, K.; Yang, W.; Wang, S.; Zhang, Z. On the convergence of fedavg on non-iid data. arXiv 2019, arXiv:1907.02189. [Google Scholar]
  3. Gadekallu, T.R. Federated Learning for Big Data: A Survey on Opportunities, Applications, and Future Directions. arXiv 2021, arXiv:2110.04160. [Google Scholar]
  4. Zhan, Y.; Zhang, J.; Hong, Z.; Wu, L.; Li, P.; Guo, S. A Survey of Incentive Mechanism Design for Federated Learning. IEEE Trans. Emerg. Top. Comput. 2021, 10, 1035–1044. [Google Scholar] [CrossRef]
  5. Karimireddy, S.P.; Kale, S.; Mohri, M.; Reddi, S.; Stich, S.; Suresh, A.T. Scaffold: Stochastic controlled averaging for federated learning. In International Conference on Machine Learning; PMLR: Cambridge, MA, USA, 2020; pp. 5132–5143. [Google Scholar]
  6. Korea Public Data Portal. Available online: https://www.data.go.kr/data/15003553/fileData.do (accessed on 21 December 2022).
  7. You, S. A Cyber-secure Framework for Power Grids Based on Federated Learning. Eng. Arch. 2020. [Google Scholar] [CrossRef]
  8. Massaoudi, M.; Abu-Rub, H.; Refaat, S.S.; Chihi, I.; Oueslati, F.S. Deep Learning in Smart Grid Technology: A Review of Recent Advancements and Future Prospects. IEEE Access 2021, 9, 54558–54578. [Google Scholar] [CrossRef]
  9. Taik, A.; Nour, B.; Cherkaoui, S. Empowering Prosumer Communities in Smart Grid with Wireless Communications and Federated Edge Learning. IEEE Wirel. Commun. 2021, 28, 26–33. [Google Scholar] [CrossRef]
  10. Zhai, S.; Jin, X.; Wei, L.; Luo, H.; Cao, M. Dynamic Federated Learning for Gmec With Time-Varying Wireless Link. IEEE Access 2021, 9, 10400–10412. [Google Scholar] [CrossRef]
  11. Liu, H.; Zhang, X.; Shen, X.; Sun, H. A federated learning framework for smart grids: Securing power traces in collaborative learning. arXiv 2021, arXiv:2103.11870. [Google Scholar]
  12. Wen, M.; Xie, R.; Lu, K.; Wang, L.; Zhang, K. FedDetect: A Novel Privacy-Preserving Federated Learning Framework for Energy Theft Detection in Smart Grid. IEEE Internet Things J. 2021, 9, 6069–6080. [Google Scholar] [CrossRef]
  13. Zhao, L.; Li, J.; Li, Q.; Li, F. A Federated Learning Framework for Detecting False Data Injection Attacks in Solar Farms. IEEE Trans. Power Electron. 2021, 37, 2496–2501. [Google Scholar] [CrossRef]
  14. Li, D.; Luo, Z.; Cao, B. Blockchain-based federated learning methodologies in smart environments. Clust. Comput. 2021, 25, 2585–2599. [Google Scholar] [CrossRef]
  15. Al-Quraan, M.; Khan, A.; Centeno, A.; Zoha, A.; Imran, M.A.; Mohjazi, L. FedTrees: A Novel Computation-Communication Efficient Federated Learning Framework Investigated in Smart Grids. arXiv 2022, arXiv:2210.00060. [Google Scholar]
  16. Wang, Y.; Bennani, I.L.; Liu, X.; Sun, M.; Zhou, Y. Electricity consumer characteristics identification: A federated learning approach. IEEE Trans. Smart Grid 2021, 12, 3637–3647. [Google Scholar] [CrossRef]
  17. Zhao, Y.; Xiao, W.; Shuai, L.; Luo, J.; Yao, S.; Zhang, M. A Differential Privacy-enhanced Federated Learning Method for Short-Term Household Load Forecasting in Smart Grid. In Proceedings of the 2021 7th International Conference on Computer and Communications (ICCC), Chengdu, China, 10–13 December 2021; pp. 1399–1404. [Google Scholar] [CrossRef]
  18. Fernandez, J.D.; Menci, S.P.; Lee, C.; Fridgen, G. Secure Federated Learning for Residential Short Term Load Forecasting. arXiv 2021, arXiv:2111.09248. [Google Scholar]
  19. Zhou, X.; Feng, J.; Wang, J.; Pan, J. Privacy-preserving household load forecasting based on non-intrusive load monitoring: A federated deep learning approach. PeerJ Comput. Sci. 2022, 8, e1049. [Google Scholar] [CrossRef] [PubMed]
  20. Savi, M.; Fabrizio, O. Short-term energy consumption forecasting at the edge: A federated learning approach. IEEE Access 2021, 9, 95949–95969. [Google Scholar] [CrossRef]
  21. Taik, A.; Cherkaoui, S. Electrical Load Forecasting Using Edge Computing and Federated Learning. In Proceedings of the ICC 2020–2020 IEEE International Conference on Communications (ICC), Dublin, Ireland, 7–11 June 2020; pp. 1–6. [Google Scholar] [CrossRef]
  22. Zhang, P.; Wang, C.; Jiang, C.; Han, Z. Deep Reinforcement Learning Assisted Federated Learning Algorithm for Data Management of IIoT. IEEE Trans. Ind. Inform. 2021, 17, 8475–8484. [Google Scholar] [CrossRef]
  23. Saputra, Y.M.; Hoang, D.T.; Nguyen, D.N.; Dutkiewicz, E.; Mueck, M.D.; Srikanteswara, S. Energy Demand Prediction with Federated Learning for Electric Vehicle Networks. In Proceedings of the 2019 IEEE Global Communications Conference (GLOBECOM), Waikoloa, HI, USA, 9–13 December 2019; pp. 1–6. [Google Scholar]
  24. Huang, X.; Li, P.; Yu, R.; Wu, Y.; Xie, K.; Xie, S. Fedparking: A federated learning based parking space estimation with parked vehicle assisted edge com-puting. IEEE Trans. Veh. Technol. 2021, 70, 9355–9368. [Google Scholar] [CrossRef]
  25. Zhang, Y.; Tang, G.; Huang, Q.; Wang, Y.; Wu, K.; Yu, K.; Shao, X. FedNILM: Applying Federated Learning to nilm Applications at the Edge. In IEEE Transactions on Green Communications and Networking; IEEE: Piscataway Township, NJ, USA, 2022; p. 1. [Google Scholar] [CrossRef]
  26. Hudson, N.; Hossain, J.; Hosseinzadeh, M.; Khamfroush, H.; Rahnamay-Naeini, M.; Ghani, N. A Framework for Edge Intelligent Smart Distribution Grids via Federated Learning. In Proceedings of the 2021 International Conference on Computer Communications and Networks (ICCCN), Athens, Greece, 19–22 July 2021; pp. 1–9. [Google Scholar] [CrossRef]
  27. Nightingale, J.S.; Wang, Y.; Zobiri, F.; Mustafa, M.A. Effect of Clustering in Federated Learning on Non-IID Electricity Consumption Prediction. In Proceedings of the 2022 IEEE PES Innovative Smart Grid Technologies Conference Europe (ISGT-Europe), Novi Sad, Serbia, 10–12 October 2022; pp. 1–5. [Google Scholar] [CrossRef]
  28. Briggs, C.; Fan, Z.; Andras, P. Federated learning for short-term residential energy demand forecasting. arXiv 2021, arXiv:2105.13325. [Google Scholar]
  29. Yan, G.; Hao, W.; Jian, L. Seizing Critical Learning Periods in Federated Learning. In Proceedings of the AAAI Conference on Artificial Intelligence; AAAI Press: Palo Alto, CA, USA, 2022. [Google Scholar]
  30. Lu, Y.; Huang, X.; Zhang, K.; Maharjan, S.; Zhang, Y. Blockchain Empowered Asynchronous Federated Learning for Secure Data Sharing in Internet of Vehicles. IEEE Trans. Veh. Technol. 2020, 69, 4298–4311. [Google Scholar] [CrossRef]
  31. Lim, W.Y.B.; Xiong, Z.; Miao, C.; Niyato, D.; Yang, Q.; Leung, C.; Poor, H.V. Hierarchical Incentive Mechanism Design for Federated Machine Learning in Mobile Networks. IEEE Internet Things J. 2020, 7, 9575–9588. [Google Scholar] [CrossRef]
  32. Lee, H.; Liu, Y.; Kim, D.; Li, Y. Robust Convergence in Federated Learning through Label-wise Clustering. arXiv 2021, arXiv:2112.14244. [Google Scholar]
  33. Zhou, T.; Konukoglu, E. FedFA: Federated Feature Augmentation. arXiv 2023, arXiv:2301.12995. [Google Scholar]
Figure 1. Federated Learning architecture visualization in solar energy generation environment.
Figure 1. Federated Learning architecture visualization in solar energy generation environment.
Electronics 12 01580 g001
Figure 2. Visualization sample (a,b) of daily solar energy generation and its distribution (c,d).
Figure 2. Visualization sample (a,b) of daily solar energy generation and its distribution (c,d).
Electronics 12 01580 g002
Figure 3. RNN-based models’ training MAE result for experiments 1 and 2: rows 1–3 and column 1 display the MAE-LSTM, BiLSTM, and GRU with training dataset  ( D 1 , D 2 )  and target dataset  D 3 . Figures in rows 1–3 and columns 2–3 show MAE-LSTM, BiLSTM, and GRU with training dataset  ( D 1 , D 3 ) ( D 2 , D 3 ) , and target dataset  D 2 . D 1 , respectively. Figures in rows 4–6 columns 1–3 was trained with  ( D 1 , D 2 , D 3 ) , and their target set was  D 1 ,   D 2 , D 3 , respectively.
Figure 3. RNN-based models’ training MAE result for experiments 1 and 2: rows 1–3 and column 1 display the MAE-LSTM, BiLSTM, and GRU with training dataset  ( D 1 , D 2 )  and target dataset  D 3 . Figures in rows 1–3 and columns 2–3 show MAE-LSTM, BiLSTM, and GRU with training dataset  ( D 1 , D 3 ) ( D 2 , D 3 ) , and target dataset  D 2 . D 1 , respectively. Figures in rows 4–6 columns 1–3 was trained with  ( D 1 , D 2 , D 3 ) , and their target set was  D 1 ,   D 2 , D 3 , respectively.
Electronics 12 01580 g003aElectronics 12 01580 g003b
Figure 4. LSTM training performance of experiment 3. Figures in row 1 display LSTM results with a scale of 0–1200, and row 2 has a narrow scale of 0–100. Figures in row 3 show the BiLSTM result, and row 4 shows the GRU result. The Figures in sequential columns in each row indicate local 1,2, and 3.
Figure 4. LSTM training performance of experiment 3. Figures in row 1 display LSTM results with a scale of 0–1200, and row 2 has a narrow scale of 0–100. Figures in row 3 show the BiLSTM result, and row 4 shows the GRU result. The Figures in sequential columns in each row indicate local 1,2, and 3.
Electronics 12 01580 g004
Figure 5. Heatmap of the  P C C ( L ( i , y e a r ) , L ( i , y e a r ) ) .
Figure 5. Heatmap of the  P C C ( L ( i , y e a r ) , L ( i , y e a r ) ) .
Electronics 12 01580 g005
Figure 6. FL training results in training (a): validation loss (b), training loss at local epochs 1 and 70 (c,d), and validation loss at local epochs 1 and 70 (e,f).
Figure 6. FL training results in training (a): validation loss (b), training loss at local epochs 1 and 70 (c,d), and validation loss at local epochs 1 and 70 (e,f).
Electronics 12 01580 g006
Figure 7. MAE values for FL: (a,b) the training and validation losses on daily updates; (c,d) the training and validation loss for yearly updates.
Figure 7. MAE values for FL: (a,b) the training and validation losses on daily updates; (c,d) the training and validation loss for yearly updates.
Electronics 12 01580 g007
Table 1. δ ^ ( L i , L i )  values.
Table 1. δ ^ ( L i , L i )  values.
δ ^ ( L 1 , L 2 )
δ ^ ( L 1 , L 3 )
δ ^ ( L 2 , L 3 )
20171.60 ± 0.851.55 ± 1.031.86 ± 0.90
20181.66 ± 0.951.60 ± 0.991.86 ± 0.91
20191.57 ± 0.881.60 ± 0.991.83 ± 0.91
20201.71 ± 0.751.61 ± 1.001.97 ± 0.86
August 20211.87 ± 0.871.64 ± 0.971.98 ± 0.86
Overall1.67 ± 0.871.60 ± 1.001.89 ± 0.89
Table 2. Final training result for experiments 1–3.
Table 2. Final training result for experiments 1–3.
Experiment 1
Local 1Local 2Local 3
TrainValTrainValTrainVal
LSTM60,298.2658,202.9664,383.7962,592.67662,806.25635,606.56
GRU60,169.0458,801.2764,059.0564,715.78640,914.86641,708.94
BiLSTM59,735.6958,777.6664,374.0865,708.52646,646.01651,371.92
Experiment 2
Local 1Local 2Local 3
TrainValTrainValTrainVal
LSTM39.419.6693.1671.93374.1075.47
GRU47.7336.6730.8627.75443.16310.01
BiLSTM41.7458.5635.7812.7356.5759.75
Experiment 3
Local 1Local 2Local 3
TrainValTrainValTrainVal
LSTM21.561.2827.4112.20159.3727.50
GRU29.8619.7632.156.80204.2230.82
BiLSTM22.431.8114.138.16119.5816.97
Table 3. Average MAE result for local combinations.
Table 3. Average MAE result for local combinations.
Local 1Local 2Local 3
FedAvg ( L 1 , L 2 , L 3 )472,308.44516,164.471,334,568.25
FedAvg ( L 1 , L 2 )183,788.52220,938.301,605,387.38
FedAvg ( L 1 , L 3 )11,498.8230,333.071,790,837.25
FedAvg ( L 2 , L 3 )228,898.06201,307.982,012,914.50
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, H. Towards Convergence in Federated Learning via Non-IID Analysis in a Distributed Solar Energy Grid. Electronics 2023, 12, 1580. https://doi.org/10.3390/electronics12071580

AMA Style

Lee H. Towards Convergence in Federated Learning via Non-IID Analysis in a Distributed Solar Energy Grid. Electronics. 2023; 12(7):1580. https://doi.org/10.3390/electronics12071580

Chicago/Turabian Style

Lee, Hyeongok. 2023. "Towards Convergence in Federated Learning via Non-IID Analysis in a Distributed Solar Energy Grid" Electronics 12, no. 7: 1580. https://doi.org/10.3390/electronics12071580

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop