Next Article in Journal
Feature of Review Papers in Nanotechnology and Applied Nanosciences
Previous Article in Journal
Segmenting Urban Scene Imagery in Real Time Using an Efficient UNet-like Transformer
Previous Article in Special Issue
Short-Term Forecasting of Ozone Concentration in Metropolitan Lima Using Hybrid Combinations of Time Series Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FedDeep: A Federated Deep Learning Network for Edge Assisted Multi-Urban PM2.5 Forecasting

1
College of Computer and Software, Hohai University, Nanjing 210024, China
2
NARI Technology Co., Ltd., Nanjing 210047, China
3
Shenzhen Urban Transport Planning Center Co., Ltd., Shenzhen 518000, China
*
Authors to whom correspondence should be addressed.
The authors contributed equally to this work.
Appl. Sci. 2024, 14(5), 1979; https://doi.org/10.3390/app14051979
Submission received: 22 December 2023 / Revised: 6 February 2024 / Accepted: 21 February 2024 / Published: 28 February 2024
(This article belongs to the Special Issue Air Quality Prediction Based on Machine Learning Algorithms II)

Abstract

:
Accurate urban P M 2.5 forecasting serves a crucial function in air pollution warning and human health monitoring. Recently, deep learning techniques have been widely employed for urban P M 2.5 forecasting. Unfortunately, two problems exist: (1) Most techniques are focused on training and prediction on a central cloud. As the number of monitoring sites grows and the data explodes, handling a large amount of data on the central cloud can cause tremendous computational pressures and increase the risk of data leakages. (2) Existing methods lack an adaptive layer to capture the varying impacts of different external factors (e.g., weather conditions, temperature, and wind speed). In this paper, a federated deep learning network (FedDeep) is developed for edge-assisted multi-urban P M 2.5 forecasting. First, we assign each urban region to an edge cloud server (ECS). An external spatio-temporal network (ESTNet) is then deployed on each ECS. Data from different urban regions are uploaded to the corresponding ECS for training, which avoids processing all the data on the central cloud and effectively alleviates computational pressure and data leakage issues. Second, in ESTNet, we develop a gating fusion layer to adaptively fuse external factors to improve prediction accuracy. Finally, we adopted P M 2.5 data collected from air quality monitoring sites in 13 prefecture-level cities, Jiangsu Province for validation. The experimental results proved that FedDeep outperformed the advanced baselines in terms of prediction accuracy and model efficiency.

1. Introduction

After entering the 21st century, the issue of P M 2.5 has become increasingly severe and garnered significant attention [1]. Especially in urban regions, P M 2.5 significantly affects human health, economic development, and social progress [2]. The prediction of P M 2.5 concentrations can assist in better tackling urban air quality issues. Following the production of accurate urban P M 2.5 estimations, the relevant authorities can implement more powerful strategies, such as regulating industrial production and issuing travel warnings [3,4].
Currently, many urban P M 2.5 forecasting methods are being developed, and some of them, relying on deep learning technologies, can accomplish high-precision prediction. Unfortunately, most researchers train their models on a central cloud [5,6]. For instance, Du et al. [6] deployed a deep-learning-based method DAQFF on the central cloud for air pollution prediction using two real-world datasets. Experimental results proved that DAQFF achieved satisfactory prediction accuracy. As the number of monitoring sites increases and the volume of collected data sharply rises, massive amounts of data and computational tasks are uploaded to the server, which leads to two issues: (1) This can put tremendous computational pressure on the central cloud, manifested by extended training times and high GPU memory usage. Hence, it is difficult to achieve timely prediction. For instance, in some large-scale air pollution predictions, i.e., with a large number of sensors, the central cloud typically takes approximately 30 min to complete an epoch of training. (2) We store all data on this server, thus creating a greater security risk [7,8,9]. The main reason for this is that the central cloud typically consolidates a large volume of data. In the event of a data leakage, this could lead to incalculable losses.
External factors, e.g., wind speed, temperature, precipitation, humidity, etc., have a significant impact on changes in urban P M 2.5 . For instance, temperature influences the atmosphere and ventilation states. The deposition of particulate matter is affected by precipitation and humidity. High wind speeds accelerate the diffusion of particulate matter [10,11,12]. However, the existing models neglect these external features or adopt equal weights for various external features when integrating them into prediction approaches [13,14,15,16]. For instance, Liu et al. [15] fused meteorological factors with air pollution data through direct summation in air pollution prediction. As a result, these methods fail to adaptively capture different impacts of diverse external features. It is challenging to develop an adaptive mechanism to extract the different impacts originating from external factors. One main reason exists: it is nontrivial to quantify the effects of external factors on urban P M 2.5 and assign weights when integrating them with forecasting approaches.
In summary, two urgent issues need solutions in urban P M 2.5 forecasting: (1) Dealing with massive amounts of data on a central cloud generates huge computational pressures and security risks. (2) The different impacts caused by various external features on urban P M 2.5 cannot be adaptively extracted. To tackle these issues, we present a novel federated deep learning network entitled FedDeep for multi-urban P M 2.5 forecasting. FedDeep is inspired by horizontal federated learning. In particular, we first deploy an edge computing server (ECS) [17] separately in each urban region. The data collected by the sensors within the urban region are sent to the corresponding ECS. Leveraging the ability of edge computing and data caching, the computing pressure on the central cloud can be effectively alleviated through deploying an ECS adjacent to the source of data generation. Meanwhile, data leakage on the central cloud can be circumvented. Second, the external spatio-temporal network (ESTNet) is stored in each ECS for training and uploading gradients to the ESTNet that is deployed on the central cloud. For each ESTNet, we design a gating fusion layer to fuse urban P M 2.5 features and external factors according to the diverse effects of various external factors. This is achieved by categorizing the external factors according to their types, then using one-hot encoding and fully connected networks (FCs) to embed them into the same space for fusion, and finally leveraging non-linear functions for weighted integration with urban P M 2.5 features. In this way, our designed model can discern variations in the impacts generated by different external factors, which is of great significance for identifying mutations in urban P M 2.5 . The main contributions of this work are as follows:
  • Since multiple urban monitoring sites produce massive amounts of P M 2.5 data, high computational pressure and data leakage levels occur when leveraging only a central cloud for all data and computing task. In light of this, we present an edge-based federated learning architecture. Specifically, we deploy a separate ECS in each urban region. The data acquired by the sensors in each urban region are uploaded to the corresponding ECS instead of the central cloud for training, which effectively mitigates the computational pressures on the central cloud and prevents security risks. Subsequently, the gradients trained on the ECS are uploaded to the central cloud for parameter updates, and then the model deployed on the central cloud predicts the P M 2.5 in all urban regions, which creates a “distributed training, centralized prediction” mode.
  • We design a novel ESTNet model, in which a gating fusion layer is proposed to adaptively fuse external factors and urban P M 2.5 features based on the different impacts of external features.
  • The experimental results on 13 prefecture-level cities in Jiangsu Province confirmed that FedDeep outperformed other state-of-the-art baselines. Specifically, FedDeep achieved the highest accuracy in predicting P M 2.5 for the subsequent 12 h. Meanwhile, FedDeep had a short training time and low GPU usage, thus showing its high efficiency.
The remainder of this paper is organized as follows: Section 3 introduces our studied region, the data used, and the architecture of our proposed model. The experiment description and result analysis are given in Section 4. Finally, Section 5 concludes this work.

2. Related Work

2.1. Air Pollution Prediction

Currently, air pollution prediction methods can be generally classified into three categories: contained deterministic, statistical, and deep learning methods. Deterministic methods adopt mathematical expressions, meteorological theories, and the physical–chemical reactions of air pollution to simulate the transformation and diffusion processes of pollutants [18]. Common methods include the weather research and forecasting model [19]. However, deterministic methods need to utilize a lot of prior knowledge, which limits their prediction performance. To address this limitation, statistical methods have been proposed, which employ principles of statistics to discovery the relationships between external factors (e.g., meteorological data) and air pollution. Statistical methods can be divided into two categories: classical statistical methods and shallow machine learning methods. Specifically, classical statistical methods include the historical average (HA) [20]. For instance, HA [20] was developed to forecast air quality in Delhi. Unfortunately, these methods are based on linear assumptions, which impacts their prediction precision in a nonlinear problem like air pollution forecasting. In light of these factors, shallow machine learning methods (e.g., support vector regression (SVR) [21]) were proposed. Nonetheless, these methods fail to consider spatio-temporal correlations in air pollution data.
With the advent of deep learning technology [22], the potential of this technology for extracting spatio-temporal features is gradually being explored. Many researchers have adopted deep-learning-based methods for air pollution prediction [23,24,25]. For instance, Sayeed et al. proposed a convolutional neural network and long short-term memory (CNN-LSTM) [24] to predict ozone concentrations over 24 h. Specifically, a convolutional neural network (CNN) was proposed to extract spatial dependencies. Long short-term memory (LSTM) was used to capture temporal dependencies. A CNN can only model the road network in a standard form, e.g., 2D-matrix. Despite the benefits of convolutional operations in mining spatial dependencies between neighboring sites, there are some crucial spatial dependencies, such as the relationships between distant sensors, that are challenging to express in the standard form. In view of this, graph neural networks (e.g., graph attention network (GAT)) have attracted a lot of attention because of their ability to represent a sensor network using graph theory. This can completely preserve the original spatial relationships in the sensor network. For example, Han et al. [25] leveraged a graph attention network and long short-term memory (GAT-LSTM) to predict P M 2.5 over 24 h in Beijing. In particular, spatial relationships between nodes in the graph were extracted by GAT, which introduced a spatial attention inspired by the transformer [26] model. LSTM was adopted to capture temporal dependencies. Although these models achieve accurate predictions, they are trained and predicted on a central cloud. This can put computational pressure on the server. Meanwhile, the security risk increases. Moreover, these methods fail to fuse external factors in an effective manner.

2.2. Federated Learning

Federated learning (FL) [27] is a widely used collaborative learning architecture. FL avoids uploading training data to the central cloud, allowing ECSs to locally train a shared global model using their own data. It has been applied in various fields [28,29]. For instance, Gholizadeh and Musilek [28] predicted individual and aggregate electrical loads based on FL. In air pollution prediction, FL is feasible because multiple monitoring sensors enable knowledge sharing, leading to improved prediction accuracy.

3. Materials and Methods

In this section, we consider multi-urban P M 2.5 forecasting based on FL [27]. A novel algorithm entitled FedDeep was designed, which is shown in Figure 1. In the following, we overview the process of multi-urban P M 2.5 prediction with FedDeep.

3.1. Overview

FedDeep is customized based on historical P M 2.5 concentrations and external factors (e.g., weather conditions, temperature, and wind speed). In particular, the data collected by the air quality monitoring and meteorological sites are transmitted to the ECS through wireless links.
A system schematic of FedDeep is shown in Figure 1. An ECS deployed in each urban region is responsible for data processing and computing offloading. Each ECS offers computing resources for ESTNets to be trained using P M 2.5 and external factors from corresponding urban regions. For instance, Figure 1 indicates that the entire Jiangsu Province can be regarded as 13 urban regions; that is, divided according to the boundaries of prefecture-level cities. Note that the study area and data in this section are consistent with that of the subsequent experimental simulations. The ECSs are defined as S = S 1 , S 2 , , S M , where M = 13 denotes the total number of ECSs. D = D 1 , D 2 , , D M is assumed as the set of local dataset. D i serves as the local dataset i originated from urban region i, which is transmitted to the i t h ECS. Let ESTNet = ESTNet 1, ESTNet 2, …, ESTNet M represent the set of local E S T N e t . Specifically, the local E S T N e t i is deployed on ECS i for training using the local dataset D i , and then the trained gradients are uploaded to the central cloud to update the parameters of the global E S T N e t . In this manner, the computational pressure of the central cloud can be decreased through deploying E S T N e t in a distributed way. In addition, the central cloud does not have to store all the data, which decreases the risk of data leakage.
In summary, realization of multi-urban P M 2.5 forecasting can be organized into four steps: In the first step, a ECS receives P M 2.5 and external factors collected by sensors through wireless links. In the following step, the ECS processes these data for training the local ESTNet. Prior to model training, the model parameters in the ECS need to be encrypted and shared with models in other ECSs. In view of this, an ECS is required to encrypt these parameters before sending, and then the other ECSs need to decrypt them after receiving. In the last step, ECSs train corresponding ESTNets through local datasets, and then upload the trained gradient to the central cloud to update the parameters of the global ESTNet.

3.2. Local Estnet

Whether in the central cloud or edge, we deploy an ESTNet with the same structure. As shown in Figure 2, ESTNet consists of four components: input embedding layer, gating fusion layer, spatio-temporal learning layer, and prediction layer. Specifically, ESTNet was designed as a hierarchical architecture. In the following, we detail how the data flows through each component from the the perspective of the local ESTNet deployed on the ECS i.
(1) Input Embedding Layer: For the local dataset D i , considering the collaborative training of urban P M 2.5 and external factors, it is necessary to standardize the dimensions of these two data types. Assuming historical P timesteps of urban P M 2.5 observations X = X t 1 , X t 2 , , X t P R P × n i × F , we adopt two-layer FCs to convert X from F into d m o d e l dimensions, which can be denoted as H X = H t 1 , H t 2 , , H t P R P × n i × d model , where n i represents the number of air quality monitoring sites in an urban region i.
External factors, e.g., meteorological conditions, have a significant impact on urban P M 2.5 . There are discrepancies in the impact of different external factors [10,11,12]. In this work, we chose weather conditions, temperature, and wind speed as external factors. To highlight different degrees of impact, we classified weather conditions into five different types: sunny, overcast, rainy, snowy, and hazy, and then employ one-hot encoding to convert each type into a vector. According to experience, we categorized the temperature into 10 levels, with a temperature difference of 5 °C for each level. Based on the official classification, wind speed was projected into 13 levels. Temperature and wind speed were normalized proportionally in the range [0, 1]. Finally, we concatenated all external factors of the same timestep into a one-dimensional tensor E R Z , where Z stands for the dimension of the concatenated tensor. The external factors of the different sites and timesteps were concatenated to obtain the external factor representations of urban region i, which is formalized as T E R P × n i × Z . To facilitate follow-up processing, two-layer FCs are utilized to transform the T E to d m o d e l -dimension, which can be represented as H T E R P × n i × d model .
(2) Gating Fusion Layer: A common strategy, i.e., to extract the impact of external factors on urban P M 2.5 , is to directly concatenate external factor data with urban P M 2.5 data [30,31]. Unfortunately, it is unreasonable to view the effects of external factors through concatenating with urban P M 2.5 features with the same weight. In light of this, a novel gating fusion layer was developed to adaptively integrate external factors with urban P M 2.5 features, depending on the significance of each part. The detailed architecture of the gating fusion layer is presented in Figure 2.
With embeddings, urban P M 2.5 features and external factors are denoted as H X R P × n i × d model and H T E R P × n i × d model , respectively. For the purpose of generating the weight z, we adopt a s i g m o i d function σ to process H X and H T E :
z = σ H X · W H X z + H T E · W H T E z + b z ,
where W H X z R d model × d model , W H T E z R d model × d model , and b z stand for the trainable parameters.
Furthermore, the fused representations H R P × n i × d model are acquired according to the following equation:
H = ( H X ( 1 z ) ) + ( z H T E ) ,
where ⊙ stands for the element-wise product.
(3) Spatio-Temporal Learning Layer: After receiving the fused representations, the spatio-temporal learning layer is deployed to capture high-level features, which contains multi-head self attention (MHSA) and LSTM. In particular, this layer is a pipeline structure. MHSA is first adopted to mine spatial dependencies, and then temporal dependencies are extracted through LSTM. A detailed description of MHSA and LSTM is given as follows:
MHSA: MHSA was first proposed by Vaswani et al. [26] for natural language processing and now plays a vital role in spatio-temporal mining tasks. This is a technique that dynamically adjusts the weights of input data, to concentrate on more important information. In urban P M 2.5 prediction, there are dynamic and complex spatial dependencies among monitoring sites. As a result, a L layer stacked MHSA is used to capture spatial dependencies by dynamically allocating weights to various sites, as shown in Figure 3. Specifically, in the urban region i, let H R P × n i × d model denote the input of L-layer MHSA. In the l t h layer, its input is the output of the l 1 t h layer l−1 H S R P × n i × d model , in which the representation of site s i at timestep t j is l−1 h s s i , t j R d model . For sensor s i at timestep t j , the attention coefficient α s i , s v m at the single m t h head is calculated as follows:
α s i , s v m = exp ρ s i , s v m r n i exp ρ s i , s r m ,
where n i stands for all sites in the urban region i. ρ s i , s v m represent the correlations among site s i and s v , and are acquired through performing the inner product of the key vector of s i and the query vector of s v :
ρ s i , s v m = f q m h s s i , t j l 1 , f k m h s s i , t j l 1 d ,
here, * , * stands for the inner product. f q m and f k m are nonlinear transformations of the query and key vector, which can be expressed generically as
f ( x ) = ReLU ( W x + b ) ,
in this setting, we denote the activation function as R e L U , and W and b are the learnable weights and bias, respectively. After obtaining ρ s i , s v m , the dynamic spatial representations output by the l t h layer are formally given as
h s s i , t j m l = r n i α s i , s r m f v m h s s r , t j l 1 ,
h s s i , t j l = BN m = 1 M h s s i , t j m l W s + h s s i , t j l 1 ,
where f v m is the formula (5) for obtaining the value vector in the m t h head. | | and BN separately represent the concentration and batch normalization. M is the total number of heads. W s denotes the learnable weights. By iterative L layer operations utilizing Equations (3)–(7), the final output of spatial representations in the site s i at timestep t j is presented as h s s i , t j L R d model . We extend the above calculation to all n i sites at historical P timesteps to acquire the spatial representations of the urban region i, H S L R P × n i × d model .
LSTM: Since urban P M 2.5 typically constitutes a time series, these data can be modeled under consideration of the dynamic temporal dependencies among different timesteps. In view of this, we employ long short-term memory (LSTM) to capture temporal dependencies [32] behind MHSA. Figure 2 indicates the specific structure of the LSTM. The LSTM is composed of multiple LSTM units, in which each unit has the same structure. A unit corresponds to processing data at a timestep, and there are connections among units. In detail, at the timestep t j , the LSTM unit contains one memory cell and three gates, e.g., a forget gate F t j , input gate I t j , and output gate O t j . The contents of memory cell state C t j are controlled by F t j and I t j . O t j can determine the amount of C t j to be projected into the current output H S T t j R n i × d model of the LSTM unit. The computation process of the LSTM unit at timestep t j is as follows:
F t j = σ g W F L H S t j + U F H S T t j 1 + b F , I t j = σ g W I L H S t j + U I H S T t j 1 + b I , O t j = σ g W O L H S t j + U O H S T t j 1 + b O , C t j = F t j C t j 1 + I t j tanh W C L H S t j + U C H S T t j 1 + b C , H S T t j = O t j tanh C t j ,
where W F , W I , W O , and W C denote weight matrices for the input of dynamic spatial representations L H S t j R n i × d model at timestep t j . U F , U I , U O , and U C represent the weight matrices allocated to the state of the memory cell H S T t j 1 from the previous LSTM unit. b F , b I , b O , and b C are bias vectors. To provide non-linearity, we leverage sigmoid functions as σ g . ⊙ is the element-wise product. Each LSTM unit outputs a feature matrix. Thus, the spatio-temporal representations H S T R n i × d model are acquired through summing them together, which formalizes as
H S T = H S T t 1 + H S T t 2 + + H S T t P ,
(4) Prediction Layer After th spatio-temporal learning layer, the prediction layer is utilized to generate output sequences. Specifically, in [33], a sequence-to-sequence (seq2seq) architecture was developed. It was first applied to machine translation and later to the field of air pollution prediction. Seq2seq consists of an encoder and a decoder. The encoder is leveraged to encode the input sequence to a semantic vector, and then the decoder decodes this vector to the target sequence. Each unit structure in the decoder remains consistent, and the semantic vector serves as the input for each prediction timestep of the decoder. Therefore, we utilize the decoder of Seq2seq to generate the prediction sequence, where each unit employs an LSTM unit. Taking the future timestep t P + k as an example, the output of the decoder at the t P + k timestep is given by
F t P + k = σ g W F H S T + U F H Y t P + k 1 + b F , I t P + k = σ g W I H S T + U I H Y t P + k 1 + b I , O t P + k = σ g W O H S T + U O H Y t P + k 1 + b O , C t P + k = F t P + k C t P + k 1 + I t P + k tanh W C H S T + U C H Y t P + k 1 + b C , H Y t P + k = O t P + k tanh C t P + k ,
where W F , W I , W O , and W C are weight matrices for the input of the spatio-temporal representations H S T . U F , U I , U O , and U C denote the weight matrices allocated to the state of memory cell H Y t P + k 1 from the previous LSTM unit. b F , b I , b O , and b C stand for bias. Subsequently, FCs are utilized to transform H Y t P + k R n i × d model into the desired output Y t P + k R n i × F . We implement the same operation on the decoder of all predicted timesteps and gain all the predicted sequences Y R Q × n i × F .
Finally, the parameters of the local ESTNet deployed on ECS i are optimized through a minimizing mean squared error (MSE) loss function, represented as
MSE = s n = 1 n i t q = 1 Q y s n , t q y ^ s n , t q 2 n i × Q + λ 2 W 2 ,
here, n i is the total number of sites in the urban region i, and Q denotes the length of the prediction sequence. The predicted and observed values at the timestep t q on the site s n are y t q , s n and y ^ t q , s n , respectively. λ represents the regularization. W is the trainable parameter.

3.3. Parameter Update for Global ESTNet

While training the local ESTNet, we also need to perform parameter updates for the global ESTNet. Specifically, after each local ESTNet runs an epoch, its trained gradient is uploaded to the global ESTNet for updating. When each local ESTNet has run all epochs, the global ESTNet completes the gradient update. Then, the global ESTNet is used to predict the P M 2.5 of future Q timesteps in all urban regions.

4. Results and Discussion

In this section, we performed several experiments on datasets from 13 prefecture-level cities of Jiangsu province to validate the predictive performance of our proposed model.

4.1. Experimental Settings

4.1.1. Study Area and Data

We evaluated FedDeep on 13 prefecture-level cities in Jiangsu Province.
Jiangsu province, positioned on the Yangtze River in East China, is one of the most developed regions in terms of economy. The province boasts a substantial population and highly developed industry. It covers a region of about 107,200 square kilometers and is divided into 13 prefecture-level cities. In 2019, the total population crossed 80.5 million [34,35]. In the past few years, the urban P M 2.5 problem in Jiangsu Province has become very serious, attributed to rapid economic development. Thus, to assist government departments in solving air pollution issues, we selected research regions in 13 prefecture-level cities of Jiangsu province, containing Nanjing, Suzhou, Wuxi, Changzhou, Zhenjiang, Nantong, Taizhou, Yangzhou, Yancheng, Huaian, Suqian, Xuzhou, and Lian yungang. A total of 133 air quality monitoring sites were referenced, of which the number of sites per urban is depicted in Table 1. Each site composed an hourly averaged concentration of P M 2.5 , ranging from 31 May 2018 to 31 December 2021. The external factors utilized in this paper were derived from district-level meteorological sites. We first measured the distance among P M 2.5 monitoring sites and each meteorological station. Weather data from the nearest meteorological station were then leveraged as external factors for the P M 2.5 monitoring site. The external factors of each site contained weather conditions, temperature, and wind speed, for which the data collection period was aligned with P M 2.5 . In the global ESTNet, we treated the 133 sites as predictive objectives. In each local ESTNet, the predictive objective was the number of sites contained in the corresponding prefecture-level city.

4.1.2. Baselines

FedDeep for multi-urban P M 2.5 forecasting was contrasted with other state-of-the-art baselines, as follows:
  • HA [20]: The average historical P M 2.5 sequence was adopted to forecast future multi-urban P M 2.5 concentrations.
  • SVR [21]: A support vector with RBF Kernel was designed for multi-step P M 2.5 forecasting.
  • LSTM [23]: This is a variant recurrent neural network (RNN) used to mine dynamic temporal dependencies in P M 2.5 data and produce predictive sequences.
  • CNN-LSTM [24]: This model integrates CNN and LSTM to capture the spatio-temporal correlations in P M 2.5 data.
  • GAT-LSTM [25]: This model mines spatial dependencies from a non-Euclidean spatial perspective and then applies LSTM to model temporal dependencies.

4.1.3. Metrics

We formulated five metrics to compare the performance of FedDeep and the baselines. These can be divided into two categories: (1) mean absolute error (MAE), root mean squared error (RMSE), and coefficient of determination ( R 2 ) were used to evaluate the prediction accuracy. (2) We tested the model efficiency through the training time and GPU memory. Details of the five metrics are described below:
(a) Mean absolute error (MAE):
M A E = 1 n i × Q s n = 1 n i t q = 1 Q y s n , t q y ^ s n , t q .
(b) Root mean squared error (RMSE):
R M S E = 1 n i × Q s n = 1 n i t q = 1 Q y s n , t q y ^ s n , t q 2 .
(c) Coefficient of determination ( R 2 ):
R 2 = 1 s n = 1 n i t q = 1 Q y s n , t q y ^ s n , t q 2 s n = 1 n i t q = 1 Q y s n , t q y ¯ 2 .
where y ¯ denotes the average of the observed values.
(d) Training time: Training time (TT) refers to the total time overhead throughout the training process till the model implemented a satisfactory accuracy. We could measure the efficiency of training using TT. For example, if prediction accuracies are at the same level, the models with shorter training time are more efficient.
(e) GPU Memory: This was used as the spatial overhead of the model, denoted as GPUM; that is, the consumption of GPU memory during training. For instance, a model with smaller usage of GPU memory has fewer learnable parameters and only demands a minimal amount of computational resources, which reflects a higher efficiency.

4.1.4. Parameter Settings

In this experiment, the implementation details of FedDeep were as shown in Table 2. The values of P and Q were set to 12, i.e., the P M 2.5 concentrations for the following 12 h were predicted according to the P M 2.5 concentrations and external factors for the previous 12 h. The pytorch library with 1 A100 and 13 Nvidia 3090 Ti GPU cards were applied in the experiment. In particular, 1 A100 card was treated as the central cloud, and 13 ECSs were simulated with 13 Nvidia 3090 Ti GPU cards. The size of the dataset in each ECS was consistent with the number of sites in Table 1. Each dataset used in the corresponding ECS was segmented into training, validation, and test sets based on the chronological order, with a split ratio of 7:1:2. In addition, the Z-Score normalization method was utilized as data normalization. For each local ESTNet deployed on the corresponding ECS, we selected the stochastic gradient descent as an optimizer to minimize the loss function MSE of 64 epochs. The batch size and learning rate were set to 32 and 0.0001. In addition, there were several hyperparameters in each local ESTNet: the number of heads in MHSA N h , the dimensions of each head N d , the MHSA layers L M H S A , the LSTM layers in the spatio-temporal learning layer L L S T M ( S T ) , and the LSTM layers in the decoder of seq2seq L L S T M ( d e c ) . These hyperparameters were fine-tuned on the validation set and the best prediction accuracy was observed with the following settings: N h = 4 , N d = 8 , L M H S A = 3 , L L S T M ( S T ) = 1 , and L L S T M ( d e c ) = 1 . After receiving the optimal N h and N d , we could determine d m o d e l = 32 . The parameter configuration of the global ESTNet was the same as that of the local ESTNet.

4.2. Experimental Results

4.2.1. Performance Comparisons

We compared FedDeep with five state-of-the-art baselines, including HA [20], SVR [21], LSTM [23], CNN-LSTM [24], and GAT-LSTM [25] for P M 2.5 forecasting of the next 12 h for the 133 sites distributed in the 13 prefecture-level cities of Jiangsu province. The comparison results are displayed in Table 3. TT in this table stands for the total training time. GPUM indicates the utilization of GPU memory during the training process. `-’ indicates that the model does not adopt GPU for running. FedDeep (w/o) indicates that only one central cloud was used to train and infer urban P M 2.5 .
(1) Prediction Accuracy Analysis. We measured the forecasting accuracy of the models with three evaluation metrics (e.g., MAE, RMSE, and R 2 ), and found that FedDeep outperformed the other baslines, thus demonstrating the effectiveness of our proposed model. In detail, the non-deep learning models HA and SVR had the lowest prediction accuracy. This was mainly because the simple structures of these models are unsuitable for meeting the requirements of processing multi-site P M 2.5 data from multiple urban regions. Compared with HA and SVR, LSTM presented a higher prediction accuracy. Nonetheless, LSTM only attended to temporal dependencies in P M 2.5 , and its accuracy was lower than the models that simultaneously captured spatio-temporal correlations (CNN-LSTM, GAT-LSTM, and FedDeep). Among the three spatio-temporal networks, CNN-LSTM was inferior to the others. One potential reason exists: the distributions of monitoring sites were irregular and treated as a graph in each urban region. When CNN is adopted to capture spatial dependencies, this requires converting the site data into a standard form, such as a 2D-matrix, which brings about a loss of crucial spatial information. In contrast, GAT-LSTM and FedDeep directly processed the graph structure, thus preserving more spatial information. Compared to GAT-LSTM, FedDeep provided a higher prediction accuracy. This result demonstrates that our proposed model could effectively mine spatio-temporal correlations in urban multi-site P M 2.5 data by taking advantage of its sensible structural design.
(2) Model Efficiency Analysis. Relying on TT and GPUM, we observed that FedDeep achieved an acceptable performance with respect to model efficiency. To improve prediction accuracy of the model as much as possible, the baselines spent more time on training and utilized more GPU memory in the multi-urban P M 2.5 forecasting, especially for the deep learning-based baselines. For instance, for total training time, the TT of FedDeep was significantly shorter than LSTM, CNN-LSTM, and GAT-LSTM. This revealed that time overheads can be reduced through a federated training approach. In addition, due to the small-scale structural design, the GPU memory usage of FedDeep was less than the other deep-learning-based baselines. This was due to the small size of the processed individual urban regions on the ECSs, so FedDeep had a relatively low GPU memory overhead, even with a large number of urban regions. Moreover, although FedDeep and FedDeep(w/o) were comparable in their prediction accuracy, FedDeep’s training time and GPU memory usage were significantly lower than FedDeep(w/o)’s. Thus, this proved the effectiveness of our proposed model in terms of efficiency.

4.2.2. Case Study

We conducted a case study to intuitively visualize the fitting results of our proposed model. We selected the Pukou site in Nanjing and Nanmen site in Suzhou, and plotted the fitting results for 500 continuous hours utilizing SVR, GAT-LSTM, and FedDeep, as depicted in Figure 4. Some conclusions can be drawn: (1) SVR could only identify linear dependencies, making it challenging to handle the complex non-linearities in P M 2.5 data. It led to the poorest fitting performance. (2) In contrast to SVR, GAT-LSTM could capture spatio-temporal correlations, enhancing its fitting capacity. Unfortunately, GAT-LSTM failed to learn the abrupt change in P M 2.5 at these two sites. (3) Compared with GAT-LSTM, FedDeep accomplished the best fitting performance, whether in the abrupt change or the smooth phase of the P M 2.5 sequence. The main reason for this was that our proposed method integrated external factors, assisting in learning the patterns in abrupt changes and smooth phases in the P M 2.5 sequence. This also reflects the fact that external factors are one of the triggers that lead to mutations in an P M 2.5 sequence.

4.2.3. Effect of Hyperparameters

The prediction results of FedDeep under different hyperparameter configurations for the future 12 h forecasting on 133 sites are presented in Figure 5. When one hyperparameter was revised, the other hyperparameters maintained their optimal settings ( N h = 4 , N d = 8 , L M H S A = 3 , L L S T M ( S T ) = 1 , and L L S T M ( d e c ) = 1 ). From Figure 5a–c, we can observe that when the model structure was more complex, it was more likely to underfit, while when the model was more simple, it was easier to overfit. Figure 5d,e exhibit stacking one layer of LSTM in the spatio-temporal learning layer, and the decoder of seq2seq accomplished the best accuracy. This was due to the fact that stacking too many LSTM layers causes error accumulation.

5. Conclusions

With the increase in the number of air quality monitoring sites and the explosion in the volume of data, centralized training of urban P M 2.5 prediction methods has led to several issues, such as computational pressures and security risks. To tackle these challenges, a FedDeep model was developed for multi-urban P M 2.5 forecasting in this paper. FedDeep uses a federated learning architecture. We first allocated an ECS corresponding to each urban region. ESTNets were then deployed on each ECS and the central cloud. We uploaded the data generated by each urban region to the corresponding ECS for training, instead of the central cloud, and then uploaded the trained gradients to the central cloud for parameter updating. Subsequently, all urban P M 2.5 predictions were carried out through the ESTNet on the central cloud, thus reaching a “distributed training, centralized prediction” mode. In this way, the computational load on the central cloud was alleviated, and the risk level of data leakage was reduced. Second, in ESTNet, a gated fusion layer was proposed to allow the model to consider the various impacts produced by different external factors (e.g., weather conditions, temperature, and wind speed). Third, we conducted an experiment on 13 prefecture-level cities in Jiangsu Province. The experimental results indicated that our proposed model outperformed the other state-of-the-art baselines, in both prediction accuracy and efficiency.

Author Contributions

The authors contributed equally to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant No. 62271190, and Research on Distribution Room Condition Sensing Early Warning and Distribution Cable Operation and Inspection Smart Decision Making Technology No. 524609220092.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding authors.

Conflicts of Interest

Authors Yue Hu and Wangyong Guo were employed by the company NARI Technology Co., Ltd., Author Meng Chen was employed by the company Shenzhen Urban Transport Planning Center Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNN-LSTMConvolutional Neural Network and Long Short-Term Memory
ECSEdge Computing Server
ESTNetExternal Spatio-Temporal Network
FedDeepFederated Deep Learning Network
GAT-LSTMGraph Attention Network and Long Short-Term Memory
GPUMGPU memory
HAHistorical Average
MAEMean Absolute Error
MHSAMulti-Head Self Attention
RMSERoot Mean Squared Error
R 2 Coefficient of determination
SVRSupport Vector Regression
TTTraining Time

References

  1. Zhang, B.; Rong, Y.; Yong, R.; Qin, D.; Li, M.; Zou, G.; Pan, J. Deep learning for air pollutant concentration prediction: A review. Atmos. Environ. 2022, 119347. [Google Scholar] [CrossRef]
  2. Chen, Y.; Yu, T.; Yang, B.; Zhang, X.S.; Qu, K. Many-objective optimal power dispatch strategy incorporating temporal and spatial distribution control of multiple air pollutants. IEEE Trans. Ind. Inform. 2019, 15, 5309–5319. [Google Scholar] [CrossRef]
  3. Shaban, K.B.; Kadri, A.; Rezk, E. Urban air pollution monitoring system with forecasting models. IEEE Sens. J. 2016, 16, 2598–2606. [Google Scholar] [CrossRef]
  4. Wen, C.; Liu, S.; Yao, X.; Peng, L.; Li, X.; Hu, Y.; Chi, T. A novel spatiotemporal convolutional long short-term neural network for air pollution prediction. Sci. Total Environ. 2019, 654, 1091–1099. [Google Scholar] [CrossRef]
  5. Bekkar, A.; Hssina, B.; Douzi, S.; Douzi, K. Air-pollution prediction in smart city, deep learning approach. J. Big Data 2021, 8, 161. [Google Scholar] [CrossRef] [PubMed]
  6. Du, S.; Li, T.; Yang, Y.; Horng, S.J. Deep air quality forecasting using hybrid deep learning framework. IEEE Trans. Knowl. Data Eng. 2019, 33, 2412–2424. [Google Scholar] [CrossRef]
  7. Aono, Y.; Hayashi, T.; Wang, L.; Moriai, S. Privacy-preserving deep learning via additively homomorphic encryption. IEEE Trans. Inf. Forensics Secur. 2017, 13, 1333–1345. [Google Scholar]
  8. Puthal, D.; Sahoo, B.P.; Mishra, S.; Swain, S. Cloud computing features, issues, and challenges: A big picture. In Proceedings of the 2015 International Conference on Computational Intelligence and Networks, Odisha, India, 12–13 January 2015; pp. 116–123. [Google Scholar]
  9. Singh, A.; Chatterjee, K. Cloud security issues and challenges: A survey. J. Netw. Comput. Appl. 2017, 79, 88–115. [Google Scholar] [CrossRef]
  10. Bai, Y.; Zeng, B.; Li, C.; Zhang, J. An ensemble long short-term memory neural network for hourly PM2.5 concentration forecasting. Chemosphere 2019, 222, 286–294. [Google Scholar] [CrossRef]
  11. Hrdličková, Z.; Michálek, J.; Kolář, M.; Veselỳ, V. Identification of factors affecting air pollution by dust aerosol PM10 in Brno City, Czech Republic. Atmos. Environ. 2008, 42, 8661–8673. [Google Scholar] [CrossRef]
  12. Dickson, R. Meteorological factors affecting particulate air pollution of a city. Bull. Am. Meteorol. Soc. 1961, 42, 556–560. [Google Scholar] [CrossRef]
  13. Xue, T.; Zheng, Y.; Geng, G.; Zheng, B.; Jiang, X.; Zhang, Q.; He, K. Fusing observational, satellite remote sensing and air quality model simulated data to estimate spatiotemporal variations of PM2. 5 exposure in China. Remote Sens. 2017, 9, 221. [Google Scholar] [CrossRef]
  14. Han, J.; Liu, H.; Xiong, H.; Yang, J. Semi-supervised air quality forecasting via self-supervised hierarchical graph neural network. IEEE Trans. Knowl. Data Eng. 2022, 35, 5230–5243. [Google Scholar] [CrossRef]
  15. Liu, B.; Yu, X.; Chen, J.; Wang, Q. Air pollution concentration forecasting based on wavelet transform and combined weighting forecasting model. Atmos. Pollut. Res. 2021, 12, 101144. [Google Scholar] [CrossRef]
  16. Wang, H.; Zhang, L.; Wu, R.; Cen, Y. Spatio-temporal fusion of meteorological factors for multi-site PM2. 5 prediction: A deep learning and time-variant graph approach. Environ. Res. 2023, 239, 117286. [Google Scholar] [CrossRef]
  17. Pan, J.; McElhannon, J. Future edge cloud and edge computing for internet of things applications. IEEE Internet Things J. 2017, 5, 439–449. [Google Scholar] [CrossRef]
  18. Ni, X.; Huang, H.; Du, W. Relevance analysis and short-term prediction of PM2. 5 concentrations in Beijing based on multi-source data. Atmos. Environ. 2017, 150, 146–161. [Google Scholar] [CrossRef]
  19. Wang, P.; Wang, P.; Chen, K.; Du, J.; Zhang, H. Ground-level ozone simulation using ensemble WRF/Chem predictions over the Southeast United States. Chemosphere 2022, 287, 132428. [Google Scholar] [CrossRef]
  20. Gourav.; Rekhi, J.K.; Nagrath, P.; Jain, R. Forecasting air quality of Delhi using ARIMA model. In Advances in Data Sciences, Security and Applications: Proceedings of ICDSSA 2019; Springer: Singapore, 2020; pp. 315–325. [Google Scholar]
  21. Liu, H.; Li, Q.; Yu, D.; Gu, Y. Air quality index and air pollutant concentration prediction based on machine learning algorithms. Appl. Sci. 2019, 9, 4069. [Google Scholar] [CrossRef]
  22. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  23. Zhao, J.; Deng, F.; Cai, Y.; Chen, J. Long short-term memory-Fully connected (LSTM-FC) neural network for PM2. 5 concentration prediction. Chemosphere 2019, 220, 486–492. [Google Scholar] [CrossRef]
  24. Sayeed, A.; Choi, Y.; Eslami, E.; Lops, Y.; Roy, A.; Jung, J. Using a deep convolutional neural network to predict 2017 ozone concentrations, 24 h in advance. Neural Netw. 2020, 121, 396–408. [Google Scholar] [CrossRef] [PubMed]
  25. Han, S.; Dong, H.; Teng, X.; Li, X.; Wang, X. Correlational graph attention-based Long Short-Term Memory network for multivariate time series prediction. Appl. Soft Comput. 2021, 106, 107377. [Google Scholar] [CrossRef]
  26. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  27. Nguyen, D.V.; Zettsu, K. Spatially-distributed federated learning of convolutional recurrent neural networks for air pollution prediction. In Proceedings of the 2021 IEEE International Conference on Big Data (Big Data), Orlando, FL, USA, 15–18 December 2021; pp. 3601–3608. [Google Scholar]
  28. Velentzas, P.; Vassilakopoulos, M.; Corral, A. GPU-aided edge computing for processing the k nearest-neighbor query on SSD-resident data. Internet Things 2021, 15, 100428. [Google Scholar] [CrossRef]
  29. Ghimire, B.; Rawat, D.B. Recent advances on federated learning for cybersecurity and cybersecurity for federated learning for internet of things. IEEE Internet Things J. 2022, 9, 8229–8249. [Google Scholar] [CrossRef]
  30. Ma, J.; Yu, Z.; Qu, Y.; Xu, J.; Cao, Y. Application of the XGBoost machine learning method in PM2. 5 prediction: A case study of Shanghai. Aerosol Air Qual. Res. 2020, 20, 128–138. [Google Scholar] [CrossRef]
  31. Lian, M.; Liu, J. Single Pollutant Prediction Approach by Fusing MLSTM and CNN. In Proceedings of the International Conference on Knowledge Science, Engineering and Management, Singapore, 6–8 August 2022; pp. 129–140. [Google Scholar]
  32. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  33. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. Adv. Neural Inf. Process. Syst. 2014, 27, 3104–3112. [Google Scholar]
  34. Zheng, H.; Xu, Z.; Wang, Q.; Ding, Z.; Zhou, L.; Xu, Y.; Su, H.; Li, X.; Zhang, F.; Cheng, J. Long-term exposure to ambient air pollution and obesity in school-aged children and adolescents in Jiangsu province of China. Environ. Res. 2021, 195, 110804. [Google Scholar] [CrossRef]
  35. Zhou, Y.; Zhao, Y.; Mao, P.; Zhang, Q.; Zhang, J.; Qiu, L.; Yang, Y. Development of a high-resolution emission inventory and its evaluation and application through air quality modeling for Jiangsu Province, China. Atmos. Chem. Phys. 2017, 17, 211–233. [Google Scholar] [CrossRef]
Figure 1. The overview of FedDeep.
Figure 1. The overview of FedDeep.
Applsci 14 01979 g001
Figure 2. The architecture of the local ESTNet deployed on the ECS i. First, fully connected networks (FCs) are adopted to transform urban P M 2.5 and external factors into H X P × n i × d model and H T E P × n i × d model , respectively. Then, the gating fusion layer is utilized to fuse these two features, which is represented as H P × n i × d model . Here, the fused feature at timestep t j is H t j n i × d model . Furthermore, a spatio-temporal learning layer is proposed to mine spatio-temporal correlations and the output representations of timestep t j are H S T t j n i × d model . Finally, we leverage the decoder of seq2seq joint FCs to produce the prediction sequences, where Y t P + k n i × F denotes the predicted value of n i sensors in the urban region i at the predicted timestep t P + k . In addition, ⊙ is a dot product. σ and t a n stand for the activation functions s i g m o i d and t a n h .
Figure 2. The architecture of the local ESTNet deployed on the ECS i. First, fully connected networks (FCs) are adopted to transform urban P M 2.5 and external factors into H X P × n i × d model and H T E P × n i × d model , respectively. Then, the gating fusion layer is utilized to fuse these two features, which is represented as H P × n i × d model . Here, the fused feature at timestep t j is H t j n i × d model . Furthermore, a spatio-temporal learning layer is proposed to mine spatio-temporal correlations and the output representations of timestep t j are H S T t j n i × d model . Finally, we leverage the decoder of seq2seq joint FCs to produce the prediction sequences, where Y t P + k n i × F denotes the predicted value of n i sensors in the urban region i at the predicted timestep t P + k . In addition, ⊙ is a dot product. σ and t a n stand for the activation functions s i g m o i d and t a n h .
Applsci 14 01979 g002
Figure 3. Multi-head self attention (MHSA) of the spatial dependencies among different sensors in the urban region i.
Figure 3. Multi-head self attention (MHSA) of the spatial dependencies among different sensors in the urban region i.
Applsci 14 01979 g003
Figure 4. P M 2.5 Prediction results with the comparison models. (a) SVR for Pukou site of Nanjing city, (b) SVR for Nanmen site of Suzhou city, (c) GAT-LSTM for Pukou site of Nanjing city, (d) GAT-LSTM for Nanmen site of Suzhou city, (e) FedDeep for Pukou site of Nanjing city, and (f) FedDeep for Nanmen site of Suzhou city.
Figure 4. P M 2.5 Prediction results with the comparison models. (a) SVR for Pukou site of Nanjing city, (b) SVR for Nanmen site of Suzhou city, (c) GAT-LSTM for Pukou site of Nanjing city, (d) GAT-LSTM for Nanmen site of Suzhou city, (e) FedDeep for Pukou site of Nanjing city, and (f) FedDeep for Nanmen site of Suzhou city.
Applsci 14 01979 g004
Figure 5. Experimental results for different hyperparameter configurations. The vertical axis represents the values of MAE, RMSE, and R 2 .
Figure 5. Experimental results for different hyperparameter configurations. The vertical axis represents the values of MAE, RMSE, and R 2 .
Applsci 14 01979 g005
Table 1. Detailed descriptions about the distribution of the air quality monitoring sites in 13 prefecture-level urban regions, Jiangsu province, and the types of air pollution and external factors.
Table 1. Detailed descriptions about the distribution of the air quality monitoring sites in 13 prefecture-level urban regions, Jiangsu province, and the types of air pollution and external factors.
Urban RegionThe Number
of Sites
The Variables of
Air Pollution
The Variables of External Factor
Nanjing13 P M 2.5 weather conditions, temperature,
and wind speed
Suzhou24
Wuxi16
Changzhou13
Zhenjiang8
Nantong9
Taizhou7
Yangzhou6
Yancheng7
Huaian6
Suqian5
Xuzhou10
Lian yungang9
Table 2. Implementation details of FedDeep.
Table 2. Implementation details of FedDeep.
Each local ESTNetThe number of heads in MHSA N h = 4
The dimensions of each head N d = 8
The MHSA layers L M H S A = 3
LSTM layers in the spatio-temporal learning layer L L S T M ( S T ) = 1
LSTM layers in the decoder of seq2seq L L S T M ( d e c ) = 1
The Global ESTNetThe number of heads in MHSA N h = 4
The dimensions of each head N d = 8
The MHSA layers L M H S A = 3
LSTM layers in the spatio-temporal learning layer L L S T M ( S T ) = 1
LSTM layers in the decoder of seq2seq L L S T M ( d e c ) = 1
d m o d e l 32
Historical timesteps P = 12
Prediction timesteps Q = 12
The Central Cloud1 A100 card
ECS13 Nvidia 3090 Ti GPU cards
Batchsize32
Learning rate0.0001
Epochs64
OptimizerStochastic Gradient Descent
Table 3. The performance comparison of the 12 h P M 2.5 forecasting task at 133 air quality monitoring sites of 13 prefecture-level cities in Jiangsu province.
Table 3. The performance comparison of the 12 h P M 2.5 forecasting task at 133 air quality monitoring sites of 13 prefecture-level cities in Jiangsu province.
ModelMAERMSER2TTGPUM
HA23.1936.570.86924.87 min-
SVR22.7532.360.87825.76 min-
LSTM17.6626.340.90443.55 min1.53 GB
CNN-LSTM15.6424.900.91354.19 min2.45 GB
GAT-LSTM13.1321.820.94567.20 min3.27 GB
FedDeep (w/o)12.4319.320.97365.28 min3.12 GB
FedDeep12.3819.190.97228.32 min0.87 GB
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hu, Y.; Cao, N.; Guo, W.; Chen, M.; Rong, Y.; Lu, H. FedDeep: A Federated Deep Learning Network for Edge Assisted Multi-Urban PM2.5 Forecasting. Appl. Sci. 2024, 14, 1979. https://doi.org/10.3390/app14051979

AMA Style

Hu Y, Cao N, Guo W, Chen M, Rong Y, Lu H. FedDeep: A Federated Deep Learning Network for Edge Assisted Multi-Urban PM2.5 Forecasting. Applied Sciences. 2024; 14(5):1979. https://doi.org/10.3390/app14051979

Chicago/Turabian Style

Hu, Yue, Ning Cao, Wangyong Guo, Meng Chen, Yi Rong, and Hao Lu. 2024. "FedDeep: A Federated Deep Learning Network for Edge Assisted Multi-Urban PM2.5 Forecasting" Applied Sciences 14, no. 5: 1979. https://doi.org/10.3390/app14051979

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop