Next Article in Journal
A Resilient LoRa-Based Solution to Support Pervasive Sensing
Next Article in Special Issue
Graph Embedding-Based Money Laundering Detection for Ethereum
Previous Article in Journal
Automatic Fabric Defect Detection Method Using AC-YOLOv5
Previous Article in Special Issue
A Review of Cuckoo Filters for Privacy Protection and Their Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Privacy Robust Aggregation Method Based on Federated Learning in the IoT

1
Shanghai Lanchang Automation Technology Co., Ltd., Shanghai 201906, China
2
College of Artificial Intelligence, Nanjing Agricultural University, Nanjing 210095, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(13), 2951; https://doi.org/10.3390/electronics12132951
Submission received: 27 May 2023 / Revised: 26 June 2023 / Accepted: 30 June 2023 / Published: 5 July 2023
(This article belongs to the Special Issue Novel Methods Applied to Security and Privacy Problems)

Abstract

:
Federated learning has been widely applied because it enables a large number of IoT devices to conduct collaborative training while maintaining private data localization. However, the security risks and threats faced by federated learning in IoT applications are becoming increasingly prominent. Except for direct data leakage, there is also a need to face threats that attackers interpret gradients and infer private information. This paper proposes a Privacy Robust Aggregation Based on Federated Learning (PBA), which can be applied to multiple server scenarios. PBA filters outliers by using the approximate Euclidean distance calculated from binary sequences and the 3 σ criterion. Then, this paper provides correctness analysis and computational complexity analysis on the aggregation process of PBA. Moreover, the performance of PBA is evaluated concerning ensuring privacy and robustness in this paper. The results indicate that PBA can resist Byzantine attacks and a state-of-the-art privacy inference, which means that PBA can ensure privacy and robustness.

1. Introduction

Internet of Things (IoT) applications can truly realize smart cities, which would require a massive number of IoT devices [1]. Owing to remarkable growth in IoT network size and data volume, these IoT devices can be made possible by centralized machine learning methods [2]. But as centralized machine learning techniques need to transfer the data to a centralized server for training, they are faced with the inherent issues of privacy leakage [3]. Moreover, centralized machine learning may not be practicable with large and dispersed data. Therefore, federated learning (FL) has been introduced to build intelligent and privacy-enhanced IoT systems [4,5,6].
FL is a distributed collaborative approach that coordinates the data training without sharing datasets [7], while FL itself has privacy concerns, it does not completely guarantee privacy [8,9,10]. Firstly, FL aims to train a model on multiple dispersed devices collaboratively by transferring local model updates, while only local update needs to be shared during the training process, there is also a high possibility of personal information leakage. This means that if sufficient measures are not taken, it will cause serious data privacy leakage risks. Secondly, the model inversion attack is a type of privacy attack that tries to infer the training data [11,12]. Deep Leakage from Gradients (DLG) [13] and Generative Adversarial Networks (GAN) [14] attacks are both types of model inversion attacks. DLG attack is a gradient-based attack method that can infer the input data of a model by accessing its gradient information. GAN attack is an attack method based on a generative adversarial network. It can generate data similar to the target model by training a generator network and then infer the input data of the model. Furthermore, due to the autonomy of workers in FL environments, any worker that participates in the process may be malicious and disrupt the security of FL [15]. These malicious workers can submit threatening local updates and malicious models, intentionally interfere with the convergence of the global model, and affect the efficiency of FL. Above all, FL is still affected by privacy inference [13,16,17] and Byzantine attacks [18,19]. And these security risks and threats faced by FL in IoT applications are also increasingly prominent [12,20,21]. It is essential to study privacy protection and Byzantine-robust FL methods in IoT scenarios.
The contributions of this paper are summarized as follows:
  • We design an approximate Euclidean distance calculated by the sign matrix and data matrix, where the sign matrix and data matrix are composed of the sign and absolute value of the model update from each worker. It can decrease the computational complexity.
  • We propose a privacy robust aggregation method based on FL (PBA) in a multi-server scenario. PBA is implemented based on approximate Euclidean distance and adopts scientific 3 σ criterion instead of sorting to screen outliers. We also analyze the correctness and computational complexity of PBA.
  • We conduct experiments and evaluate the performance of PBA concerning privacy security, Byzantine robustness, and time costs. The results indicate that PBA can ensure privacy and robustness with less time.
The rest of this paper is organized as follows: Section 2 summarizes the works related to ours while highlighting the differences; Section 3 introduces the threat model and some basic techniques; Section 4 designs a privacy robust aggregation rule based on approximate Euclidean distance and analyzes the correctness and computational complexity; Section 5 conducts experimental evaluation; Section 6 concludes the paper.

2. Related Work

Compared with traditional machine learning, FL already increases the level of privacy. It mitigates the transmission of sensitive data and prevents a third party from performing learning tasks on the unpermitted individual data [22]. However, the local updates/gradients uploaded by individuals, especially parts of personal data that are sensitive to specific features or values, may reveal sensitive information about users’ data. Moreover, bad data may be uploaded because of device malfunction or malicious behavior.
Many studies have been conducted to solve the issues mentioned above. Currently, studies on privacy-preserving FL use either secure multi-party computation (MPC) or differential privacy (DP). By combining the locally trained classifiers, Pathak et al. [23] suggested a DP-based global classifier. A framework was proposed that adjusted the objective function during the training process to achieve DP [24]. Other methods based on encryption include homomorphic encryption (HE) and secure aggregation protocol [25]. A privacy-preserving deep learning (PPDL) algorithm was proposed [26], in which a number of distributed participants work together to train a deep learning model using local data. They established a trade-off between efficiency and security for the number of clients taking part in the training process. For Byzantine resilience, Krum [27] computed distance based on Euclidean norm between gradients to filter outliers. Different from Krum [27], coordinate-median and trimmed-mean [28] performed element-wise aggregation rules. After that, Yang et al. [29,30] presented ByRDiE and BRIDGE based on trimmed-mean from centralized to decentralized systems. At every epoch for every regular client, BRIDGE [30] performs the exchange and updates its local model’s coordinates, whereas ByRDiE [29] executes multiple scalar exchanges and updates. Essentially combining Krum [27] and a variant of trimmed-mean [28], Bulyan [31] was proposed to reduce the leeway of Byzantine clients. To provide robustness against a poisoning attack that lowers global accuracy, a lightweight federated multi-task learning framework [32] was proposed. Anomaly detection [33], principal component analysis [34], and so on were applied to identify malicious clients. Instead of considering the relationship between privacy and robustness, these efforts concentrated on privacy protection or Byzantine resilience in FL.
To achieve privacy-preserving and Byzantine-resilience, So et al. [35] proposed a FL algorithm based on secure aggregation. Secret sharing (SS) is used, and any two parties need to negotiate a random number to ensure the security of the information. However, this may require a large number of rounds of interaction, leading to higher communication complexity. Guo et al. [36] proposed an Uniform Byzantine-resilient Aggregation Rule (UBAR) to defeat an arbitrary number of Byzantine nodes. The main idea is to use training samples to test the performance of parameters that are benign clients. This method can converge quickly, but the balance between privacy protection and Byzantine robustness needs to be struck. FoolsGold [37] calculated the cosine similarity between the historical gradients of the clients to mitigate sybils. However, this scheme directly analyzed and calculated the gradients in plain text, which resulted in privacy leakage.

3. Preliminary

In this section, we introduce some techniques involved in this paper. The symbols and notations used in the remainder of this paper are given in Table 1.

3.1. Federated Learning

Data are the cornerstone of model training in the field of artificial intelligence. However, data is frequently present in the form of data islands. Data processing in a centralized way is the direct remedy for data islands, but data leakage may occur during the collection and processing stages [38]. To this end, FL is developed and received widespread attention.
FL seeks to train a global collectively on numerous datasets distributed on individual devices without explicitly transferring data samples [39,40]. In IoT networks, several IoT devices can act as workers who interact with a server to train neural networks. The server first initializes a global model. Each worker downloads the most recent model, computes its own model update, and then sends the computed update to the server. Subsequently, the server aggregates all local updates and updates the global model [5]. Local training, aggregation, update, and other steps mentioned above are repeated until the termination condition is met. The preset training epochs or global model accuracy can be considered as termination conditions.
Although the development of FL solves the problem of data islands and reduces the risk of data leakage, many threats and challenges still need to be addressed urgently. The most core problems include the weakness of communication efficiency [41,42,43], the defects of privacy security [44,45,46], the lack of incentive mechanisms [47,48] and so on.

3.2. Privacy Protection Technology

Privacy computing is a set of technologies that can achieve the goal of “usable and invisible” data. Homomorphic encryption (HE), secret sharing (SS), and differential privacy (DP) are presented in this section.

3.2.1. Homomorphic Encryption

HE [49] is a type of encryption that preserves the functionality and format of the encrypted data while allowing a third party to perform some compute operations on the encrypted data. It is usually defined as:
If an encryption scheme satisfies the following equation, then the scheme is called homomorphic with respect to the operation ★:
E ( m 1 ) E ( m 2 ) = E ( m 1 m 2 ) , m 1 , m 2 M
where E represents the encryption algorithm, and M represents the collection of all information [50].
HE can be used to withstand various security threats, such as membership inference attacks [51], DLG attack [13], chosen plain text attacks [52], etc. Therefore, it has good application prospects in the field of information security.

3.2.2. Secret Sharing

SS is a technique that divides a secret into multiple parts and distributes them to different participants. The secret can only be reconstructed when specific conditions are met [53,54].
Taking the Shamir [55] as an example, it is a classic ( t , n ) threshold SS method. A secret owner and a group of participants are involved in this method. The owner splits a secret into n shards and transfers them to n participants. The secret can be recovered only by simultaneously obtaining t shards. The specific implementation process of Shamir [55] is as follows:
  • Encryption
    Assuming there is a secret S, take any t 1 random numbers and construct a polynomial f ( x ) = a 0 + a 1 · x + a 2 · x 2 + + a ( t 1 ) · x ( t 1 ) , where a 0 = S . It should be noted that all operations are performed in a finite field. Then take any n numbers, substitute x 1 , x 2 , , x n into the polynomial to obtain f ( x 1 ) , f ( x 2 ) , , f ( x n ) 4 . Send ( x i , f ( x i ) ) to the participants.
  • Decryption
    Polynomial coefficient calculation using secret shards held by any t participants:
    a 0 + a 1 · x 1 + a 2 · ( x 1 ) 2 + + a t 1 · ( x 1 ) t 1 = y 1 , a 0 + a 1 · x 2 + a 2 · ( x 2 ) 2 + + a t 1 · ( x 2 ) t 1 = y 2 , a 0 + a 1 · x 3 + a 2 · ( x t ) 2 + + a t 1 · ( x t ) t 1 = y t .
    After obtaining the coefficients, substitute them into the polynomial, take x = 0 , and you can obtain the secret S = a 0 .

3.2.3. Differential Privacy

DP [56] protects privacy by merging client data with added noise. This noise makes it difficult for attackers to determine which data belongs to a specific client, thereby increasing the privacy of the data. Local differential privacy (LDP) and global differential privacy (GDP) are variants of DP [57], which are used to protect privacy.
For LDP [58], each worker’s data is added with noise. On the contrary, GDP [59] needs to send all workers’ data to the central server for processing. This makes GDP more suitable for situations where the entire dataset needs to be analyzed. Obviously, the difference between LDP and GDP is that LDP is a method to achieve DP on a single dataset, while GDP is a method to add noise on multiple datasets.

3.3. Outlier Detection Method

Outliers, which may be caused by human errors, mechanical failures, and changes in system behavior, or natural deviations in the environment, can be identified through outlier detection [60], a method that discovers significant deviations from normal or mean values in given data.

3.3.1. 3 σ Criterion

In 3 σ criterion [61], outliers are defined as a collection of measurements that deviate from the mean by more than three times the standard deviation. Under normal distribution, the chance of values appearing beyond 3 σ from the mean is p ( | x μ | > 3 σ ) 0.003 , which is an extremely unusual event. If the data is not normally distributed, a · σ can be used to characterize it. Therefore, Curtis et al. [62] proposed the z-score method, which measures the distance of a certain raw score x from the mean in standard deviation units. That is,
z = x μ σ
z-score typically considers data points more than three times the standard deviation from the mean as outliers. In other words, data points with z-score  > a are considered outliers, where a can be set according to specific circumstances.

3.3.2. DBSCAN

Density-based Spatial Clustering of Applications with Noise (DBSCAN) [63] is a density-based spatial clustering algorithm. This technique finds clusters of any shape in a noisy spatial database and divides areas with sufficient density into one class. The core idea of DBSCAN is to start from a certain core point and continuously expand to a region where the density can reach so as to obtain a maximized region containing core points and boundary points. Any two points in the region are connected in density. Moreover, some studies use DBSCAN to perform outlier filters. The disadvantage is that the application of DBSCAN results in a longer time required for model training.

4. Privacy Robust Aggregation Based on Federated Learning

In this section, we present the system structure and threat model. Moreover, we design an approximate Euclidean distance calculated by the sign matrix and data matrix and propose a Privacy Robust Aggregation Based on Federated Learning (PBA) aggregation rule.

4.1. System Structure and Threat Model

Figure 1 depicts our system model. There are C customers in the overall structure, of which f is Byzantine. The gradients sent by these Byzantine clients may interfere with the convergence and precision of the model. These clients interact with S ( S 2 ) servers. These servers are required to carry out the identification of outliers and global model updates.
On the server side, the threat model is classed as honest-but-curious (semi-honest). This means that servers follow the established protocol without deviating but could try to extract private information from the data transmitted with them. So, if these servers obtain raw gradients, they might employ established methods to extract sensitive client data. Only a fraction of clients may be Byzantine. They may intentionally produce false data in an effort to undermine the training.

4.2. Algorithm Overview

The privacy-robust aggregation rule PBA proposed in this paper is implemented based on S servers, where S 2 . Suppose that there are n participants in the system, out of which f are Byzantine. The servers in the system are honest and curious (semi-honest) but do not collide with each other. This implies that they follow algorithmic processes, do not disclose each other’s information, but may attempt to access user privacy and restore private data. Similarly, n participants are also honest and curious. However, f malicious participants might upload random data, leading to the deviation of the global model from its standard training process. Figure 1 illustrates the overall framework of PBA, where g i j represents the gradient information sent by participant C i to the server S j .
During the model initialization stage, S servers negotiate to create S ( S 1 ) 2 random vectors. These vectors need to satisfy the following equation:
i = 1 S ( S 1 ) 2 ( S i + 1 ) · r i = c
where r i represents the i-th vector generated by servers and c is a constant. Then, each client downloads the most recent global model and computes local updates. Before uploading the updates to the servers, the following operations need to be performed:
  • Differential Privacy and Denoise
    To prevent the data leakage in servers, client C i creates private noise and adds it to the local gradient g i , i.e., g ˜ i g i + N ( 0 , 2 · σ 2 ) . But noises added to the gradients have an impact on model convergence and accuracy. Hence, It is necessary to denoise the gradients after adding noises. Kolmogorov–Smirnov (KS) distance is used as a measure of denoising [64]. C i applies it to denoise the gradient. That is, g ^ i K S ( g ˜ i , N ) , where K S ( ) represents the KS measure.
    The above operations can not only preserve the characteristics of DP to reduce the risk of data leakage but also eliminate the impact of noise on the model convergence and accuracy [64].
  • Gradient Encoding
    Each client needs to perform some mask operations to provide further protection. The specific masking process is shown below:
    g ^ i j = g ^ i + p : p j r p + p : p j r p
    where g ^ i j represents the encoded gradient that C i sends to the server S j .
Above all, C i generates the gradient g ^ i j based on its own local update g i , which undergoes DP, denoise, and encoding operations and sends to the server S j . Then, when S j receives the gradient information, it needs to calculate the approximate Euclidean distance between two gradients, perform outlier detection through 3 σ criterion, and combine gradients.
After receiving C encoded gradient information, S j calculates the approximate Euclidean distance. And the approximate distance between C i and C j is temporarily recorded as A E D ( i , j ) . Then, S j detects outliers in the approximate distances based on 3 σ criterion. In the application of PBA, outliers can be defined as values with deviations from the mean exceeding a times, where a is not necessarily three. Thirdly, S j calculates the sum of all gradients based on the filtration results, i.e.,
g S j = g i N S i g i + n · ( p : p j r p q : q > j r q )
The computation result g S j is sent to other servers while S j receives corresponding results from them. S j combines gradients as follows:
g = 1 S g S j ( n f ) · c
Finally, each server completes this epoch by updating the global model. During the initialization phase, S servers negotiate and generate random vectors r i . Then, The process of PBA is shown in Algorithm 1, where f A E D ( · ) represents the function that calculates the approximate Euclidean distance.
Algorithm 1 PBA.
Input:
  The global model in the t-th round, x t ;
Output:
  The gradient aggregated in the t-th round, g;
  1:
for client C i , i = 1 , 2 , , C : do
  2:
   randomly sample from the local dataset for local training g i = f ( x t , ξ i ) ;
  3:
   add private noise g ˜ i g i + N ( 0 , 2 σ 2 ) ;
  4:
   denoise g ^ i K S ( g ˜ i , N ) ;
  5:
   encode g ^ i j = g ^ i + p : p j r p + p : p > j r p ;
  6:
   send g ^ i j to the server S j ;
  7:
end for
  8:
for server S j , j = 1 , 2 , , S : do
  9:
   calculate approximate Euclidean distance A E D ( i , j ) = f A E D ( g ^ p j , g ^ q j ) ;
10:
   filter out outliers based on the 3 σ criterion, denoted as N S i ;
11:
   calculate the sum of gradients g S i = g i N S j + n · ( p : p i r p + p : p > i r p ) ;
12:
   send the g S j to other servers and receive gradients from them;
13:
   compute the average g = 1 S · g S j ( n f ) · c ;
14:
end for
15:
return g

4.3. Approximate Euclidean Distance

Due to the characteristics of data dispersion and multi-device collaborative training, Training Time Attack (TTA), including Model Poisoning Attack and Data Poisoning Attack, is a significant threat to the security of FL system [65]. Studies explored robust aggregation techniques, such as Krum [27] and Median [28], to handle devices that send corrupted updates to the server. Wang et al. [66] proposed that the cosine value between any two gradients is only related to their direction, not magnitude. Therefore, they replace the gradient with its sign rather than specific values, represented as a binary sequence, to highlight the direction differences between any two gradients. However, this method ignores the impact of gradient sizes on model convergence. In distance-based robust aggregation methods, Euclidean distance is the most typical distance measure. It is computed as follows:
d i s t a n c e ( i , j ) = ( g i [ k ] g j [ k ] ) 2
where g i [ k ] and g j [ k ] denote the k-th bit of local training gradients g i and g j , respectively. However, distance-based robust aggregation methods could incur significant computational overhead with numerous clients in the system [9,67].
A method which transfers only gradient signs and computes cosine similarity through the binary sequence is proposed, which greatly reduces the communication overhead of FL system [66]. Based on this method, an approximate Euclidean distance based on binary sequences is designed. Let g i and g j denote the gradients of C i and C j , respectively, and g denote their average gradient. The approximate distance between them is computed as follows:
  • Servers form sign matrices and data matrices by taking the sign and absolute value of g i and g j , respectively. The sign matrices are denoted as S i and S j , while the data matrices are denoted as D i and D j .
  • An XOR operation between the two sign matrices is performed, i.e., s i g n ( i , j ) = S i S i . At the same time, servers need to subtract the two data matrices element-wise, take the absolute value, and compare it to g. Specifically,
    d a t a ( i , j ) = d a t a ( i , j ) [ k ] = 0 , d a t a ( i , j ) [ k ] < g [ k ] ; d a t a ( i , j ) [ k ] = 1 , d a t a ( i , j ) [ k ] g [ k ] ;
    where d a t a ( i , j ) [ k ] = | D i [ k ] D j [ k ] | .
  • s i g n ( i , j ) and d a t a ( i , j ) are combined, resulting in four different binary sequences: 00, 01, 10, and 11, which are represented as 0, 1, 2, and 3 in decimal form to reflect the degree of maliciousness of the gradients under different scenarios. Specifically,
    A E D ( i , j ) = 0 , ( s i g n , d a t a ) = 00 1 , ( s i g n , d a t a ) = 01 2 , ( s i g n , d a t a ) = 10 3 , ( s i g n , d a t a ) = 11
The four binary sequences mentioned above represent four scenarios: same-sign with a small value difference, same-sign with a large value difference, opposite-sign with a small value difference, and opposite-sign with a large value difference.

4.4. Theoretical Analysis

4.4.1. Correctness Analysis

PBA first performs DP for local updates and then reduces the noise according to the added noise in DP. In addition, due to the post-processing property of DP, the execution of denoising does not affect the differential property and can still maintain the model accuracy losslessly [64,66]. Therefore, privacy security can still be guaranteed. We provide the correctness analysis and proof for aggregation based on gradient encoding.
Firstly, the correctness of distance computed on encoded gradients is analyzed. For S i , the gradient received from clients can be represented as g ^ x i = g ^ x + p : p i r p + p : p > i r p , x [ n ] . Therefore, the distance of the gradients between any two clients C x and C y can be expressed as
d i s t a n c e ( i , j ) = ( g ^ x i g ^ y i ) 2 = [ ( g ^ x + p : p i r p + p : p > i r p ) ( g ^ y + p : p i r p + p : p > i r p ) ] 2 = ( g ^ x g ^ y ) 2
It can be seen that the result computed on the server S i is consistent with the result whose gradients are not encoded.
Then, a correctness proof of aggregation is provided. Suppose among n clients, f malicious clients are selected. The gradient sum calculated by server S i is represented as g S j = g i N S i g i + n · ( p : p j r p q : q > j r q ) . Then,
g S i = [ g i N S i g i + n · ( p : p j r p q : q > j r q ) ] = S · g i N g i + ( n f ) · [ S · r 1 + + ( S i + 1 ) · r i + + r S ] = S · [ g i N g i + ( n f ) · c ]
Thus, the gradient for the model update can be computed as follows:
g = 1 S · g S j ( n f ) · c = g i N g i
where N refers to the set of potential benign gradients.

4.4.2. Complexity Analysis

PBA is based on approximate Euclidean distance, which can reduce computational overheads. Hence, we analyze the computational complexity of approximate Euclidean distance and compare it with Euclidean distance.
Proposition 1.
Let n represent the dimension of a model update, k represent the bits of each dimension. Suppose that the time required to perform a one-bit XOR is unit time, the time required for addition and subtraction is t times the unit time, and the time required for multiplication and division is t times the unit time. Then, the computational complexity of approximate Euclidean distance is
O = ( 2 · k + 1 ) · n
Proof. 
For approximate Euclidean distance, it needs to perform an XOR operation between two sign matrices, which requires n times unit time. Moreover, subtraction is performed between data matrices, and the required time is n · k times unit time. Finally, a comparison operation spends n · k times unit time. Hence, the time spent computing an approximate Euclidean distance is O = ( 2 · k + 1 ) · n . □
For Euclidean distance, the calculation method is shown in Formula (8), including subtraction, multiplication, addition, and square root operations. From this, it can be concluded that the computational complexity of Euclidean distance is O = 2 · ( t + 2 ) · n · k . Then, after comparison, the computational complexity of approximate Euclidean distance is about 1 ( t + 1 ) · k times that of Euclidean distance.

5. Performance Evaluation

5.1. Experiment Setup

All trials are conducted on a computer equipped with an AMD Ryzen 7 5800H, Radeon Graphics 3.20 GHz, and NVIDIA GeForce GTX 1650. And PyTorch [68] is used for training in Python. In addition, a task for object recognition is developed that involves n clients working together to train a convolutional neural network (CNN) with two convolutional layers and two fully connected layers.
Assuming there are S samples and L categories in the dataset, two methods are studied for dividing the dataset among the clients:
  • Independent Identically Distribution (IID)
    In this method, data is shuffled and split into S C samples per client.
  • Non-Independent Identically Distribution (Non-IID)
    In this method, the data set is sorted and divided into C parts of size S C , assigning one part to each client. Non-IID data will be used to study the privacy security of PBA in order to reflect the leakage of specific sensitive information from a client.
    In the experiment, the MNIST [69] image dataset is used to train a CNN model, which is a handwritten digit (0–9) dataset consisting of 60,000 training samples and 10,000 test samples, with each sample being a 28 × 28 grayscale image
Moreover, in order to evaluate the Byzantine robustness, several Byzantine attacks are considered in the experiment:
  • Gaussian Attack (GA)
    This is an aimless poisoning attack aimed at reducing the accuracy of the model. Specifically, malicious clients can randomly sample from the Gaussian distribution and upload it as local update parameters, namely:
    g ^ [ d ] = G a u s s i a n ( μ , σ 2 )
    Among them, μ represents the mathematical expectation of a Gaussian distribution, and σ 2 represents its variance.
  • Label Flipping Attack (LFA)
    This is a targeted poisoning attack. Malicious clients flip the labels of local data to generate error gradients [66], especially flipping the labels of each sample from x to L 1 x , x ( 0 , , L 1 ) .
Furthermore, some baselines are compared with PBA, such as Krum [27], trimmed-mean [28] and UBAR [36]. And following performance metrics will be considered:
  • Global Accuracy: Test the accuracy of the trained model on the validation set;
  • Loss Value: Test the cross-entropy of the training model on the validation machine;
  • Reconstructed Images: Image reconstruction under DLG attack.
A total of C = 20 clients are set up, with f Byzantine clients. We compare the global accuracy and loss of the model under different numbers of malicious clients. When evaluating the security of PBA, the DLG attack will be conducted to see if user-sensitive information in images can be reconstructed.

5.2. Performance Analysis

Through a series of experiments, a comparison and analysis of the security and robustness of PBA are conducted. The model performance of PBA is evaluated concerning ensuring security and robustness.

5.2.1. Security Evaluation

To evaluate the security of PBA, the DLG attack is implemented, and LeNet5 is used as the generator and model training network. Moreover, we take the MNIST dataset sample “7” as an example to observe whether the reconstructed image can be recognized. In this part, we simulate malicious attackers intercepting the model updates of the parties participating in the training and attempting to recover sensitive information. We take the FL method without any protection as the baseline and compare the reconstructed images of these two methods. Each attack result displays the reconstruction results of every ten rounds within 500 rounds.
Firstly, the DLG attack is conducted on the baseline, which is not perturbed. Although there is still some noise in the data sample in the 10th round, as shown in Figure 2a, it could already be recognized by the human eye. Hence, it can be indicated that information leakage occurs in this method. Subsequently, we conduct the DLG attack on PBA. When the gradient is perturbed in PBA, the DLG attack cannot reconstruct the user’s sensitive image, as shown in Figure 2b.
Moreover, the DLG attack is conducted on different degrees of perturbation, and the attack results are all unable to reconstruct the image. That is, the attack failed. The experimental results show that PBA is effective in resisting DLG attacks and can achieve the goal of protecting user privacy.

5.2.2. Robustness Evaluation

In the evaluation of robustness, we take the FL method without any protection as a baseline method to compare with PBA, Krum [27], trimmed-mean [28], and UBAR [36]. We, respectively, set Byzantine clients accounting for 0%, 10%, and 30% of the total, comparing, and analyzing global accuracy changes under Gaussian attacks and label flipping attacks.
To begin with, the performance without malicious clients and attacks is tested and compared with the presence of Byzantine clients. The results as shown in Figure 3. Compared with the baseline, the accuracy of PBA is close to other methods, about 88%. In the same way, the loss of PBA, about 0.4, is lower or similar to that of other methods. Moreover, the convergence speed of PBA is almost the same as the baseline.
After that, the performance of PBA is compared with that of Krum [27], trimmed-mean [28], and UBAR [36] with the different number of Byzantine clients. Faced with the same attack, the degree to which model performance is affected by Byzantine clients is related to the number of malicious clients in the system.
We take GA to simulate a fraction of clients uploading bad data. From Figure 4, the model accuracy of all methods with f = 6 Byzantine clients in the system is lower than that with f = 2 . When there are few malicious clients, the global accuracy of PBA is better than Krum [27] and UBAR [36], almost equal to the accuracy without Byzantine clients. But when malicious clients account for 30%, the performance of PBA and trimmed-mean [28] decreases significantly. On the contrary, the performance of UBAR [36] and Krum [27] can maintain more stability. Moreover, we conduct LFA to evaluate the model performance. As shown in Figure 5, although LFA is more severe than GA, the model performance of all methods is almost not affected with few Byzantine clients in the system, the accuracy of about 85%. However, if there are 60% malicious clients in the system, the performance of methods, except UBAR [36], is reduced remarkably. Krum [27] and trimmed-mean [28] can not achieve convergence. This means Byzantine clients reach the goal of disrupting the model. Although PBA and UBAR [36] can achieve convergence, their accuracy is influenced significantly. Hence, PBA is unsuitable for the system with many Byzantine clients.
Based on the above experiments, the performance of PBA is close to the baseline without Byzantine clients and attacks. However, if there are many malicious clients in the system and the attacks are strong, the performance of PBA is greatly compromised. Although PBA did not always perform well in experiments, it could reduce the time cost compared to other methods.

5.2.3. Cost Evaluation

Wang et al. [66] proposed a method called Brief that calculates cosine similarity using binary operations and then implements B2A to calculate the Euclidean distance further. Similarly, PBA uses binary sequences to calculate the Euclidean distance, somewhat reducing the time cost. Hence, we measure the time spent by different methods in an epoch. And the presentation is the average time required for five epochs. The results are shown in Table 2.
Because of B2A and clustering used in Brief [66], it needs a long time to finish a training epoch, about 57 s. Moreover, to achieve high accuracy, UBAR [36] reuses training samples to test the performance, which takes some time. The time spent by other distance-based methods is not significantly different. For PBA, it is based on approximate Euclidean distance and 3 σ criterion. PBA, computing approximate Euclidean distance by binary sequences, has lower complexity than other distance-based methods. Compared with the clustering method in Brief [66], the 3 σ criterion with lower complexity is scientific.

6. Conclusions

In order to solve the issue that previous federated learning methods cannot quickly execute outlier detection, this paper proposes a method to estimate the Euclidean distance using binary sequences that reflects the differences between gradients. Based on this approach, PBA, a 3 σ criterion-based federated learning security robust method, is proposed. Moreover, this paper analyzes the correctness and complexity of PBA and evaluates the performance of PBA on open datasets. The results show that its privacy security can be guaranteed and Byzantine robustness can be achieved in some conditions. PBA offers good convergence performance and somewhat lowers the computational cost as compared to other approaches. However, faced with severe attacks or many Byzantine clients in the system, PBA can maintain stable performance.
We may try to deploy this method to multi-server scenarios that require a shorter time to train the FL model. In addition, in the future, we hope to further study how to better integrate FL with the IoT, which can protect privacy, achieve Byzantine resilience, and be more lightweight.

Author Contributions

Conceptualization, Q.L. and X.W.; methodology, X.W.; validation, Q.L., X.W., and S.R.; formal analysis, X.W.; writing—original draft preparation, X.W.; writing—review and editing, Q.L. and S.R.; supervision, Q.L.; project administration, S.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Research on Key Technologies for Greenhouse Gas Emission Reduction and Low Carbon Farming in Poultry Farming (BE2022310).

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable for this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Khan, L.U.; Yaqoob, I.; Imran, M.; Han, Z.; Hong, C.S. 6G wireless systems: A vision, architectural elements, and future directions. IEEE Access 2020, 8, 147029–147044. [Google Scholar] [CrossRef]
  2. Qi, K.; Yang, C. Popularity prediction with federated learning for proactive caching at wireless edge. In Proceedings of the 2020 IEEE Wireless Communications and Networking Conference (WCNC), Seoul, Republic of Korea, 25–28 May 2020; IEEE: Toulouse, France, 2020; pp. 1–6. [Google Scholar]
  3. Khan, L.U.; Saad, W.; Han, Z.; Hossain, E.; Hong, C.S. Federated learning for internet of things: Recent advances, taxonomy, and open challenges. IEEE Commun. Surv. Tutor. 2021, 23, 1759–1799. [Google Scholar] [CrossRef]
  4. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial Intelligence and Statistics, PMLR, Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  5. Nguyen, D.C.; Ding, M.; Pathirana, P.N.; Seneviratne, A.; Li, J.; Poor, H.V. Federated learning for internet of things: A comprehensive survey. IEEE Commun. Surv. Tutor. 2021, 23, 1622–1658. [Google Scholar] [CrossRef]
  6. Liu, Z.; Guo, J.; Yang, W.; Fan, J.; Lam, K.Y.; Zhao, J. Privacy-preserving aggregation in federated learning: A survey. In IEEE Trans. Big Data; IEEE: Toulouse, France, 2022. [Google Scholar]
  7. Konečnỳ, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated learning: Strategies for improving communication efficiency. arXiv 2016, arXiv:1610.05492. [Google Scholar]
  8. Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; Cummings, R.; et al. Advances and open problems in federated learning. Found. Trends Mach. Learn. 2021, 14, 1–210. [Google Scholar] [CrossRef]
  9. Hao, M.; Li, H.; Xu, G.; Chen, H.; Zhang, T. Efficient, Private and Robust Federated Learning. In Annual Computer Security Applications Conference; ACM: New York, NY, USA, 2021. [Google Scholar]
  10. Ma, X.; Gu, L. Research and Application of Generative-Adversarial-Network Attacks Defense Method Based on Federated Learning. Electronics 2023, 12, 975. [Google Scholar] [CrossRef]
  11. Wang, K.C.; Fu, Y.; Li, K.; Khisti, A.; Zemel, R.; Makhzani, A. Variational Model Inversion Attacks. Adv. Neural Inf. Process. Syst. 2021, 34, 9706–9719. [Google Scholar]
  12. Zhao, Y.; Zhao, J.; Jiang, L.; Tan, R.; Niyato, D.; Li, Z.; Lyu, L.; Liu, Y. Privacy-Preserving Blockchain-Based Federated Learning for IoT Devices. IEEE Internet Things J. 2021, 8, 1817–1829. [Google Scholar] [CrossRef]
  13. Zhu, L.; Liu, Z.; Han, S. Deep leakage from gradients. Adv. Neural Inf. Process. Syst. 2019, 32, 14774–14784. [Google Scholar]
  14. Yang, Z.; Zhang, J.; Chang, E.C.; Liang, Z. Neural network inversion in adversarial setting via background knowledge alignment. In Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, London, UK, 11–15 November 2019; ACM: New York, NY, USA, 2019; pp. 225–240. [Google Scholar]
  15. Lin, L.; Zhang, X. PPVerifier: A Privacy-Preserving and Verifiable Federated Learning Method in Cloud-Edge Collaborative Computing Environment. IEEE Internet Things J. 2022, 10, 8878–8892. [Google Scholar] [CrossRef]
  16. Gao, W.; Guo, S.; Zhang, T.; Qiu, H.; Wen, Y.; Liu, Y. Privacy-preserving collaborative learning with automatic transformation search. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; IEEE: Toulouse, France, 2021; pp. 114–123. [Google Scholar]
  17. Geiping, J.; Bauermeister, H.; Dröge, H.; Moeller, M. Inverting gradients-how easy is it to break privacy in federated learning? Adv. Neural Inf. Process. Syst. 2020, 33, 16937–16947. [Google Scholar]
  18. Baruch, G.; Baruch, M.; Goldberg, Y. A little is enough: Circumventing defenses for distributed learning. Adv. Neural Inf. Process. Syst. 2019, 32, 8635–8645. [Google Scholar]
  19. Bagdasaryan, E.; Veit, A.; Hua, Y.; Estrin, D.; Shmatikov, V. How to backdoor federated learning. In Proceedings of the International Conference on Artificial Intelligence and Statistics, PMLR, Online, 26–28 August 2020; pp. 2938–2948. [Google Scholar]
  20. Islam, A.; Al Amin, A.; Shin, S.Y. FBI: A federated learning-based blockchain-embedded data accumulation scheme using drones for Internet of Things. IEEE Wirel. Commun. Lett. 2022, 11, 972–976. [Google Scholar] [CrossRef]
  21. Liu, J.; Huang, J.; Zhou, Y.; Li, X.; Ji, S.; Xiong, H.; Dou, D. From distributed machine learning to federated learning: A survey. Knowl. Inf. Syst. 2022, 64, 885–917. [Google Scholar] [CrossRef]
  22. Briggs, C.; Fan, Z.; Andras, P. A review of privacy-preserving federated learning for the Internet-of-Things. In Federated Learning Systems: Towards Next-Generation AI; Springer: Cham, Switzerland, 2021; pp. 21–50. [Google Scholar]
  23. Pathak, M.; Rane, S.; Raj, B. Multiparty differential privacy via aggregation of locally trained classifiers. Adv. Neural Inf. Process. Syst. 2010, 23, 1876–1884. [Google Scholar]
  24. Zhao, L.; Wang, Q.; Zou, Q.; Zhang, Y.; Chen, Y. Privacy-preserving collaborative deep learning with unreliable participants. IEEE Trans. Inf. Forensics Secur. 2019, 15, 1486–1500. [Google Scholar] [CrossRef] [Green Version]
  25. Bonawitz, K.; Ivanov, V.; Kreuter, B.; Marcedone, A.; McMahan, H.B.; Patel, S.; Ramage, D.; Segal, A.; Seth, K. Practical secure aggregation for privacy-preserving machine learning. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017; ACM: New York, NY, USA, 2017; pp. 1175–1191. [Google Scholar]
  26. Shokri, R.; Shmatikov, V. Privacy-preserving deep learning. In Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security, Denver, CO, USA, 12–16 October 2015; ACM: New York, NY, USA, 2015; pp. 1310–1321. [Google Scholar]
  27. Blanchard, P.; El Mhamdi, E.M.; Guerraoui, R.; Stainer, J. Machine learning with adversaries: Byzantine tolerant gradient descent. Adv. Neural Inf. Process. Syst. 2017, 30, 118–128. [Google Scholar]
  28. Yin, D.; Chen, Y.; Kannan, R.; Bartlett, P. Byzantine-robust distributed learning: Towards optimal statistical rates. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 5650–5659. [Google Scholar]
  29. Yang, Z.; Bajwa, W.U. ByRDiE: Byzantine-resilient distributed coordinate descent for decentralized learning. IEEE Trans. Signal Inf. Process. Over Netw. 2019, 5, 611–627. [Google Scholar] [CrossRef] [Green Version]
  30. Fang, C.; Yang, Z.; Bajwa, W.U. BRIDGE: Byzantine-resilient decentralized gradient descent. IEEE Trans. Signal Inf. Process. Over Netw. 2022, 8, 610–626. [Google Scholar] [CrossRef]
  31. Guerraoui, R.; Rouault, S. The hidden vulnerability of distributed learning in byzantium. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 3521–3530. [Google Scholar]
  32. Li, T.; Hu, S.; Beirami, A.; Smith, V. Ditto: Fair and robust federated learning through personalization. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 6357–6368. [Google Scholar]
  33. Lin, J.; Du, M.; Liu, J. Free-riders in federated learning: Attacks and defenses. arXiv 2019, arXiv:1911.12560. [Google Scholar]
  34. Tolpegin, V.; Truex, S.; Gursoy, M.E.; Liu, L. Data poisoning attacks against federated learning systems. In Proceedings of the Computer Security—ESORICS 2020: 25th European Symposium on Research in Computer Security, ESORICS 2020, Guildford, UK, 14–18 September 2020; Proceedings, Part I 25. Springer: Berlin/Heidelberg, Germany, 2020; pp. 480–501. [Google Scholar]
  35. So, J.; Güler, B.; Avestimehr, A.S. Byzantine-resilient secure federated learning. IEEE J. Sel. Areas Commun. 2020, 39, 2168–2181. [Google Scholar] [CrossRef]
  36. Guo, S.; Zhang, T.; Yu, H.; Xie, X.; Ma, L.; Xiang, T.; Liu, Y. Byzantine-resilient decentralized stochastic gradient descent. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 4096–4106. [Google Scholar] [CrossRef]
  37. Miao, Y.; Liu, Z.; Li, H.; Choo, K.K.R.; Deng, R.H. Privacy-preserving Byzantine-robust federated learning via blockchain systems. IEEE Trans. Inf. Forensics Secur. 2022, 17, 2848–2861. [Google Scholar] [CrossRef]
  38. Zhang, C.; Xie, Y.; Bai, H.; Yu, B.; Li, W.; Gao, Y. A survey on federated learning. Knowl.-Based Syst. 2021, 216, 106775. [Google Scholar] [CrossRef]
  39. Yin, X.; Zhu, Y.; Hu, J. A comprehensive survey of privacy-preserving federated learning: A taxonomy, review, and future directions. ACM Comput. Surv. 2021, 54, 1–36. [Google Scholar] [CrossRef]
  40. Xie, C.; Koyejo, S.; Gupta, I. Zeno: Distributed stochastic gradient descent with suspicion-based fault-tolerance. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6893–6901. [Google Scholar]
  41. Chen, M.; Shlezinger, N.; Poor, H.V.; Eldar, Y.C.; Cui, S. Communication-efficient federated learning. Proc. Natl. Acad. Sci. USA 2021, 118, e2024789118. [Google Scholar] [CrossRef]
  42. Luping, W.; Wei, W.; Bo, L. CMFL: Mitigating communication overhead for federated learning. In Proceedings of the 2019 IEEE 39th International Conference on Distributed Computing Systems (ICDCS), Dallas, TX, USA, 7–10 July 2019; IEEE: Toulouse, France, 2019; pp. 954–964. [Google Scholar]
  43. Hamer, J.; Mohri, M.; Suresh, A.T. Fedboost: A communication-efficient algorithm for federated learning. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 3973–3983. [Google Scholar]
  44. Hao, M.; Li, H.; Luo, X.; Xu, G.; Yang, H.; Liu, S. Efficient and privacy-enhanced federated learning for industrial artificial intelligence. IEEE Trans. Ind. Inform. 2019, 16, 6532–6542. [Google Scholar] [CrossRef]
  45. Xu, R.; Baracaldo, N.; Zhou, Y.; Anwar, A.; Ludwig, H. Hybridalpha: An efficient approach for privacy-preserving federated learning. In Proceedings of the 12th ACM Workshop on Artificial Intelligence and Security, London, UK, 15 November 2019; ACM: New York, NY, USA, 2019; pp. 13–23. [Google Scholar]
  46. Truex, S.; Liu, L.; Chow, K.H.; Gursoy, M.E.; Wei, W. LDP-Fed: Federated learning with local differential privacy. In Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking, Heraklion, Greece, 27 April 2020; ACM: New York, NY, USA, 2020; pp. 61–66. [Google Scholar]
  47. Yu, H.; Liu, Z.; Liu, Y.; Chen, T.; Cong, M.; Weng, X.; Niyato, D.T.; Yang, Q. A Fairness-aware Incentive Scheme for Federated Learning. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society; ACM: New York, NY, USA, 2020. [Google Scholar]
  48. Zhan, Y.; Li, P.; Qu, Z.; Zeng, D.; Guo, S. A learning-based incentive mechanism for federated learning. IEEE Internet Things J. 2020, 7, 6360–6368. [Google Scholar] [CrossRef]
  49. Rivest, R.L.; Adleman, L.; Dertouzos, M.L. On data banks and privacy homomorphisms. Found. Secur. Comput. 1978, 4, 169–180. [Google Scholar]
  50. Acar, A.; Aksu, H.; Uluagac, A.S.; Conti, M. A survey on homomorphic encryption schemes: Theory and implementation. ACM Comput. Surv. 2018, 51, 1–35. [Google Scholar] [CrossRef]
  51. Shokri, R.; Stronati, M.; Song, C.; Shmatikov, V. Membership inference attacks against machine learning models. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017; IEEE: Toulouse, France, 2017; pp. 3–18. [Google Scholar]
  52. Liu, X.; Li, H.; Xu, G.; Chen, Z.; Huang, X.; Lu, R. Privacy-enhanced federated learning against poisoning adversaries. IEEE Trans. Inf. Forensics Secur. 2021, 16, 4574–4588. [Google Scholar] [CrossRef]
  53. Xu, P.; Hu, M.; Chen, T.; Wang, W.; Jin, H. LaF: Lattice-based and communication-efficient federated learning. IEEE Trans. Inf. Forensics Secur. 2022, 17, 2483–2496. [Google Scholar] [CrossRef]
  54. Beimel, A. Secret-sharing schemes: A survey. In Proceedings of the Coding and Cryptology: Third International Workshop, IWCC 2011, Qingdao, China, 30 May–3 June 2011; Proceedings 3. Springer: Berlin/Heidelberg, Germany, 2011; pp. 11–46. [Google Scholar]
  55. Shamir, A. How to share a secret. Commun. ACM 1979, 22, 612–613. [Google Scholar] [CrossRef]
  56. Dwork, C. Differential privacy. In Proceedings of the Automata, Languages and Programming: 33rd International Colloquium, ICALP 2006, Venice, Italy, 10–14 July 2006; Proceedings, Part II 33. Springer: Berlin/Heidelberg, Germany, 2006; pp. 1–12. [Google Scholar]
  57. El Ouadrhiri, A.; Abdelhadi, A. Differential privacy for deep and federated learning: A survey. IEEE Access 2022, 10, 22359–22380. [Google Scholar] [CrossRef]
  58. Zhao, Y.; Zhao, J.; Yang, M.; Wang, T.; Wang, N.; Lyu, L.; Niyato, D.; Lam, K.Y. Local differential privacy-based federated learning for internet of things. IEEE Internet Things J. 2020, 8, 8836–8853. [Google Scholar] [CrossRef]
  59. Wei, K.; Li, J.; Ding, M.; Ma, C.; Yang, H.H.; Farokhi, F.; Jin, S.; Quek, T.Q.; Poor, H.V. Federated learning with differential privacy: Algorithms and performance analysis. IEEE Trans. Inf. Forensics Secur. 2020, 15, 3454–3469. [Google Scholar] [CrossRef] [Green Version]
  60. Ayadi, H.; Zouinkhi, A.; Boussaid, B.; Abdelkrim, M.N. A machine learning methods: Outlier detection in wsn. In Proceedings of the 2015 16th International Conference on Sciences and Techniques of Automatic Control and Computer Engineering (STA), Monastir, Tunisia, 21–23 December 2015; IEEE: Toulouse, France, 2015; pp. 722–727. [Google Scholar]
  61. Pukelsheim, F. The three sigma rule. AM STAT 1994, 48, 88–91. [Google Scholar]
  62. Curtis, A.E.; Smith, T.A.; Ziganshin, B.A.; Elefteriades, J.A. The mystery of the Z-score. Aorta 2016, 4, 124–130. [Google Scholar] [CrossRef]
  63. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. Density-based spatial clustering of applications with noise. Int. Conf. Knowl. Discov. Data Min. 1996, 96, 226–231. [Google Scholar]
  64. Nasr, M.; Shokri, R. Improving deep learning with differential privacy using gradient encoding and denoising. arXiv 2020, arXiv:2007.11524. [Google Scholar]
  65. Li, S.; Ngai, E.; Voigt, T. Byzantine-robust aggregation in federated learning empowered industrial iot. IEEE Trans. Ind. Inform. 2021, 19, 1165–1175. [Google Scholar] [CrossRef]
  66. Wang, R.; Wang, X.; Chen, H.; Picek, S.; Liu, Z.; Liang, K. BRIEF but Powerful: Byzantine-Robust and Privacy-Preserving Federated Learning via Model Segmentation and Secure clustering. arXiv 2022, arXiv:2208.10161. [Google Scholar]
  67. Zhai, K.; Ren, Q.; Wang, J.; Yan, C. Byzantine-robust federated learning via credibility assessment on non-IID data. arXiv 2021, arXiv:2109.02396. [Google Scholar] [CrossRef] [PubMed]
  68. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8024–8035. [Google Scholar]
  69. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
Figure 1. The overall framework of PBA.
Figure 1. The overall framework of PBA.
Electronics 12 02951 g001
Figure 2. Reconstructed images under the DLG. (a) Baseline. (b) PBA.
Figure 2. Reconstructed images under the DLG. (a) Baseline. (b) PBA.
Electronics 12 02951 g002
Figure 3. Convergence performance without malicious clients and attacks. (a) Global accuracy. (b) Cross entropy.
Figure 3. Convergence performance without malicious clients and attacks. (a) Global accuracy. (b) Cross entropy.
Electronics 12 02951 g003
Figure 4. Global accuracy under GA with different numbers of malicious clients. (a) f = 2 . (b) f = 6 .
Figure 4. Global accuracy under GA with different numbers of malicious clients. (a) f = 2 . (b) f = 6 .
Electronics 12 02951 g004
Figure 5. Global accuracy under LFA with different numbers of malicious clients. (a) f = 2 . (b) f = 6 .
Figure 5. Global accuracy under LFA with different numbers of malicious clients. (a) f = 2 . (b) f = 6 .
Electronics 12 02951 g005
Table 1. Notations.
Table 1. Notations.
NotationDescription
Cthe number of workers
Sthe number of servers
fthe number of malicious workers
Tthe number of epochs
gmodel update (gradient) 1
xglobal model
g [ k ] the k-th value of the gradient
rrandom vector
S sign matrix
D data matrix
μ mean value
σ standard deviation
η learning rate
1 “Global model” and “gradient” are used interchangeably.
Table 2. Running time of different methods.
Table 2. Running time of different methods.
MethodTime
Krum [27]10.640 s
trimmed-mean [28]17.588 s
UBAR [36]23.030 s
Brief [66]56.659 s
PBA9.391 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Q.; Wang, X.; Ren, S. A Privacy Robust Aggregation Method Based on Federated Learning in the IoT. Electronics 2023, 12, 2951. https://doi.org/10.3390/electronics12132951

AMA Style

Li Q, Wang X, Ren S. A Privacy Robust Aggregation Method Based on Federated Learning in the IoT. Electronics. 2023; 12(13):2951. https://doi.org/10.3390/electronics12132951

Chicago/Turabian Style

Li, Qingtie, Xuemei Wang, and Shougang Ren. 2023. "A Privacy Robust Aggregation Method Based on Federated Learning in the IoT" Electronics 12, no. 13: 2951. https://doi.org/10.3390/electronics12132951

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop