Next Article in Journal
Dynamic Pricing with Parametric Demand Learning and Reference-Price Effects
Next Article in Special Issue
Enhancing Brain Tumor Segmentation Accuracy through Scalable Federated Learning with Advanced Data Privacy and Security Measures
Previous Article in Journal
Explicit Properties of Apostol-Type Frobenius–Euler Polynomials Involving q-Trigonometric Functions with Applications in Computer Modeling
Previous Article in Special Issue
Federated Learning for the Internet-of-Medical-Things: A Survey
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FedISM: Enhancing Data Imbalance via Shared Model in Federated Learning

Department of Information and Computer Engineering, Chung Yuan Christian University, Taoyuan 320, Taiwan
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(10), 2385; https://doi.org/10.3390/math11102385
Submission received: 15 April 2023 / Revised: 6 May 2023 / Accepted: 18 May 2023 / Published: 20 May 2023
(This article belongs to the Special Issue New Trends in Machine Learning and Medical Imaging and Applications)

Abstract

:
Considering the sensitivity of data in medical scenarios, federated learning (FL) is suitable for applications that require data privacy. Medical personnel can use the FL framework for machine learning to assist in analyzing large-scale data that are protected within the institution. However, not all clients have the same distribution of datasets, so data imbalance problems occur among clients. The main challenge is to overcome the performance degradation caused by low accuracy and the inability to converge the model. This paper proposes a FedISM method to enhance performance in the case of Non-Independent Identically Distribution (Non-IID). FedISM exploits a shared model trained on a candidate dataset before performing FL among clients. The Candidate Selection Mechanism (CSM) was proposed to effectively select the most suitable candidate among clients for training the shared model. Based on the proposed approaches, FedISM not only trains the shared model without sharing any raw data, but it also provides an optimal solution through the selection of the best shared model. To evaluate performance, the proposed FedISM was applied to classify coronavirus disease (COVID), pneumonia, normal, and viral pneumonia in the experiments. The Dirichlet process was also used to simulate a variety of imbalanced data distributions. Experimental results show that FedISM improves accuracy by up to 25% when privacy concerns regarding patient data are rising among medical institutions.

1. Introduction

With the impact of the recent coronavirus disease 2019 (COVID-19) pandemic, there is an urgent need for rapid testing and screening to confirm whether individuals are infected. However, the large number of screening samples has overwhelmed the manpower of medical institutions. To speed up disease detection, alternative methods such as quick screening reagents are being adopted. The high number of COVID-19 patients worldwide is causing a considerable burden on the global healthcare system. Consequently, many studies have been conducted on medical and healthcare applications using deep learning (DL) for coronavirus detection. For instance, COVID-Net [1] developed a new Deep Neural Network (DNN) model to detect coronavirus from lung X-ray images. In another study [2], ImageNet [3] was used as a pre-trained model for transfer learning. The model was further combined with COVID-Net to train chest computed tomography (CT) images and investigate the performance effect of initial model parameters. However, resource-deficient institutions may face inefficiencies due to an insufficient quantity of data.
To train the model effectively, medical data are usually uploaded to a server. Traditional DL approaches rely on centralized learning (CL), where all data are collected on a server for storage and training. However, in practical scenarios, medical institutions face privacy issues when using CL technologies. The collected datasets may contain sensitive data, such as X-rays, facial data, or disease history. During the training process, data may be accidentally leaked, and sensitive information may indirectly identify user characteristics. Therefore, privacy concerns regarding patient data are rising among medical institutions.
To address privacy concerns in CL, the federated learning (FL) approach has been proposed in recent years. FL is an emerging collaborative framework for distributed machine learning, as shown in Figure 1. The FedAvg [4] algorithm, represented in Equation (1), obtains the global model wt by averaging the local model wit of each institution i. The standard optimization formulation in FL is shown in Equation (2), where F() is the loss function and wit represents the model of each institution i. The ultimate objective is to aggregate a model that can fit the data of each institution and minimize the sum of the obtained loss values.
w t = i I ω i t I
m i n i I F i ω i t
Unlike CL, the FL training process does not require pooling raw data to a centralized server or exchanging raw data between clients. In other words, medical institutions do not have to send out their patient data, which alleviates privacy issues. Instead, clients exchange only model parameters and metadata during the training process. Resource-deficient institutions can also benefit from FL aggregation by exchanging learning features with other institutions. Each institution may learn different features from different datasets in other institutions, resulting in improved overall performance. Finally, medical personnel from each institution can assess the results through the model inference.
Sharing data between medical institutions is a sensitive matter, especially when the data contain private patient information. During the transmission process, the data may be vulnerable to network attacks or leaks [5]. Therefore, regulations such as General Data Protection Regulation (GDPR) [6] have been enacted to protect the personal data of European citizens. Collecting user data in a proper and lawful manner is essential to obtain user consent.
Without sending data to the server, the performance of FL highly depends on the completeness of data in each institution. Due to data privacy, the server cannot access the raw data of the institution, which brings many challenges in FL [4], such as Non-Independent Identically Distribution (Non-IID), imbalanced dataset distribution, massively distributed participants, and limited communication rounds. Among these challenges, data imbalance is the key factor affecting the performance of the training model. Some approaches have also been proposed to improve the FedAvg approach in classical FL. FedProx [7] adapted an optimizer to make the model gradient updates smaller. FedNova [8] referred to the training step and the quantity of data needed to improve the aggregator. Federated Stochastic Gradient Descent COVID-19 Detection (FedSGDCOVID) [9] combined FedAvg with the Differential Privacy Stochastic Gradient Descent (DP-SGD) [10] algorithm to improve the FL security by adding randomized Gaussian noise to the local gradient during the aggregation step. DataSilos [11] integrated the above approaches to conduct the experiment based on the same benchmark. However, none of the state-of-the-art approaches have particularly outstanding performance in the Non-IID environment. Another experiment [12] conducted on five advanced algorithms also concluded that Non-IID has a significant impact on FL.
On the other hand, data imbalance problems often occur in healthcare scenarios. Various frameworks for distributed DL were proposed in previous works, such as Institutional Incremental Learning (IIL) and Cyclic Institutional Incremental Learning (CIIL) [13,14]. Experiments were conducted on these frameworks to explore the catastrophic forgetting problem [15] in FL. In other medical applications, FL aids in learning feature extraction for depression [16]. The authors proposed a face detection and speech emotional classifier using a Convolutional Neural Network (CNN) combined with a Support Vector Machine (SVM). FL calculates a patient sentiment factor based on the label probability to provide advice for professionals to make a complete assessment. In contrast, the authors [17] adopted an FL framework to build a breast density classification model using mammography data. The model inference for breast density is an important factor that directly affects the diagnosis. However, their work lacked data heterogeneity and may be inefficient in the Non-IID environment.
The shared data approach [18] can effectively alleviate data imbalance by allowing the server to collect shared data from institutions for data expansion during training. However, in medical scenarios, exchanging raw data is discouraged due to privacy concerns. Therefore, this paper proposes a novel shared model approach for FL called FedISM, which enhances data imbalance without requiring the exchange of shared data. FedISM not only alleviates the problem of data imbalance distribution without exchanging shared data, but it also improves test accuracy with a shared model. Additionally, FL lacks an appropriate factor to assess the completeness of a dataset. It is imperative to delve into the realm of mathematics to determine which factors are appropriate for evaluating datasets in FL. However, some evaluation factors, such as computational power, have been studied in FedCS [19]. The main goal of FedCS is to explore ways of completing training within a specified time frame in a heterogeneous computing environment. Participants with large datasets may give up due to the long computation time. Therefore, the field of federated learning currently lacks a solution to efficiently evaluate datasets among participants. Accordingly, a Candidate Selection Mechanism (CSM) is proposed to select the best choice for a shared model. The goal is to select the best institution and share its local model without relying on shared data.
An X-ray dataset was used in our experiments to simulate medical institutions for coronavirus detection. The Dirichlet process was also used to simulate imbalanced distributions among clients. The experimental results showed that the shared model improves test accuracy in the Non-IID environment. Using only 5% of the shared data to train the shared model could improve accuracy by up to 25%. Moreover, the accuracy of the model after FL training was higher than that of the shared model, demonstrating the effectiveness of the shared model in addressing the challenge of data imbalance. Additionally, a Balanced CSM was proposed to assess the important dataset of each institution. Without peeking at the raw data, Balanced CSM selected institutions with more balanced data to train the shared model. Results proved that Balanced CSM could find the best choice for the shared model. Consequently, FedISM improved accuracy by 6% in the imbalanced distribution of the Dirichlet process.
Overall, the main contributions of this paper are highlighted as follows:
  • Exploring a shared model for effectively alleviating the challenge of data imbalance in FL;
  • Designing data assessment approaches and candidate selection mechanisms to find the best institution for training the shared model;
  • Applying FedISM for multi-class detection of COVID-19 and pneumonia in medical applications;
  • Evaluating the FedISM with various data distributions to illustrate that the shared model method significantly overcomes the data imbalanced problem and improves the test accuracy.
The remainder of this paper is organized as follows: in Section 2, related works on FL and Non-IID issues are discussed. Section 3 introduces the concept of a shared model and CSM as well as the proposed FedISM. Section 4 provides details on the experimental settings and conducts an analysis of various concepts and discussions of the experimental results, as well as the advantages and limitations in different scenarios. Finally, Section 5 presents conclusions and future directions for research.

2. Related Works

Many studies have applied DL techniques to medical image training for disease diagnosis and prediction. One example is the development of three predictive models for cancer detection [20], which demonstrates that machine learning can effectively diagnose diseases from images. Several surveys [21,22,23,24,25] comprehensively discuss and analyze machine learning models for medical image detection. Machine learning techniques have improved significantly, and the models are more tolerant of noisy data. However, the challenge with DL in medicine is the large amount of medical training data that need to be labeled, making it more difficult to implement. Another critical issue is the convergence problem. A study [26] investigated several DL research studies for health informatics. Due to the expensive costs, medical datasets are not easily available, resulting in imbalanced datasets. Classical DL models need to be implemented on balanced data distributions and often require additional synthetic data. However, this solution causes dependence on fabricated biological samples.
Table 1 highlights several issues related to FL that are worth discussing. Previous studies [27,28] have conducted experiments to compare the effectiveness of different DNN models in FL for detecting lung CT and X-ray images under Independent Identically Distribution (IID). Another study proposed an FL framework [29] for detecting lung X-ray images and compared the performance of Visual Geometry Group 16 (VGG16) and Residual Network 50 (ResNet50). The results showed that FL can achieve the same performance as CL in a five-fold cross-validation method. Furthermore, the capsule network-based classification model and a blockchain-based FL approach [30,31] were proposed to securely share data without being compromised. A normalization technique was applied to normalize heterogeneous chest CT images from different hospitals for more accurate training of FL models. To tackle the lack of training datasets, the blockchain-based FedGAN framework [32] was proposed to combine FL with the Generative Adversarial Network (GAN) technique [33]. This approach was also implemented in a distributed healthcare institution with edge cloud computing for coronavirus detection. The GAN generator was used to detect COVID-19 through the data enhancement method. However, generating medical data images using generators is not recommended for medical image generation [34], since no mathematical formula can prove their validity.
Regarding the challenge of limited communication in FL, the weights trained by all participating clients need to be transmitted to a central server, which increases communication costs. To improve communication costs in FL, the dynamic fusion approach [35] was proposed. This approach dynamically determines whether the client can participate in the fusion based on the performance of its local model. The work [36] proposed an efficient FL architecture for communication by pruning and quantizing the weights of the local model. The communication cost between the server and the client can then be reduced. Genetic Clustered Federated Learning (Genetic CFL) [37] adopted a genetic algorithm to optimize hyperparameters on cluster FL to achieve convergence efficiency.
Harmonia [38] is an FL system developed by Taiwan AI Lab for practical deployment. In many FL studies, a simulator is used to simulate the experiment in order to achieve a fast and stable environment. However, to make FL closer to reality, it must be able to run on a heterogeneous system to realize the application of FL in real scenarios. Unlike most simulators, Harmonia uses Kubernetes to encapsulate basic DL computation and aggregation in containers. The Operator container is responsible for system maintenance and communication with the Application container using Remote Procedure Calls (gRPC). The Application container is used for local training, and MNIST [39] is adopted as the default training dataset. During the training process, Gitea is used as the centralized storage service, and the FedAvg algorithm is applied by default for aggregation on the server.
Regarding the Non-IID problem, the survey [40] synthesized various investigations in FL. The Non-IID problem may arise from feature distribution skew, label distribution skew, same label but different features, same features but different labels, and quantity skew. All Non-IID data distributions may lead to model inefficiency. The shared data approach [18] initially addressed this problem by exchanging the shared data during the training process. Although the Non-IID problem can be effectively alleviated via shared data, exchanging raw data still violates the spirit of FL. Federated Learning with Shared Label Distribution (FedSLD) [41] proposed an FL approach with shared label distribution to alleviate the impact of Non-IID. Results showed that the accuracy of FedSLD can be improved by sharing label information.
The client-drift study [42] proposed a new optimizer specifically for FL. The model convergence was improved in the case of heterogeneous data and reduced the impact of client drift during aggregation. FedBN [43] proposed a new FL aggregation method based on FedAvg, which allowed the batch normalization (BN) layer to be updated only by the client. Computational resource requirements are reduced during training. To address client drift in a Non-IID environment, SCAFFOLD [44] improved the FedAvg algorithm by controlling the aggregation variables. However, in an experiment [12] that synthesized five FL architectures [4,7,8,43,44] for pneumonia detection and conducted them on different data distributions, the model accuracy in these approaches was much lower in data imbalance distribution than in IID distributions.
The previous studies were certainly effective, but most of them addressed different issues compared to our work. This paper enhances the following issues:
  • Multi-class classification. Previous studies conducted binary classification experiments. However, in practical scenarios, the model for multi-class classification should help medical institutions identify symptoms faster and at a lower cost;
  • Non-IID issue. The Non-IID problem is magnified in practical scenarios. The proposed FedISM investigates the Non-IID problem with a global optimization solution and alleviates the limitation of data imbalance from IID to various degrees of Non-IID distributions;
  • Non-IID experiment. Experiments in previous studies were mainly conducted in IID distributions. However, in practical scenarios, each organization has different data characteristics depending on factors such as geography and time, resulting in inefficient model performance. Thus, Non-IID performance should be a desirable criterion for evaluating the FL algorithm.
Among the numerous studies to solve the FL problem, the shared model approach has a better aggregation capability in extreme Non-IID environments. Compared to existing approaches, the shared model approach has experimentally proven to be effective in improving the Non-IID problem of FL.

3. FedISM Method

The proposed FedISM approach in this paper aims to address the challenges of data imbalance and data privacy in medical institutions by utilizing a shared model in FL. Additionally, a Candidate Selection Mechanism (CSM) is designed to effectively choose the most appropriate institution to train the shared model without compromising the privacy of the raw data.

3.1. Shared Model for FL

The work [45] investigated the diversity of aggregation directions for institutions in Non-IID distributions from the perspective of pre-trained models. The results showed that pre-trained models can make the aggregation directions of institutions more consistent. It was also proven that the expected convergence time satisfies Equation (3), where f is the institutional model, F represents the model update parameters, |I| is the number of institutions, E is the number of local epochs, ζ is the data heterogeneity of each institution, and T is the communication round. The complexity of the formula integrates the influence of both local training and communication rounds and determines the amplitude of gradient updates, ensuring the convergence performance of FL. Adjusting the direction of model updates can make the update direction more consistent, enhance the aggregation effect, and enable the aggregated model to adapt to the data of different institutions. Accordingly, using pre-trained models can alleviate the influence caused by heterogeneous data in different institutions.
E f ω T 2 O ( F T E I + F 2 / 3 + ζ 2 / 3   T 2 / 3 )
Zhao et al. [18] proposed a shared data approach, and the training process is illustrated in Figure 2a. Their work assumes that a portion of shared data resides on the server side. When the training process starts, each institution trains a local model with the shared data from the server and the local data residing on the institution side. In this way, institutions can expand their data to increase the amount of training data and alleviate the Non-IID problem. To this end, when an institution has only one data label, an expansion of the dataset on the server has all data labels. While applying the shared data scheme, each institution can access all the data from the server. Therefore, data expansion is beneficial for local training, and the server eventually aggregates a more efficient global model through the local models. However, transmitting the dataset violates the principle of avoiding raw data exchange in FL and also imposes additional burdens on the network.
In view of this, this paper proposes a shared model approach inspired by the concept of the shared data approach. This paper applies the concept of a pre-trained model with a shared model to achieve more efficient optimization. The training process of our proposed method is shown in Figure 2b. In the shared model approach, the network traffic during the training process can be reduced because shared data does not need to be exchanged in FL. Moreover, the training process is further modified using the FedProx approach [7] to simplify model gradient updates. Each institution trains its local model and aggregates the model before and after the local training. In the server, all local models, along with the previous round of global models, are collected for aggregation. Hence, the model gradient is updated by a small magnitude, as per the shared model. The detailed training processes are described as follows.
  • Initially, the server owns the shared data. In each round of training, the server first trains a global model using the shared data to obtain a shared model;
  • After training the shared model, the server sends the shared model as a global model to each institution;
  • Each institution performs local training to generate a local model, which is then aggregated with the global model;
  • Each institution sends its local model back to the server for aggregation;
  • The server aggregates local models from all institutions from the previous round of the global model.
The model update profile shown in Figure 3 highlights the significant differences that can arise for FL training when dealing with IID and Non-IID data distributions. In the case of IID, each institution has the same data distribution and data quantity, making it easy to find the global optimum during aggregation because the local update process of each institution does not differ too much. However, in the case of Non-IID data distribution, it is easy to get stuck in a local optimum because the data distribution and quantity of data in each institution are different. Some features can only be learned by specific institutions but can be forgotten due to different data distributions, leading to a decrease in FL model efficiency. Through the shared model, institutions can continue training based on the shared model and modify the algorithm to slow down the update process, resulting in better aggregation results for the global model.
The experiments presented in Section 4.1 demonstrate that the proposed algorithm effectively updates the shared model towards global optimization. The novel approach is able to achieve good performance even with a smaller portion of the whole dataset and even when institutions face the catastrophic forgetting problem after local training. The results indicate that using a shared model can improve accuracy in cases of an extremely imbalanced data distribution. With only 5% of shared data used for training, a 25% improvement in accuracy can be obtained. Higher accuracy can be achieved when more shared data are used to train the shared model.
However, training the shared model with shared data on the server may not be feasible in a practical medical scenario. Moreover, sharing data is difficult to achieve due to data sensitivity. These datasets may only be used for scientific research purposes and may not be openly accessible. In the next subsection, a CSM is proposed among institutions to find an alternative solution and replace the ideal shared data.

3.2. Candidate Selection Mechanism (CSM)

To address privacy concerns, this paper proposes a Candidate Selection Mechanism (CSM) to train a shared model using data from the best candidate institution. The assessment factor used in the CSM is initially determined by the sample size of the dataset and the number of labels at the local institution. Let |L| denote the total number of labels and Li denote the number of labels at a particular institution i. Let S denote the total data quantity across all institutions and Si denote the quantity of data at a particular institution i. The preliminary score is defined by the parameter β, which weighs the different factors. When β is smaller, the institution with more data has a higher score. When β is larger, the institution with more data labels has a higher score. Hence, the preliminary score (PScore) is calculated as follows.
P S c o r e = L i β + S i S 1 β
Figure 4 illustrates how the CSM can be used to select an appropriate institution for training the shared model. In the initialization process, as shown in Figure 4a, the server collects the data quantity (Si) from all the institutions. Next, the total data quantity (S) and the total number of data labels (|L|) are transmitted to each institution. Each institution calculates the assessment score according to Equation (4), which is initially determined by the sample size of the dataset and the number of labels residing locally. Once each institution calculates its assessment score, it sends the score to the server. In the candidate and training process, as shown in Figure 4b, the server notifies the institution with the highest assessment score to start the training first. The institution with the highest Pscore sends its local model to the remaining institutions. The remaining training process is the same as the training process in the shared model, as shown in Step 3~5 of Figure 2b. In this way, the CSM replaces the shared data collection, and the workload of the server can also be reduced while maintaining data privacy.
Section 4.2 presents detailed experiments that reveal that the accuracy of institutions is not proportional to their data quantity when an imbalanced data distribution occurs. Instead, institutions with complete data labels perform better. However, this experiment was only conducted on a small scale and did not consider remaining exceptions, such as imbalanced data distribution causing the Non-IID problem. The next subsection introduces more advanced factors that can be applied in the CSM to address this issue.

3.3. Balanced CSM

Since the CSM considers only monotonic factors, the impact scores for institutions cannot be accurately assessed. Therefore, a Balanced Candidate Selection Mechanism (Balanced CSM) was proposed to consider more advanced factors in assessing the data of an institution. Imbalanced data distribution has been shown to result in performance degradation [46]. Inconsistent sample size makes it difficult for the model to learn complete features. The data used to train the shared model in each institution should satisfy two conditions: (i) having more data labels and (ii) having more balanced data. The shared model is trained using the Balanced CSM to identify the institution with the most balanced data, leading to better performance for the shared model.
In this regard, label completeness should be considered a significant impact factor in accuracy performance. Let Si,l denote the data quantity of each label l in institution i. The sample size of each label in the institution is converted using Li. The converted score, Ci, is closer to the sum of the sample size when the institution has more complete labels. In other words, the score is calculated based on label completeness before summation. The institution that has more data labels reduces the score loss for conversion. Thus, the equation is calculated using Equation (5).
C i = l L i S i , l L i
Furthermore, the assessment score for each institution eventually considers the following critical factors:
  • Label completeness and data quantity. As shown in Equation (5), the score of data quantity is constrained by label completeness. The number of data labels at an institution determines its score for data quantity, as training a shared model requires relatively complete data and vice versa. If an institution has fewer data labels, its score for data quantity will be lower;
  • Local and overall standard deviation. The standard deviation indicates the degree of data imbalance. The more imbalanced the data distribution is, the larger the standard deviation becomes. Both local standard deviation (σi) and overall standard deviation (σall) are used to evaluate the factor of data distribution imbalance. When training the shared model with the same amount of data, it is preferable for the data distribution to be relatively balanced. The score is decreased if σi is larger than σall and vice versa;
  • Minimum sample size. Sparse sample size may result in inaccurate assessments. Equation (6) considers label completeness, but it does not account for the minimum sample size, represented by min (Si,l). In cases with sparse samples of labels, more scores are preserved during the conversion operation if there are more non-zero samples. Similarly, the standard deviation does not consider the average of the total sample size, which means that data distributions with high and low average sample sizes may have the same standard deviation. To address these limitations, min (Si,l) is needed as a factor in the assessment. When the min (Si,l) in an institution is more sparse, the score is lower and vice versa.
Therefore, each institution i considers these factors to calculate an expected score, Ei, according to Equation (6), and then it sends the result back to the server. The server selects the institution with the highest score as the candidate for training the shared model.
E i = C i min S i , l   s q r t σ i σ a l l
The shared model with Balanced CSM is incorporated into the FedISM algorithm (as shown in Algorithm 1), assuming that there are a total of I institutions participating in a round of training. All sets of labels are denoted as L, and the sum of all the sample sizes is denoted as S. Each institution i has a set of data labels denoted as Li, and the sample sizes are denoted as Si. The initialized procedures are presented in lines 2 to 9. First, the server initializes each institution i and obtains all label sets for the training process. After initialization, each institution i calculates its data quantity Si and standard deviation σi and sends them to the server. The server calculates the sum of Si to obtain S and the average of σi to obtain σall. After calculating S and σall, the server sends them to each institution. All institutions compute their expected scores Ei based on Equations (5) and (6) and send the score back to the server. The server selects the institution with the highest score as the candidate institution for training the shared model.
The candidate selection procedures are presented in lines 10 to 16. After receiving Ei, the server groups the institutions into normal and candidate sets. The institution with the highest Ei is designated as icandidate, and the remaining institutions are grouped in inormal. The icandidate represents the institution that has the most balanced data distribution and is responsible for training the shared model. The training procedures are presented in lines 17 to 24. In each round of training, the icandidate first performs LocalTraining() to obtain the local model. The local model trained by icandidate is taken as the global model and sent to institutions inormal that have not yet been trained. Each icandidate starts the LocalTraining() to get the local model and sends it to the server. After the server receives all the local models, the local model of each institution is aggregated with the previous round of the global model to obtain a new global model. In the next round, the global model is trained continuously by the same icandidate until the completion of the training rounds.
Algorithm 1: FedISM
Input: all institutions I , communication rounds T , local epochs E , learning rate η
Output: The final model ω Τ
1Server executes:
2   initial each institution i
3 L   Get all set of labels
4Get data quantity S i and standard deviation σ i from institutions i
5 S     i I S i
6 σ a l l     i I σ i I
7for each institution i I do
8 E i ExpectedScore ( S , L ,   σ a l l ,   i )
9end for
10for each institution i I do
11if  E i = = max ( E i I ) then
12   i c a n d i d a t e = i
13else
14 i n o r m a l .append ( i )
15end if
16end for
17for round t = 0 , 1 ,   ,   T 1 do
18 i c a n d i d a t e executes LocalTraining ( i , ω t ) and get local model ω Τ
19send the local model ω Τ to every i n o r m a l
20for  i i n o r m a l  do
21 ω i t     LocalTraining i , ω t
22end for
23 ω t + 1     i I S i S   ω k t + ω t / 2
24end for
25return  ω Τ
26ExpectedScore ( S ,   L ,   σ a l l ,   i ):
27Calculate C i   based on Equation (3)
28 σ i = s t d S i , l L
29Calculate E i based on Equation (4)
30return  E i
31LocalTraining i , ω t :
32 ω i t     ω t
33for epoch e = 1 ,   2 ,   , E do
34for each batch b = x , y of local data D i do
35 ω i t ω i t η L ω i t ; b
36 ω i t = ω i t + ω t / 2
37 return  ω i t to the server
To analyze the complexity of the FL algorithm, assuming that T represents the communication round, I represents the number of institutions, and L represents the time of local training. Therefore, the time complexity of traditional FedAvg is represented by Equation (7).
  O T × L × I
In FedISM, due to assessments of the completeness of a dataset, each institution calculates its own assessment score before the training process. Then, the server needs to select the institution with the highest score to prioritize the training of the shared model, followed by the other institutions. Therefore, the time complexity of FedISM is represented by Equation (8), where E represents the time to calculate the assessment score.
O I × E + I + T × L 1 + I 1 = O I × E + I + T × L × I
The time complexity of FedISM is slightly higher than that of FedAvg because FedISM requires an initialization process to calculate the scores. However, these differences are small because the calculation time of I × E + I is far less than T and L. Thus, the additional calculations can be ignored. The time complexity of FedISM approximates O (T × L × I).

4. Performance Evaluations

In practical medical scenarios, the classification problem often involves multiple classes rather than a purely binary classification. Therefore, the COVID Radiography Database [47,48] was used as our dataset to demonstrate the medical data in this paper. The dataset contains a total of 21,165 X-ray images, which are divided into four categories: COVID, lung opacity, normal, and viral pneumonia. The data distribution and some sample images are shown in Figure 5. The dataset is a compilation of publicly accessible databases, such as Chest X-Ray (CXR) [1] and Pneumonia [49]. Using this dataset provides developers with a variety of data to adapt the classification model.
For data pre-processing, the experiments take 80% of the dataset for each label as the training data and the remaining 20% as the test data. The DNN model applied in the experiments was the ResNet18 neural network [50], and the optimizer was SGD. The setting for weight_decay was 1 × 10−4, and the learning rate was adjusted to 1 × 10−3. CrossEntropyLoss was used as the loss function. The rest of the experimental configurations and data distributions are described in each subsection.

4.1. Shared Model Experiments

This paper proposes the concept of a shared model to alleviate the challenge of data imbalance without exchanging shared data. In this experiment, the number of institutions was set to four, which corresponds to the number of data labels. The data allocation principle first assigns the shared data according to the experiment ratio, while the rest of the data are private and assigned to different Non-IID configurations. In our experiments, the ratios of shared data were set at 5%, 10%, and 15% for evaluations. The private data were divided into Non-IID (1), Non-IID (2), Non-IID (3), and Non-IID (4), where Non-IID (n) represents the number of data labels owned by each institution. For example, Non-IID (1) means that each institution has only one label for its private data, while Non-IID (4) means that the sample sizes and data labels of the private data are equally distributed among all institutions. The simulation parameters were set to 100 rounds, 5 local epochs for FL, and 100 epochs for CL.
Figure 6 and Table 2 demonstrate the effectiveness of the shared model in different Non-IID distributions when compared to FedAvg and CL. The experimental results for CL are depicted in Figure 6, where the accuracy could achieve 69%, 76%, 79%, and 91% when the training data were applied at 5%, 10%, 15%, and 100%, respectively. Training with a smaller sample size results in a model that is unable to learn features completely. On the other hand, the effectiveness of the shared model with FedAvg is depicted in Table 2, where FedAvg and the shared model were trained using 5%, 10%, and 15% of the shared data in the case of Non-IID distributions. Experimental results show that the test accuracy of FedAvg was close to 90% in Non-IID (4). However, the accuracy decreased when the institutions had fewer labels. The accuracy dropped to 89%, 74%, and 52% in Non-IID (3), Non-IID (2), and Non-IID (1), respectively. When the degree of Non-IID was more severe, the local model learned more inconsistent features. The Non-IID problem dramatically decreases the performance in FL.
Table 2 also records the convergence time for each method to achieve a specified test accuracy of 80%. The results show that Non-IID has a significant impact on FedAvg under different data distributions. FedAvg could not reach the specified test accuracy within 100 rounds in some cases, such as Non-IID (1) and Non-IID (2) distributions. In the Non-IID (1) distribution, FedAvg performance was only 51.688%, making convergence impossible. Through the use of a shared model, convergence time could be effectively improved, which is particularly evident in Non-IID (1). Moreover, the convergence speed increased with the amount of data used to train the shared model.
This experiment yielded several important findings. The first is that the shared model trained by the shared data generally improved performance across different data distributions. In particular, the shared model was effective in addressing high-skew distributions with Non-IID (1) and Non-IID (2) configurations. For example, with a minimum of 5% shared data, the shared model improved the accuracy of Non-IID (1) and Non-IID (2) by 25% and 8%, respectively, compared to FedAvg. Although the performance improvements for Non-IID (3) and Non-IID (4) were less pronounced, the performance of the shared model was also comparable to that of CL when trained on 100% of the data. Notably, a 5% sharing model provided the largest performance improvement, while higher sharing rates (10% or more) tended to produce further improvements and be more stable.
Another notable finding from this experiment is the relationship between the effectiveness of the shared model and FL performance. The shared model could further improve the performance of FL. For example, when training with highly skewed Non-IID (1) using FedAvg, the accuracy was only 51%, while CL achieved 69% accuracy with 5% shared data. However, when training with a 5% shared model, the accuracy of Non-IID (1) was improved to 76%. Furthermore, if the shared model is properly trained, the aggregated model can move towards global optimization. Although institutions may still have a catastrophic forgetting problem after local training, the shared model could be a breakthrough for FL because applying 5% of shared data can significantly enhance test accuracy.
Obviously, training the shared model with shared data can achieve a significant effect. However, this assumption of existing shared data is unreasonable due to the importance of privacy issues in FL, which are magnified and examined in medical scenarios. Therefore, a CSM mechanism is desired to replace the shared data approach. In addition, experiments using highly skewed data distributions are conducted in this section and are suitable for face recognition applications [51]. Hence, alternative methods are needed to simulate various data distributions without sacrificing experimental completeness and matching reality.

4.2. CSM Experiments

The proposed CSM method enables training of the shared model without relying on the shared data stored on the server. Instead, CSM attempts to identify an institution whose data are representative of all participating institutions. The shared model is then trained using the data from this selected institution rather than the shared data. Since CSM cannot directly access the raw data from each institution, the selection process relies on accessing the metadata to make informed decisions.
While respecting data privacy constraints, the CSM method utilizes an assessment score that considers various factors to represent the importance of an institution’s data. Specifically, Equation (3) calculates the assessment score using two impact factors with different weights (β = 0.2, β = 0.5, β = 0.8) to reflect the amount and quality of the data in each institution. A smaller β value means that the institution has a larger amount of data and will receive a higher score. On the contrary, a higher β value indicates that the institution has a higher proportion of data labels and will also receive a higher score. The institution with the highest score is then selected as the candidate to train the shared model.
To account for various factors present in medical scenarios, such as disease type, sample size, and geographical location, the Dirichlet process [52] was employed to simulate probability distributions. The Dirichlet process is a conjugate prior distribution for multiple distributions, and the probability parameter α is adjusted to control the degree of data imbalance. The Dirichlet distribution is commonly used in the Bayesian generative process for data reconstruction. This process involves defining a prior distribution and specifying the hyperparameter α, where α represents the concentration parameter. Subsequently, random sampling from the distribution is performed based on α to satisfy the concentration parameter. When the value of α is small, the data distribution is more imbalanced, whereas a larger α value tends to result in a more balanced data distribution. For this experiment, the number of institutions was set to four, and the local epoch was set to five. Three different values of α (0.1, 0.5, and 1) and an IID scenario for balanced data distribution were used to simulate various data distribution scenarios. Figure 7 visualizes these data distributions, from extreme imbalance to balance, moving from left to right with α values of 0.1, 0.5, 1, and finally IID.
Table 3 presents the experimental results of CSM using different weights of β and the FedAvg method for various data distributions. Obviously, the accuracy of each method declines as the data imbalance becomes more severe. For instance, in FedAvg, the accuracy in the IID distribution could reach 90%, which is quite close to the results of CL with 100% data. However, the accuracy gradually dropped to 64% as α decreased, which is a 27% difference compared to CL. Among the CSM methods with different weights, CSM with β = 0.2 tends to choose the institution with the highest data quantity while CSM with β = 0.8 tends to select the institution with the most data labels. In a balanced data distribution, the CSM methods with different weights may pick the same institution and yield the same results.
Regarding the convergence time to achieve 85% test accuracy, FedAvg required more time to converge in various data distributions. Even FedAvg could not reach the specified test accuracy within 100 rounds in some cases. In contrast, the CSM method with β = 0.8 achieved a faster convergence time compared to other FL methods. In addition, the convergence speed was faster as α increases. Therefore, using the CSM method with β = 0.8 can lead to better convergence results in the same amount of time and improve efficiency.
In this experiment, not all institutions selected by CSM were better for training the shared model with the same performance. Specifically, the CSM with β = 0.2 performed worse than FedAvg in the imbalanced data distribution because data quantity is not dominant. For example, in the case of α = 0.1, Institution 0 had the highest data quantity, but most of the data were concentrated in the normal category, with only a small amount of data in the COVID-19 category. As a result, catastrophic forgetting of the model occurred during shared model training with imbalanced categories, leading to poorer training results.
On the other hand, the highest accuracy was achieved at CSM with β = 0.8. That is because the institution with more data labels achieved data balance, thereby eliminating the need for shared data and resulting in better training results for the shared model. However, the factors used in CSM are too simple, and none of these factors could sufficiently represent performance metrics. Thus, more advanced factors are required to assess the institutional data for further training of the shared model.

4.3. Balanced CSM Experiments

In order to improve the assessment process when there are more institutions involved, an extended version of CSM called Balanced CSM was developed. Balanced CSM incorporates additional impact factors to assess the best institution choice more accurately. These factors include label completeness and data quantity, local and overall standard deviation, and the minimum sample size. To test the effectiveness of Balanced CSM, the number of institutions in the experiment was increased to 10, with 100 rounds and 5 local epochs for FL. Dirichlet process with parameters α = 0.1, α = 0.5, and α = 1 was used to simulate different data distributions, and IID was used for balanced data distribution. The different data distributions are visualized in Figure 8.
Table 4 presents the experimental results. Similarly, the performance of FedAvg deteriorated as the data distribution became more imbalanced. In the IID case, FedAvg achieved an accuracy of 89%, which is only 2% worse than CL. However, in the case of imbalanced data distribution (α = 0.1), the accuracy of FedAvg was only 76%, which increases the gap with CL to 15%. The factors in CSM with β = 0.8 were not highly correlated with the data distribution, indicating that the data imbalance problem was prevalent in large scale institutions. Regarding the convergence time to achieve 75% test accuracy, CSM with β = 0.8 could not achieve the best solution. In the data distribution with α = 0.5, the convergence time was longer than any other method. However, in the improved Balanced CSM method, the selection mechanism was optimized to identify the best suitable institution for training so as to achieve better accuracy and convergence time.
Accordingly, the experimental results show that the accuracy of CSM with β = 0.8 was even worse than that of FedAvg at Dirichlet process with α = 0.5 or α = 1. In contrast, the accuracy of Balanced CSM was higher than the other methods. Balanced CSM utilizes three factors to identify the institution with the most balanced data, making it more effective in assessing the data for further training of the shared model.
In Figure 9, the data distribution is shown for the Dirichlet process with α = 0.5. Among all institutions, Institution 3 had the highest data quantity and all kinds of data labels. However, in this scenario, CSM with β = 0.8 selected Institution 3 to train the shared model, while the Balanced CSM selected Institution 1. Even though Institution 1 did not have the highest data quantity, it had the most balanced data among all institutions. In contrast, Institution 3 had a serious data imbalance problem and the highest standard deviation among all institutions. The large difference between the highest and lowest data quantities was 700 times, which led to an imbalance problem in the training process and resulted in low accuracy performance.
Table 5 shows the results of all institutions trained with the shared model in each distribution. Evaluation metrics use accuracy, recall, precision, and F1-score to assess different aspects of the inspection. The N-class classification (weighted average) was calculated as in Equations (9)–(12).
Accuracy = i = 0 n 1 T P i + T N i j = 0 n 1 T P j + T N j + F P j + F N j 100 %
Precision = i = 0 n 1 T P i y i y j = 0 n 1 T P j + F P j 100 %
Recall = i = 0 n 1 T P i y i y j = 0 n 1 T P j + F N j 100 %
F 1 - score = 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l 100 %
The results show that the institution selected by Balanced CSM had the best accuracy and higher performance in precision, recall, and F1-score. This means that Balanced CSM can improve learning in all categories and reduce the number of false positive and false negative cases. These factors in Balanced CSM extend the indicators of data quantity and label size, proving that Balanced CSM is more suitable for assessing institution data through experiments.
Overall, the factor in Balanced CSM aims to achieve a balanced characteristic. A balanced data distribution enables the model to learn the features of each label fairly. However, in an imbalanced data distribution, each institution trains a model with different training characteristics. As a result, Balanced CSM can find the most balanced data as a candidate to train the shared model. The shared model with the Balanced CSM can improve the accuracy of data distribution from imbalanced to balanced, particularly when the data is extremely imbalanced.

5. Conclusions and Future Works

Data imbalance is a common challenge in FL that has not been fully alleviated in previous works. This paper aimed to address this challenge by proposing the FedISM approach for COVID-19 detection. This scenario is more suitable for practical medical applications where raw data cannot be exchanged between medical institutions. By applying FedISM, data privacy and data imbalance problems between institutions can be alleviated. There are two main contributions of FedISM that increase feasibility in a practical medical scenario. First, the shared model approach enhances the shared data approach by not exchanging raw data during the training process, while also modifying the algorithm to achieve higher accuracy with smaller gradient updates. Second, while shared data is ideal, it is not practical for FL applications. Instead, the shared model can be trained by the institution as an alternative way. To achieve this, a CSM was proposed to calculate an assessment score for the dataset of each institution.
The experimental results show that FedISM can efficiently alleviate data imbalance issue by identifying the most suitable institution for training the shared model. A significant improvement was achieved when training the shared model using only 5% of the shared data. In the highly skewed distribution of Non-IID (1), FedISM improved accuracy by 25% compared to FedAvg. The shared model improves the aggregation effect and can be trained without sharing raw data. The CSM evaluates the most balanced data distribution based on several key factors to train the shared model. The Dirichlet process is used to simulate various data distributions and compare performance evaluations between the IID and Non-IID scenarios. Results indicate that Balanced CSM can further enhance accuracy by 6% in highly imbalanced data distributions.
To summarize, this paper discussed the impacts of data imbalance on test accuracy and convergence time in FL. Some issues, such as secure transmission, heterogeneous clients, and noise data during FL training, were not considered in this paper. These issues may require further investigation to gain a more comprehensive understanding. The mobility of participating clients is also interesting, and FL could potentially be combined with Internet of Vehicles (IoV) scenarios, which may encounter various issues such as communication costs and convergence time among dynamic joining/leaving clients.
In future work, three directions are planned to move forward. First, developing a framework for practical deployments of distributed FL is critical. The Taiwan AI lab has already developed a simple project using Kubernetes. The leverage of the open-source project and our work could be feasible. Second, moving towards a decentralized FL framework is also an important area for future work. Last but not least, considering factors such as late joiners or offline participants that occur in practical scenarios may bring another challenge to FL. Addressing these challenges of deploying FL in practical scenarios is also a potential topic for further research.

Author Contributions

Conceptualization, W.-C.C. and Y.-H.L.; methodology, W.-C.C. and Y.-H.L.; software, Y.-H.L.; investigation, Y.-H.L. and S.-H.F.; data curation, Y.-H.L.; writing—original draft preparation, Y.-H.L. and S.-H.F.; writing—review and editing, W.-C.C. and Y.-H.L.; visualization, Y.-H.L.; supervision, W.-C.C.; funding acquisition, W.-C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is partially supported by the National Science and Technology Council (NSTC) of Taiwan under Grant No. 111-2221-E-033-033.

Data Availability Statement

The dataset utilized in our publication can be accessed at the following link: https://www.kaggle.com/datasets/tawsifurrahman/covid19-radiography-database, accessed on 17 May 2023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, L.; Lin, Z.Q.; Wong, A. COVID-Net: A Tailored Deep Convolutional Neural Network Design for Detection of COVID-19 Cases from Chest X-ray Images. Sci. Rep. 2020, 10, 19549. [Google Scholar] [CrossRef] [PubMed]
  2. Zhao, W.; Jiang, W.; Qiu, X. Deep Learning for COVID-19 Detection Based on CT Images. Sci. Rep. 2021, 11, 14353. [Google Scholar] [CrossRef] [PubMed]
  3. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  4. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–27 April 2017; pp. 1273–1282. [Google Scholar]
  5. Sikandar, H.S.; Waheed, H.; Tahir, S.; Malik, S.U.; Rafique, W. A Detailed Survey on Federated Learning Attacks and Defenses. Electronics 2023, 12, 260. [Google Scholar] [CrossRef]
  6. Voigt, P.; Von dem Bussche, A. The EU General Data Protection Regulation (GDPR). In A Practical Guide, 1st ed.; Springer International Publishing: Charm, Switzerland, 2017; Volume 10, p. 3152676. [Google Scholar]
  7. Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated Optimization in Heterogeneous Networks. In Proceedings of the Machine Learning and Systems, MLSys, Austin, TX, USA, 2–4 March 2020; pp. 429–450. [Google Scholar]
  8. Wang, J.; Liu, Q.; Liang, H.; Joshi, G.; Poor, H.V. Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization. In Proceedings of the 34th Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6–12 December 2020; pp. 7611–7623. [Google Scholar]
  9. Ho, T.-T.; Tran, K.-D.; Huang, Y. FedSGDCOVID: Federated SGD COVID-19 Detection under Local Differential Privacy using Chest X-ray Images and Symptom Information. Sensors 2022, 22, 3728. [Google Scholar] [CrossRef]
  10. Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep Learning with Differential Privacy. In Proceedings of the ACM SIGSAC on Computer and Communications Security, Vienna, Austria, 25–27 October 2016; pp. 308–318. [Google Scholar]
  11. Li, Q.; Diao, Y.; Chen, Q.; He, B. Federated Learning on Non-IID Data Silos: An Experimental Study. In Proceedings of the 2022 IEEE 38th International Conference on Data Engineering (ICDE), Kuala Lumpur, Malaysia, 9–12 May 2022; pp. 965–978. [Google Scholar]
  12. Elshabrawy, K.M.; Alfares, M.M.; Salem, M.A.-M. Ensemble Federated Learning for Non-IID COVID-19 Detection. In Proceedings of the 2022 5th International Conference on Computing and Informatics (ICCI), New Cairo, Cairo, Egypt, 9–10 March 2022; pp. 57–63. [Google Scholar]
  13. Chang, K.; Balachandar, N.; Lam, C.; Yi, D.; Brown, J.; Beers, A.; Rosen, B.; Rubin, D.L.; Kalpathy-Cramer, J. Distributed Deep Learning Networks among Institutions for Medical Imaging. J. Am. Med. Inform. Assoc. 2018, 25, 945–954. [Google Scholar] [CrossRef]
  14. Sheller, M.J.; Edwards, B.; Reina, G.A.; Martin, J.; Pati, S.; Kotrotsou, A.; Milchenko, M.; Xu, W.; Marcus, D.; Colen, R.R. Federated Learning in Medicine: Facilitating Multi-institutional Collaborations without Sharing Patient Data. Sci. Rep. 2020, 10, 12598. [Google Scholar] [CrossRef]
  15. Kirkpatrick, J.; Pascanu, R.; Rabinowitz, N.; Veness, J.; Desjardins, G.; Rusu, A.A.; Milan, K.; Quan, J.; Ramalho, T.; Grabska-Barwinska, A. Overcoming catastrophic forgetting in neural networks. arXiv 2016, arXiv:1612.00796. [Google Scholar] [CrossRef]
  16. Chhikara, P.; Singh, P.; Tekchandani, R.; Kumar, N.; Guizani, M. Federated Learning Meets Human Emotions: A Decentralized Framework for Human–Computer Interaction for IoT Applications. IEEE Internet Things J. 2020, 8, 6949–6962. [Google Scholar] [CrossRef]
  17. Roth, H.R.; Chang, K.; Singh, P.; Neumark, N.; Li, W.; Gupta, V.; Gupta, S.; Qu, L.; Ihsani, A.; Bizzo, B.C. Federated Learning for Breast Density Classification: A Real-World Implementation. In Proceedings of the Domain Adaptation and Representation Transfer, and Distributed and Collaborative Learning, Lima, Peru, 4–8 October 2020; pp. 181–191. [Google Scholar]
  18. Zhao, Y.; Li, M.; Lai, L.; Suda, N.; Civin, D.; Chandra, V. Federated Learning with Non-IID Data. arXiv 2018, arXiv:1806.00582. [Google Scholar] [CrossRef]
  19. Nishio, T.; Yonetani, R. Client Selection for Federated Learning with Heterogeneous Resources in Mobile Edge. In Proceedings of the IEEE International Conference on Communications (ICC), Shanghai, China, 20–24 May 2019; pp. 1–7. [Google Scholar]
  20. Sidey-Gibbons, J.A.; Sidey-Gibbons, C.J. Machine Learning in Medicine: A Practical Introduction. BMC Med. Res. Methodol. 2019, 19, 64. [Google Scholar] [CrossRef] [PubMed]
  21. Erickson, B.J.; Korfiatis, P.; Akkus, Z.; Kline, T.L. Machine Learning for Medical Imaging. Radiographics 2017, 37, 505. [Google Scholar] [CrossRef] [PubMed]
  22. De Bruijne, M. Machine Learning Approaches in Medical Image Analysis: From Detection to Diagnosis. Med. Image Anal. 2016, 33, 94–97. [Google Scholar] [CrossRef] [PubMed]
  23. Lundervold, A.S.; Lundervold, A. An Overview of Deep Learning in Medical Imaging Focusing on MRI. Z. Med. Phys. 2019, 29, 102–127. [Google Scholar] [CrossRef]
  24. Prasad, V.K.; Bhattacharya, P.; Maru, D.; Tanwar, S.; Verma, A.; Singh, A.; Tiwari, A.K.; Sharma, R.; Alkhayyat, A.; Țurcanu, F.-E. Federated Learning for the Internet-of-Medical-Things: A Survey. Mathematics 2023, 11, 151. [Google Scholar] [CrossRef]
  25. Nafisah, S.I.; Muhammad, G.; Hossain, M.S.; AlQahtani, S.A. A Comparative Evaluation Between Convolutional Neural Networks and Vision Transformers for COVID-19 Detection. Mathematics 2023, 11, 1489. [Google Scholar] [CrossRef]
  26. Ravì, D.; Wong, C.; Deligianni, F.; Berthelot, M.; Andreu-Perez, J.; Lo, B.; Yang, G.-Z. Deep Learning for Health Informatics. IEEE J. Biomed. Health Inform. 2016, 21, 4–21. [Google Scholar] [CrossRef]
  27. Yan, B.; Wang, J.; Cheng, J.; Zhou, Y.; Zhang, Y.; Yang, Y.; Liu, L.; Zhao, H.; Wang, C.; Liu, B. Experiments of Federated Learning for COVID-19 Chest X-ray Images. In Proceedings of the International Conference on Artificial Intelligence and Security, Dublin, Ireland, 19–23 July 2021; pp. 41–53. [Google Scholar]
  28. Khan, S.H.; Alam, M.G.R. A Federated Learning Approach to Pneumonia Detection. In Proceedings of the 2021 International Conference on Engineering and Emerging Technologies (ICEET), Istanbul, Turkey, 27–28 October 2021; pp. 1–6. [Google Scholar]
  29. Feki, I.; Ammar, S.; Kessentini, Y.; Muhammad, K. Federated Learning for COVID-19 Screening from Chest X-ray Images. Appl. Soft Comput. 2021, 106, 107330. [Google Scholar] [CrossRef]
  30. Kumar, R.; Khan, A.A.; Kumar, J.; Golilarz, N.A.; Zhang, S.; Ting, Y.; Zheng, C.; Wang, W. Blockchain-Federated-Learning and Deep Learning Models for COVID-19 Detection using CT Imaging. IEEE Sen. J. 2021, 21, 16301–16314. [Google Scholar] [CrossRef]
  31. Durga, R.; Poovammal, E. FLED-Block: Federated Learning Ensembled Deep Learning Blockchain Model for COVID-19 Prediction. Front. Public Health 2022, 10, 892499. [Google Scholar] [CrossRef]
  32. Nguyen, D.C.; Ding, M.; Pathirana, P.N.; Seneviratne, A.; Zomaya, A.Y. Federated Learning for COVID-19 Detection with Generative Adversarial Networks in Edge Cloud Computing. IEEE Internet Things J. 2021, 9, 10257–10271. [Google Scholar] [CrossRef]
  33. Creswell, A.; White, T.; Dumoulin, V.; Arulkumaran, K.; Sengupta, B.; Bharath, A.A. Generative Adversarial Networks: An Overview. IEEE Signal Process. Mag. 2018, 35, 53–65. [Google Scholar] [CrossRef]
  34. DuMont Schütte, A.; Hetzel, J.; Gatidis, S.; Hepp, T.; Dietz, B.; Bauer, S.; Schwab, P. Overcoming Barriers to Data Sharing with Medical Image Generation: A Comprehensive Evaluation. NPJ Digit. Med. 2021, 4, 141. [Google Scholar] [CrossRef] [PubMed]
  35. Zhang, W.; Zhou, T.; Lu, Q.; Wang, X.; Zhu, C.; Sun, H.; Wang, Z.; Lo, S.K.; Wang, F.-Y. Dynamic-Fusion-Based Federated Learning for COVID-19 Detection. IEEE Internet Things J. 2021, 8, 15884–15891. [Google Scholar] [CrossRef] [PubMed]
  36. Cetinkaya, A.E.; Akin, M.; Sagiroglu, S. A Communication Efficient Federated Learning Approach to Multi Chest Diseases Classification. In Proceedings of the 2021 6th International Conference on Computer Science and Engineering (UBMK), Ankara, Turkey, 15–17 September 2021; pp. 429–434. [Google Scholar]
  37. Kandati, D.R.; Gadekallu, T.R. Genetic Clustered Federated Learning for COVID-19 Detection. Electronics 2022, 11, 2714. [Google Scholar] [CrossRef]
  38. Federated Learning Made Easy. Available online: https://github.com/ailabstw/harmonia (accessed on 1 May 2023).
  39. Deng, L. The MNIST Database of Handwritten Digit Images for Machine Learning Research. IEEE Signal Process. Mag. 2012, 29, 141–142. [Google Scholar] [CrossRef]
  40. Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; Cummings, R. Advances and Open Problems in Federated Learning. Found. Trends Mach. Learn. 2019, 14, 1–210. [Google Scholar] [CrossRef]
  41. Luo, J.; Wu, S. FedSLD: Federated Learning with Shared Label Distribution for Medical Image Classification. In Proceedings of the IEEE 19th International Symposium on Biomedical Imaging (ISBI), Kolkata, India, 28–31 March 2022; pp. 1–5. [Google Scholar]
  42. Gao, L.; Fu, H.; Li, L.; Chen, Y.; Xu, M.; Xu, C.-Z. FedDC: Federated Learning with Non-IID Data via Local Drift Decoupling and Correction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10112–10121. [Google Scholar]
  43. Li, X.; Jiang, M.; Zhang, X.; Kamp, M.; Dou, Q. FedBN: Federated Learning on Non-IID Features via Local Batch Normalization. arXiv 2021, arXiv:2102.07623. [Google Scholar]
  44. Karimireddy, S.P.; Kale, S.; Mohri, M.; Reddi, S.; Stich, S.; Suresh, A.T. Scaffold: Stochastic Controlled Averaging for Federated Learning. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 5132–5143. [Google Scholar]
  45. Nguyen, J.; Wang, J.; Malik, K.; Sanjabi, M.; Rabbat, M. Where to Begin? On the Impact of Pre-Training and Initialization in Federated Learning. arXiv 2022, arXiv:2210.08090. [Google Scholar]
  46. Batista, G.E.; Prati, R.C.; Monard, M.C. A Study of the Behavior of Several Methods for Balancing Machine Learning Training Data. ACM SIGKDD Explor. Newsl. 2004, 6, 20–29. [Google Scholar] [CrossRef]
  47. Chowdhury, M.E.; Rahman, T.; Khandakar, A.; Mazhar, R.; Kadir, M.A.; Mahbub, Z.B.; Islam, K.R.; Khan, M.S.; Iqbal, A.; Al Emadi, N. Can AI Help in Screening Viral and COVID-19 Pneumonia? IEEE Access 2020, 8, 132665–132676. [Google Scholar] [CrossRef]
  48. Rahman, T.; Khandakar, A.; Qiblawey, Y.; Tahir, A.; Kiranyaz, S.; Kashem, S.B.A.; Islam, M.T.; Al Maadeed, S.; Zughaier, S.M.; Khan, M.S. Exploring the Effect of Image Enhancement Techniques on COVID-19 Detection using Chest X-ray Images. Comput. Biol. Med. 2021, 132, 104319. [Google Scholar] [CrossRef] [PubMed]
  49. Kermany, D.; Zhang, K.; Goldbaum, M. Labeled Optical Coherence Tomography (OCT) and Chest X-ray Images for Classification. Mendeley Data 2018, 2, 2. [Google Scholar]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  51. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A Unified Embedding for Face Recognition and Clustering. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  52. Blei, D.M.; Ng, A.Y.; Jordan, M.I. Latent Dirichlet Allocation. J. Mach. Learn. Res. 2003, 3, 993–1022. [Google Scholar]
Figure 1. Illustration of FL training paradigm.
Figure 1. Illustration of FL training paradigm.
Mathematics 11 02385 g001
Figure 2. Illustrations of the shared data approach compared with the shared model approach. (a) Shared data approach: (a-1) server owns a portion of the shared data and sends both the data and the global model to each institution; (a-2) each institution performs local training; (a-3) each institution sends its local model to the server; (a-4) server aggregates the local models to obtain the global model. (b) Shared model approach: (b-1) server owns a portion of the shared data and trains the global model to obtain the shared model; (b-2) server sends the shared model to each institution; (b-3) each institution performs local training; (b-4) each institution sends its local model to the server; (b-5) server aggregates the local models to obtain the global model.
Figure 2. Illustrations of the shared data approach compared with the shared model approach. (a) Shared data approach: (a-1) server owns a portion of the shared data and sends both the data and the global model to each institution; (a-2) each institution performs local training; (a-3) each institution sends its local model to the server; (a-4) server aggregates the local models to obtain the global model. (b) Shared model approach: (b-1) server owns a portion of the shared data and trains the global model to obtain the shared model; (b-2) server sends the shared model to each institution; (b-3) each institution performs local training; (b-4) each institution sends its local model to the server; (b-5) server aggregates the local models to obtain the global model.
Mathematics 11 02385 g002
Figure 3. FL model aggregation profiles under different data distributions.
Figure 3. FL model aggregation profiles under different data distributions.
Mathematics 11 02385 g003
Figure 4. Illustrations of the candidate selection mechanism of the shared model. (a) Initialized process: (a-1) each institution sends the amount of data to the server; (a-2) server summarizes the total amount of data and the amount of labels for the task, and then sends the results back to the institution; (a-3) each institution then calculates the score and sends the value to the server. (b) Candidate and training process: (b-4) the server sorts the institutions based on the scores calculated by each institution and selects the institution with the highest score for training first; (b-5) the model trained by the institution with the highest score is sent to the remaining institutions for training; (b-6) each institution sends its local model to the server; (b-7) server aggregates the local models to obtain the global model.
Figure 4. Illustrations of the candidate selection mechanism of the shared model. (a) Initialized process: (a-1) each institution sends the amount of data to the server; (a-2) server summarizes the total amount of data and the amount of labels for the task, and then sends the results back to the institution; (a-3) each institution then calculates the score and sends the value to the server. (b) Candidate and training process: (b-4) the server sorts the institutions based on the scores calculated by each institution and selects the institution with the highest score for training first; (b-5) the model trained by the institution with the highest score is sent to the remaining institutions for training; (b-6) each institution sends its local model to the server; (b-7) server aggregates the local models to obtain the global model.
Mathematics 11 02385 g004
Figure 5. Data distributions and samples of COVID Radiography Database [47,48].
Figure 5. Data distributions and samples of COVID Radiography Database [47,48].
Mathematics 11 02385 g005
Figure 6. CL with different proportions of training data.
Figure 6. CL with different proportions of training data.
Mathematics 11 02385 g006
Figure 7. Visualization of data distributions for Dirichlet α = 0.1, 0.5, 1, and IID for four institutions. The size of circles in the figure indicates the sample size. A larger circle represents the label with more data, while a smaller circle represents the label with less data.
Figure 7. Visualization of data distributions for Dirichlet α = 0.1, 0.5, 1, and IID for four institutions. The size of circles in the figure indicates the sample size. A larger circle represents the label with more data, while a smaller circle represents the label with less data.
Mathematics 11 02385 g007
Figure 8. Visualization of data distributions for Dirichlet α = 0.1, 0.5, 1, and IID for ten institutions. The size of circles in the figure indicates the sample size. A larger circle represents the label with more data, while a smaller circle represents the label with less data.
Figure 8. Visualization of data distributions for Dirichlet α = 0.1, 0.5, 1, and IID for ten institutions. The size of circles in the figure indicates the sample size. A larger circle represents the label with more data, while a smaller circle represents the label with less data.
Mathematics 11 02385 g008
Figure 9. Data distributions of Dirichlet α = 0.5 in each institution with the standard deviation of sample size.
Figure 9. Data distributions of Dirichlet α = 0.5 in each institution with the standard deviation of sample size.
Mathematics 11 02385 g009
Table 1. Summaries of existing FL approaches for COVID-19 detection.
Table 1. Summaries of existing FL approaches for COVID-19 detection.
WorkMulti-ClassApproachData DistributionContribution
Ho et al. [9]CNN-Spatial Pyramid Pooling, Artificial Neural Network, DP-SGD,IID/Non-IIDPrivacy
Yan et al. [27]MobileNet, ResNet18, ResNeXt, COVID-Net-Comparing different existing methods
Khan et al. [28] CNN, ResNet50, VGG16, AlexNetIIDComparing different existing methods
Feki et al. [29] VGG16, ResNet50IID/Non-IIDNon-IID
Rajesh et al. [30] SegCaps, share data-Improving training accuracy by feature extraction
Durga et al. [31] Ensembled capsule networks,
extreme learning machines
Non-IIDImproving training accuracy by feature extraction
Nguyen et al. [32]GANIIDA blockchain framework for FL with GAN
Zhang et al. [35]Dynamic FusionIID/Non-IIDCommunication cost
Cetinkaya et al. [36]New CNN model, weight pruningIIDCommunication cost
Kandati et al. [37]Genetic algorithms, clustered FLIIDOptimizing FL hyperparameter
FedISM
(Our Proposed)
Shared model, CSMIID/Non-IIDNon-IID problem
Table 2. Method comparisons of shared model and FedAvg in various data distributions.
Table 2. Method comparisons of shared model and FedAvg in various data distributions.
Data DistributionMethodBest AccuracyConvergence Time (80%)
Non-IID (1)FedAvg51.688%-
Shared Model 5%76.906%-
Shared Model 10%80.141%71 rounds
Shared Model 15%81.487%47 rounds
Non-IID (2)FedAvg73.530%-
Shared Model 5%81.558%66 rounds
Shared Model 10%85.714%8 rounds
Shared Model 15%85.832%3 rounds
Non-IID (3)FedAvg88.925%5 rounds
Shared Model 5%88.760%3 rounds
Shared Model 10%88.949%4 rounds
Shared Model 15%89.964%3 rounds
Non-IID (4)FedAvg90.035%4 rounds
Shared Model 5%90.389%3 rounds
Shared Model 10%90.129%3 rounds
Shared Model 15%90.507%3 rounds
Table 3. Detailed method results for FedAvg and CSM with different weights for various Dirichlet processes.
Table 3. Detailed method results for FedAvg and CSM with different weights for various Dirichlet processes.
Data DistributionMethodBest AccuracyConvergence Time (70%)
α = 0.1FedAvg64.203%-
CSM β = 0.262.809%-
CSM β = 0.572.325%57 rounds
CSM β = 0.872.325%57 rounds
α = 0.5FedAvg86.493%5 rounds
CSM β = 0.286.446%3 rounds
CSM β = 0.586.446%3 rounds
CSM β = 0.887.532%2 rounds
α = 1FedAvg88.476%2 rounds
CSM β = 0.289.586%1 round
CSM β = 0.589.586%1 round
CSM β = 0.889.586%1 round
IIDFedAvg90.318%2 rounds
CSM β = 0.290.649%1 round
CSM β = 0.590.649%1 round
CSM β = 0.890.649%1 round
100% dataCL91.334%-
Table 4. Detailed method results for FedAvg, CSM with β = 0.8, and Balanced CSM with various Dirichlet processes.
Table 4. Detailed method results for FedAvg, CSM with β = 0.8, and Balanced CSM with various Dirichlet processes.
Data DistributionMethodBest AccuracyConvergence Time (75%)
α = 0.1FedAvg76.174%13 rounds
CSM β = 0.880.850%8 rounds
Balanced CSM82.266%7 rounds
α = 0.5FedAvg87.508%7 rounds
CSM β = 0.886.942%14 rounds
Balanced CSM88.311%3 rounds
α = 1FedAvg87.508%4 rounds
CSM β = 0.888.736%3 rounds
Balanced CSM89.279%2 rounds
IIDFedAvg89.209%5 rounds
CSM β = 0.889.964%2 rounds
Balanced CSM89.964%2 rounds
100% dataCL91.334%-
Table 5. Comparisons against different methods in terms of accuracy, precision, recall, and F1-score.
Table 5. Comparisons against different methods in terms of accuracy, precision, recall, and F1-score.
No.LabelAccuracyPrecisionRecallF1-ScoreMechanism
Dirichlet (α = 0.1)0303820174.64078.45474.64073.033
1270597477.80479.59077.80475.894
2286623779.97680.12879.97679.515
3125385232080.21375.03580.21377.459
41372262477.75779.11077.75775.307
500560563.25977.02363.25963.218
649000072.94072.45572.94067.343
7319001466.51775.55166.51758.353
81748236719073.71976.66973.71970.127CSM β = 0.2
9418464941182.26782.75182.26780.978CSM β = 0.8,
Balanced CSM
Dirichlet (α = 0.5)02211199977387.24987.27087.24987.169
159268638751388.31288.35688.31288.237Balanced CSM
23940302187.93488.04387.93487.935
3430030373686.56486.59286.56486.473CSM β = 0.2 & β = 0.8
40524136513687.03787.03387.03787.022
5116513447087.76987.86187.76987.672
6266912087.72187.99687.72187.696
71462253217787.81687.95787.81687.725
8959919273087.88788.09387.88787.774
9393514130114088.17088.18488.17088.102
Dirichlet (α = 1)030865974911889.28089.29189.28089.199Balanced CSM
168752210026988.99689.01188.99688.983
2132465135111088.45388.56388.45388.463
38645072620088.97388.93188.97388.931
425178010214788.73788.69088.73788.676
52014217212288.85588.90688.85588.831
6126275781788.47788.50488.47788.413
7483141721588.14688.25588.14688.120
845135764214488.38388.48888.38388.223
9573118163724488.73788.75288.73788.650CSM β = 0.2 & β = 0.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chung, W.-C.; Lin, Y.-H.; Fang, S.-H. FedISM: Enhancing Data Imbalance via Shared Model in Federated Learning. Mathematics 2023, 11, 2385. https://doi.org/10.3390/math11102385

AMA Style

Chung W-C, Lin Y-H, Fang S-H. FedISM: Enhancing Data Imbalance via Shared Model in Federated Learning. Mathematics. 2023; 11(10):2385. https://doi.org/10.3390/math11102385

Chicago/Turabian Style

Chung, Wu-Chun, Yan-Hui Lin, and Sih-Han Fang. 2023. "FedISM: Enhancing Data Imbalance via Shared Model in Federated Learning" Mathematics 11, no. 10: 2385. https://doi.org/10.3390/math11102385

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop