Next Article in Journal
Endoscopic Ultrasound Quality Metrics in Clinical Practice
Next Article in Special Issue
Conventional Machine Learning versus Deep Learning for Magnification Dependent Histopathological Breast Cancer Image Classification: A Comparative Study with Visual Explanation
Previous Article in Journal
Local Bone Mineral Density, Subcutaneous and Visceral Adipose Tissue Measurements in Routine Multi Detector Computed Tomography—Which Parameter Predicts Incident Vertebral Fractures Best?
Previous Article in Special Issue
The Utility of Deep Learning in Breast Ultrasonic Imaging: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cloud Computing-Based Framework for Breast Cancer Diagnosis Using Extreme Learning Machine

1
Department of Computer Science and Engineering, DAV University, Jalandhar 144 012, Punjab, India
2
Department of Computer Science and Engineering, Thapar Institute of Engineering and Technology, Patiala 147004, Punjab, India
3
Chitkara University School of Engineering and Technology, Chitkara University, Himachal Pradesh, India
4
Information Systems Department, College of Computer Science and Information Technology, University of Anbar, 55431 Ramadi, Anbar, Iraq
5
Department of Applied Informatics, Vytautas Magnus University, 44404 Kaunas, Lithuania
6
Faculty of Applied Mathematics, Silesian University of Technology, 44-100 Gliwice, Poland
7
Faculty of Applied Computing and Technology (FACT), Noroff University College, 4608 Kristiansand, Norway
8
Department of Electrical—Electronics Engineering, Trakya University, Edirne 22030, Turkey
*
Author to whom correspondence should be addressed.
Diagnostics 2021, 11(2), 241; https://doi.org/10.3390/diagnostics11020241
Submission received: 6 December 2020 / Revised: 28 January 2021 / Accepted: 29 January 2021 / Published: 4 February 2021
(This article belongs to the Special Issue Machine Learning in Breast Cancer Diagnosis and Prognosis)

Abstract

:
Globally, breast cancer is one of the most significant causes of death among women. Early detection accompanied by prompt treatment can reduce the risk of death due to breast cancer. Currently, machine learning in cloud computing plays a pivotal role in disease diagnosis, but predominantly among the people living in remote areas where medical facilities are scarce. Diagnosis systems based on machine learning act as secondary readers and assist radiologists in the proper diagnosis of diseases, whereas cloud-based systems can support telehealth services and remote diagnostics. Techniques based on artificial neural networks (ANN) have attracted many researchers to explore their capability for disease diagnosis. Extreme learning machine (ELM) is one of the variants of ANN that has a huge potential for solving various classification problems. The framework proposed in this paper amalgamates three research domains: Firstly, ELM is applied for the diagnosis of breast cancer. Secondly, to eliminate insignificant features, the gain ratio feature selection method is employed. Lastly, a cloud computing-based system for remote diagnosis of breast cancer using ELM is proposed. The performance of the cloud-based ELM is compared with some state-of-the-art technologies for disease diagnosis. The results achieved on the Wisconsin Diagnostic Breast Cancer (WBCD) dataset indicate that the cloud-based ELM technique outperforms other results. The best performance results of ELM were found for both the standalone and cloud environments, which were compared. The important findings of the experimental results indicate that the accuracy achieved is 0.9868, the recall is 0.9130, the precision is 0.9054, and the F1-score is 0.8129.

1. Introduction

In recent decades, breast cancer has been a predominant cause of mortality amongst women [1,2]. Approximately 15% of mortalities among women are caused by breast cancer [3]. According to future projections of the World Health Organization (WHO), by 2040, the number of cases of breast cancer is predicted to reach 2.7 million worldwide [4]. The situation is alarming for many developing countries and has crippled medical facilities, where the medical staff have been overwhelmed by the COVID-19 pandemic. Early and accurate detection of breast cancer can lead to an early start in treatment and can increase the chances of survival [5], but it is difficult to diagnose cancer in the early stages, and therefore, techniques that can improve accurate detection of breast cancer are always solicited.
In the past few years, cloud computing [6] has emerged as a strong alternative to using costly locally managed computing resources. It is an on-demand service in which data can be processed and managed by storing it on the network of remote servers. Cloud computing services provide a browser-based dashboard, making it easier for the IT personnel to access the services provided by cloud service providers. Cloud computing is convenient for integrating data on the cloud, making it easier to update medical records. Moreover, cloud computing provides a large number of resources that can accommodate huge datasets of biomedical images or speech data [7]. A critical feature of cloud computing is the high availability of the services that can help healthcare industries provide uninterrupted services with less downtime [8].
Cloud computing services are essential for developing assisted living environments [9]. Furthermore, cloud computing services can be used to monitor patients, elderly people, and those with disabilities in remote or inaccessible villages and towns in many underdeveloped countries, where medical facilities and expertise are not readily available [10]. In these areas, women with breast cancer are often left undiagnosed, and ultimately, it is too late when they reach doctors available in larger cities. Doctors can use cloud computing to diagnose patients who cannot reach them due to a lack of financial resources. They can also use cloud computing for guidance through telehealth [11] and telemedicine [12], which includes the transmission of various medical data, such as high-resolution biomedical photographs and patient video recordings from remote areas to other geographic locations, where specialist physicians and large hospitals are situated. Cloud computing also enables essential services such as a rapid search tool for blood and organ donors in the case of emergencies [13].
Studies have investigated machine learning methods for increasing the accuracy of disease diagnoses [14,15,16,17] in addition to advanced medical imaging techniques such as mammography. One such method is extreme learning machine (ELM) [18], which is a kind of artificial neural network (ANN). ELM has been applied to various applications such as colorectal cancer [19]; thyroid disease [20]; Parkinson’s disease [21]; brain tumors [22]; osteoarthritis [23]; and most recently, COVID-19 pneumonia [24] diagnosis. Due to the advantages of fast learning speed and low computational cost, ELMs have become popular for solving many complex problems. The main contributions of this study are as follows:
  • A design of a cloud-based diagnosis system to monitor remote user health data for breast cancer diagnosis is proposed. Through an analysis of consumer health data stored on cloud servers, the method is flexible enough to diagnose and classify a variety of diseases.
  • ELM is used to classify patient data for breast cancer detection.
  • The ELM model is compared with other traditional classification algorithms. Large datasets are supported using the cloud to reduce execution time; these classification models are compared using the cloud as well as a standalone platform.
  • To further improve the model’s classification performance, feature selection is used to remove irrelevant features, and the hidden layer nodes of ELM are tuned.
  • The best performance results of ELM for both standalone and cloud environments are compared.
The remainder of this paper is organized a follows: In Section 2, related work is presented; in Section 3, a description of the methodology used in this work is provided; in Section 4, we discuss the setup of the experimental environment; in Section 5, we discuss the various results obtained in this study; and, in Section 6, we discuss the implications of the results as well as the conclusions and future work.

2. Related Work

The diagnosis of breast cancer disease is an area of interest for many researchers [25]. Below, we discuss some of the disease diagnosis systems. Gupta et al. [26] proposed a heart disease prediction system using a cloud environment. Various algorithms including Random Forest (RF), J48, multi-layer perceptron, Naïve Bayes (NB), Binary Discriminant, Boosted tree, AdaBoost, and Support Vector Machine (SVM) were executed on the Cleveland dataset. Firstly, the algorithms were implemented on the standalone system by using various evaluation criteria. Later, the three algorithms that had the best accuracy were ensembled in the cloud environment. Saba et al. [27] discussed a framework in which breast cancer cells can be detected and classified using cytology images. Furthermore, features that incorporate shape were used to detect tumor cells using ANNs and an NB classifier. Goncalves et al. [28] discussed an approach to early breast cancer diagnosis. This work followed two different strategies. The first step involved in the classification process used ANN, and the second step focused on SVM. Rodriguez-Ruiz et al. [29] evaluated an Artificial Intelligent (AI) system against radiologists in the detection of breast cancer using digital mammograms. The results acquired proved that the AI system was able to detect breast cancer far more accurately than radiologists. Ragab et al. [30] suggested a system for the diagnosis of breast tumors. This system is segmented into two parts in which the features are retrieved using the deep convolutional network and support vector machines are used for obtaining better accuracy. Kashif et al. [31] suggested a hybrid model for predicting breast cancer from mammography images. First, the images were segmented and the features were extracted using mammogram processing and then classification was performed using the extracted features.
Hamed et al. [32] proposed using the You Only Look Once (YOLO) and RetinaNet models for breast cancer recognition while achieving 91% accuracy of five mammogram image datasets. Ak [33] discussed various approaches of machine learning and applied them to the Wisconsin Diagnostic Breast Cancer (WBCD) dataset, focusing on comparative analysis and data visualization. Jeyanathan et al. [34] extracted features from breast thermograms using wavelet, curvelet, and contourlet transform for breast cancer recognition, achieving an accuracy of 91%, a sensitivity of 87%, and a specificity of 90% using the AdaBoost classifier. Abdar et al. [35] used voting and stacking techniques to construct a two-layer nested ensemble (NE) model with single classifiers (naïve Bayes and BayesNet), which was tested on the WDBC dataset, achieving an accuracy of 98.07%. Dhahri et al. [36] compared the performance of KNN, SVM, Decision Trees (DT), Random Forest (RF), AdaBoost, Gradient Boosting (GB), Gaussian Naïve Bayes (GNB), Linear Discriminant Analysis (LDA), quadratic discriminant analysis (QDA), linear regression, and extra trees classifier, while the features were selected using Genetic Programming (GP) optimization. The AdaBoost classifier seemed to exhibit the best accuracy of 98.24% on the WDBC dataset. Khan et al. [37] adopted pretrained CNNs (GoogLeNet, VGGNet, and ResNet), which were fed into a fully connected network layer for the classification of malignant and benign cells using average pooling classification, which achieved a 97.52% accuracy on two breast microscopic image datasets.
McKinney et al. [38] proposed an AI system that outperformed human experts in breast cancer prediction on mammogram images. Memon et al. [39] suggested using a modified recursive feature selection algorithm that achieved 99% accuracy with an SVM classifier on the WDBC dataset. Ronoud and Asadi [40] suggested using the genetic algorithm (GA) to evolve the number of hidden layers and neurons and to finetune the network weights and biases of the deep belief network (DBN). Finally, DBN was combined with an ELM classifier, which achieved 99.75% accuracy on the Breast Cancer Wisconsin—Original (WBCO) data and an accuracy of 99.12% on the WDBC dataset. Ting et al. [41] proposed a deep classification algorithm to detect and classify breast cancer in mammogram images, achieving an accuracy of 90.50% and a specificity of 90.71%. Vijayarajeswari et al. [42] combined Hough transform for feature extraction from mammograms and SVM for classification while achieving an accuracy of 94% on a small dataset of images. Wu et al. [43] suggested a deep CNN for breast cancer recognition, achieving an AUC of 0.89 over a large dataset of mammogram images. Assiri et al. [44] suggested using ensemble classification, which combined logistic regression learning, SVM with stochastic gradient descent optimization, and multilayer perceptron network, with a hard voting mechanism. This scheme achieved 99.42% accuracy on the WBCD dataset.
Table 1 presents a summary of some of the prominent disease diagnosis systems. Even though previous studies were promising, there is still room for improvement and development in breast cancer diagnosis methods. Some noteworthy facts used as inspiration are as follows:
  • Most of the studies did not consider feature selection and ELM as their primary algorithm for the diagnosis of breast cancer.
  • The most important issue is that many of the previous studies restricted their models to standalone systems, and thus, they are not available anytime and anywhere.
  • Many of these studies are unique to a particular field of study, but the approach should apply to all fields.
To solve the above issues, this study contributes to the following points:
  • ELM is considered as the primary classification algorithm.
  • To further improve the model’s classification performance, feature selection is used and the hidden layer nodes of ELM are tuned.
  • The ELM model is deployed in the cloud environment.

3. Cloud-Based Breast Cancer Diagnosis Model

This research proposes the design of a cloud-based breast cancer diagnosis system that provides monitoring of remote user health data for the identification of breast cancer. When analyzing consumer health data stored on cloud servers, the method is flexible enough to diagnose and classify a variety of diseases. However, in this paper, we concentrated primarily on only one case of usage, namely defining the disease as “cancerous” or “noncancerous”. The outline of our proposed architecture is shown in Figure 1. In the proposed architecture, the patient goes to a remote healthcare center in their village, where the healthcare service provider collects the data from the patient, such as x-rays and other health parameters, and sends the data via the Internet to a doctor; then, the doctor uploads the data to the cloud platform for further processing.
In the cloud, the processing occurs in two stages. It has been established by past researchers that attribute selection improves the performance of machine learning methods [56,57,58]. Therefore, in the first stage, the gain ratio method is used to identify the significant features and to then remove the insignificant ones. The purpose of this step is to reduce the computational complexity. In the second stage, the classification is applied using ELM.

3.1. Gain Ratio

The gain ratio [59] is a single attribute evaluation method that uses ranking to eliminate irrelevant attributes. It improves the information gain method by removing the favoritism towards attributes with many values by normalizing the formula of information gain using the information value of the split. It is a filter method that performs noniterative computation on the dataset to find the relevance of the attribute by using the following formula:
G a i n   r a t i o   ( X ) = I n f o r m a t i o n   G a i n   ( X ) H ( X ) ,
where H ( X ) =   i P j   log 2   P j , where P j is the probability of having a class j.

3.2. Extreme Learning Machine (ELM)

ELM [60] is a type of feed-forward neural network usually used for classification, regression, clustering, small estimate, compression, and pattern learning with either a single layer or various layers of hidden nodes, where the parameters of hidden nodes that include biases and weights need not be adjusted. On the other hand, the parameters of hidden nodes can be allocated randomly and never changed or can be inherited from their ancestors without alteration. These models learn extremely quicker than networks trained with backpropagation. The prevalent learning procedure used in feed-forward neural networks is the learning procedure for backpropagation, where propagating from the output to the input gradients can be determined. However, backpropagation possesses many problems. The training process is very time-consuming in most applications as weights and biases are rationalized afterward each iteration. To achieve maximum accuracy, the weight magnitude is disregarded in this model, due to which the output becomes worse over time. The local minima also affect the efficiency of the learning algorithm for backpropagation. ELM is a feed-forward network that removes the barrier of updating weights and biases. It focuses not only on minimum training error but also on achieving the lowest weight standards that increase the overall efficiency of this model. The problem of trapping in local minima is handled using simple alternatives avoiding such trivial issues. Figure 2 presents the working of ELM.
For H arbitrary samples ( p i ,   t i ) , where p i = [ p i 1 ,   p i 2 . p i n ] T   Q n and   t i = [ t i 1 ,   t i 2 . t i m ] T   Q m , the standard single-hidden layer feedforward neural networks (SLFNs) with activation function f ( · ) and G hidden nodes can be written as
i = 1 G w i f i ( p j ) = i = 1 G w i f ( a i ×   p j + c i ) =   o j ,   ( j = 1 , 2 ,   H ) ,
where a i = [ a i 1 ,   a i 2 . a i n ] T is the weight vector linking ith hidden node and input nodes,   w i = [ w i 1 ,   w i 2 . w i n ] T   is the weight vector linking ith hidden node to output node, c i is the threshold of hidden node, and o j = [ o j 1 ,   o j 2 . o j m ] T is the jth output vector of SLFNs.
Standard SLFNs with G hidden nodes and activation function f ( · ) can estimate these H illustrations with zero error, which means that j = 1 G | | o j t j | | = 0 and that there exist ω i, a i , and c i such that
j = 1 G w i f ( a i × y j + c i ) = t j   ( j = 1 , 2 , H ) ,
The above equation can be summarized as follows:
M w = T ,
where
M ( a 1 , , a G ,   c 1 , ,   c G ,   y 1 , ,   y G ) =   [ f ( a 1 × y 1 + c 1 ) f ( a G × y 1 + c G ) f ( a 1 × y H + c 1 ) f ( a G × y H + c G ) ] H   ×   G ,
w = [ w 1 T . . . w Ň T ] G   ×   n ,
T = [ t 1 T . . . t N T ] G   ×   n ,
where M is called an output matrix of hidden layer and the kth column of M is the output of the kth hidden node according to inputs y 1 ,   y 2 . y H . The solution of the linear system is
w =   M 1 T ,
where M−1 is the Moore–Penrose generalized inverse of matrix M.
The output function of ELM is defined as
g ( y ) = p ( y ) w = p ( y )   M 1 T ,
In ELM training, there are three key parameters. These are training set K = { ( y j ,   t j ) | y j Q n , t j     Q m , j = 1 ,   . .   H } ; the hidden node output function f ( a i ,   c i ,   y j ) ; and the hidden node number G. Once the values of the parameters are set properly, the training process of ELM can be initiated. Firstly, ELM randomly generates values for the G pair of hidden nodes parameters ( a i ,   c i ) . Then, the output matrix M is generated using Equation (4) according to the input and arbitrarily produced parameters. Then, the output weight vector ω is generated using Equation (8). The classification outcome of test data tuples can be forecasted using Equation (9) after the training phase is completed.
The training of ELM is performed as follows.
Input a training set A = { ( a i ,   d i ) | a i   X n ,   d i   X m ,   i = 1 , . , N } , activation function f ( x ) , and number of hidden neurons N.
  • The weights wi of input and bias bi are allocated randomly.
  • The output matrix M of the hidden layer is computed.
  • Compute the output weight w as
w = M × T ,
where M and T are represented by Equations (4) and (5), respectively.

3.3. Evaluation Criteria

The key idea of this study is to diagnose an input sample whether it belongs to a class of positive samples or belongs to negative samples. There are four possibilities of prediction, which can be described using the terms shown in Table 2.
Further, Table 3 shows the formulae of evaluation metrics. Classification accuracy is the total number of data tuples correctly classified out of the total number of classifications. Precision is the number of positive outcomes correctly classified out of the total positive outcomes forecasted by the classifier. Recall is the proportion of correct predictions of positives to the total number of actual positives. Kappa is a helpful evaluation metric, but due to its complexities, it is underutilized. This metric helps in problems in which there is multiclass classification. F-score represents the harmonic mean between precision and recall with values falling in [0, 1]. It shows the accuracy and reliability of the classifier.

4. Research Materials and Methods

The experimental methodology of this study is divided into two parts. Firstly, the authors considered multiple classification models that include K-nearest neighbors [61], Naïve Bayes [62], Perceptron network [63], AdaBoost [64], and Support Vector Machine [65] and compared all these classification models with the ELM on the standalone environment, and later, the ELM model was deployed on the cloud environment. Firstly, the parameters of ELM were varied, and among those, the best results were determined. Later, the best model of ELM was compared with different classification models. The experimental steps are shown in Figure 3. Both the standalone and cloud computing environments are discussed below.

4.1. Cloud Environment

In the cloud environment, Platform-as-a-Service (PaaS) was used on the Amazon EC2 (Amazon.com, Inc., Seattle, Washington, DC, USA) cloud to deploy the ELM models that were compared on the standalone system. The main reason for deploying the models on the cloud environment was to decrease the execution time and to increase the accuracy. Furthermore, shifting the models on the cloud also helps the models remain readily available anytime and anywhere. The virtual machines that have been used in the cloud environment are all based on a LINUX operating system. Later, the results from both the cloud environment and standalone system were compared.
This process was also deployed on the Amazon EC2 cloud environment. The operating system of the virtual machine that was used on the cloud platform was Ubuntu (Canonical Ltd., London, United Kingdom), and the other parameters such as the number of CPUs, HDD space, and RAM varied. The instances m4.xlarge and c5.xlarge were utilized, which have an Intel Xeon (Intel Corporation, Santa Clara, CA, USA) processor. The feature subset selection was performed using Weka ver. 3.8 (University of Waikato, New Zealand) [66].

4.2. Standalone Environment

The standalone system that was used to carry out the experimentation had the following hardware configuration: (1) a memory of 8 GB, (2) Intel i5-7200 u (Intel Corporation, Santa Clara, CA, USA) with a base clock speed of 2.71 GHz, and (3) an HDD space of 1 TB. In this environment, various classification models were implemented using PyCharm IDE ver. 2020.2 (JetBrains s.r.o., Prague, Czech Republic) [67] on the WBCD dataset and were evaluated using various evaluation metrics.

4.3. Collection of Data

The Wisconsin Breast Cancer Diagnosis (WBCD) [68] dataset was used for the experiment. The dataset consisted of 569 entries and 32 attributes, with the diagnosis attribute signifying malignant or benign. Table 4 shows the description of the 32 attributes.
To find the subset of relevant attributes for the classification process using the gain ratio method, ELM with 100 hidden nodes was used. The gain ratio ranked the attributes according to their relevance. To find the final subset of attributes, the accuracy of ELM was found with the first n attributes, where n varied from 3 to 32, and it was observed that the first 14 attributes in that rank list gave the maximum accuracy. In the order of rank, the serial number of these attributes were 24, 22, 25, 29, 9, 8, 28, 4, 5, 15, 7, 12, 14, and 27; hence, these 14 attributes were selected for further experiments.

5. Results

This section contains the results that were collected from both the standalone and cloud environments, and the results were compared to visualize the performance as we shifted from the standalone environment to the cloud environment.

5.1. Performance Analysis on Standalone Environment

First, AdaBoost, SVM, naïve Bayesian, perceptron, and KNN and, then, the ELM model were executed on a standalone environment.

5.1.1. Performance Analysis of ELM with Different Hidden Nodes

An attempt was made by the authors to improve the accuracy of ELM by altering the number of nodes in the hidden layer in the ELM model. The minimum hidden layer node count was 50, and the maximum was 250. It was observed that ELM had a performance boost and gave better results when the hidden layer nodes count was altered. Table 5 summarizes the results, while Figure 4 presents a visual illustration.
It is clear from Table 5 that the ELM has maximum accuracy when the number of hidden layer nodes is 200, followed by the number of hidden layer nodes set at 250, 150, 100, and 50. Although the Kappa value when the number of hidden layer nodes is 200 is lower compared to when there are 50, 100, and 150 hidden layer nodes, it can be observed that the recall and F-score values for the 200 and 150 hidden layer nodes are higher compared to the values for the other hidden layer nodes. Hence, from the above comparison using various metrics, it can be observed that the ELM model gives the best accuracy (0.969) when the number of hidden layer nodes is taken as 200, followed by 250 (0.9648), 150 (0.956), 100 (0.945), and 50 (0.934). As the number of nodes increases in the hidden layers, each input is handled with multiple neurons, which reduces the load on a single neuron making computation less complex, but as the number of nodes exceeds a certain limit, the computation splits up among different neurons, making it more complex to handle the inputs, ultimately reducing the overall performance of the model [69].

5.1.2. Performance Comparison of ELM with Various Classification Models

In this section, the performance of various classifiers is compared with the ELM classifier with 200 nodes in a standalone environment. Note that 80% of the tuples in the dataset are used for training and the rest of the 20% tuples are used for testing. Metrics of evaluation such as Kappa statistics, accuracy, precision, recall, and F-score were used for comparison. Table 6 summarizes the results, while Figure 5 presents a visual illustration.
From Table 6, by comparing the values of accuracy, ELM provides the highest accuracy of 0.9692 whereas perceptron provides the lowest accuracy of 0.8304. Moreover, ELM has the best recall value (1.00) among all the other classifiers. From the results shown above, it is observed that the ELM-based model provides the highest accuracy of classification followed by SVM, KNN, naïve Bayes (NB), AdaBoost, and perceptron.

5.2. Performance Analysis on Cloud Environment (Amazon EC2)

After comparing the various models in a standalone environment, we observed that the ELM outperformed all the other traditional classification procedures. Therefore, the ELM model was deployed on the cloud environment on virtual machines with different configurations.

Analysis of ELM Performance Using Different Hidden Layer Nodes

As observed, varying the hidden layer node number in the ELM increased its efficiency in the standalone environment. Therefore, the same procedure was followed in the cloud environment to boost the performance of ELM by varying the number of nodes in the hidden layer of the ELM model. Table 7 presents the results obtained on different virtual CPUs (vCPUs) and RAM. Below are the results that were obtained in the cloud environment.
From Table 7, it can be observed that, when the number of hidden layer nodes is set to 200, ELM gives a better performance when there are 4 vCPUs and 16 GB RAM, and the same is the case when there are 8 vCPUs and 32 GB RAM. However, there is an increase in performance when the number of hidden layer nodes is set to 250 and when the number of vCPUs is 16 and RAM is 64 GB, and ELM obtains the best accuracy of 0.9868 when the number of vCPUs is 36 and RAM is 60 GB. It can be seen that, as the number of virtual CPUs, RAM, and number of hidden layer nodes increase, the classification accuracy increases. It can therefore be seen from the above comparison that ELM provides the best classification accuracy of 0.9868 based on different metrics when the number of hidden layer nodes is set to 250 and when vCPU is 36 and RAM is 60 GB. Hence, the ELM model with 250 hidden layer nodes implemented on a virtual machine with 36 vCPUs and 60 GB of RAM is considered best for the classification of breast cancer. Finally, the results are visualized in Figure 6.

5.3. Performance Comparison of ELM on the Cloud Environment and Standalone Environment

The best performance results of ELM were taken for both the standalone and cloud environment, and a comparison was made.
Figure 7 shows ELM’s success in the standalone and cloud environments with specific hidden layer nodes. The classification accuracy of ELM with 250 hidden layers nodes in a standalone environment is 0.9648, whereas in a cloud environment, the classification accuracy is 0.9868 when the number of vCPUs is 36 and RAM is 60 GB. This proves that ELM has a better classification accuracy in the cloud environment when diagnosing patients, whether they are suffering from breast cancer or not, utilizing the data provided by them.
For execution time, there is an improvement of about 18% (from 3.35 s on the standalone computer to 2.81 s on the cloud environment) when we deployed the model to the cloud environment. ELM has a shorter training time since it is a single-layer network feedforward. When a model is deployed on a cloud environment, the execution time is further reduced as resources are available in bulk and allows for computation in a shorter time.
One of the prominent causes of mortality among women is breast cancer. Early determination of this cancer increases survival chances, but women residing in medically underserved areas do not have access to specialist doctors. Machine learning and cloud computing services have drawn the attention of various researchers for developing disease prediction systems, such as [70,71,72,73,74,75,76,77,78], as a feasible option in remote diagnostics, where cloud computing provided Platform-as-a-Service (PaaS) to obtain resources on demand.

6. Conclusions

This paper proposed a framework for cloud-based breast cancer diagnosis using Extreme Learning Machine (ELM) as a classifier. Cloud computing can provide unceasing services anytime and anywhere, which is beneficial for the healthcare industry as they can access the system whenever they want. Moreover, the cloud environment also provides resources that improve the overall classification accuracy of the proposed model. The main advantage of ELM is that the parameters such as weights and biases need not be adjusted, which makes it extremely faster and simpler than all the other gradient-based learning algorithms. In this context, this study proposed a cloud-based architecture for the diagnosis of breast cancer, which collected the data of the patient at remote healthcare centers established near villages and sent the data using cloud services to specialist doctors for analysis and for the provision of further instructions to the patients.
We implemented various classifiers on the WBCD dataset for the diagnosis of breast cancer. Firstly, the gain ratio method was used to select the most relevant attributes and to discard irrelevant attributes. Secondly, various state-of-the-art algorithms were applied and compared with ELM on the standalone system. Further, the ELM model was deployed on the cloud environment using the Amazon EC2 cloud platform. ELM models with different hidden layer nodes were compared on the cloud environment, and the results obtained during experimentation validated that the accuracy of ELM increased in the cloud environment. Therefore, in that case, cloud computing will provide a stable platform since it provides higher accuracy and less execution time than the standalone platform.
The performance of the cloud-based ELM was compared with some state-of-the-art technologies for disease diagnosis. The results achieved on the Wisconsin Diagnostic Breast Cancer (WBCD) dataset indicate that the cloud-based ELM technique outperforms other techniques. The best performance results of ELM were taken from both the standalone and cloud environments, and a comparison has been made. The findings of the experimental results indicate that the accuracy achieved is 0.9868, the recall is 0.9130, the precision is 0.9054, and the F1-score is 0.8129.
In the future, this framework can be further extended by using more resources in a cloud environment that may further increase the classification accuracy of the proposed framework. Moreover, various parameters of ELM can be tuned to further increase the performance of the proposed framework, and the proposed model can also be implemented in the area of image processing, under which various applications such as character recognition, medical imaging, satellite imagery, and photograph enhancement can be implemented.

Author Contributions

Conceptualization, V.L., H.S., A.A., B.S., M.A.M., R.D., S.K., and K.C., methodology, V.L., H.S., A.A., M.A.M., R.D., and K.C., software, V.L., H.S., investigation, V.L., H.S., A.A., B.S., M.A.M., R.D., S.K, and K.C., writing—original draft preparation, V.L., H.S., A.A., B.S., writing—review and editing, V.L., H.S., A.A., B.S., M.A.M., R.D., S.K., and K.C., funding acquisition, R.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available dataset was analyzed in this study. This data can be found here: https://archive.ics.uci.edu/ml/machine-learning-databases/breast-cancer-wisconsin/.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Momenimovahed, Z.; Salehiniya, H. Epidemiological characteristics of and risk factors for breast cancer in the world. Breast Cancer 2019, 11, 151–164. [Google Scholar] [CrossRef] [Green Version]
  2. Mohammed, M.A.; Al-Khateeb, B.; Rashid, A.N.; Ibrahim, D.A.; Abd Ghani, M.K.; Mostafa, S.A. Neural network and multi-fractal dimension features for breast cancer classification from ultrasound images. Comput. Electr. Eng. 2018, 70, 871–882. [Google Scholar] [CrossRef]
  3. Azamjah, N.; Soltan-Zadeh, Y.; Zayeri, F. Global Trend of Breast Cancer Mortality Rate: A 25-Year Study. Asian Pac. J. Cancer Prev. 2019, 20, 2015–2020. [Google Scholar] [CrossRef] [PubMed]
  4. World Health Organization. WHO Report on Cancer: Setting Priorities, Investing Wisely and Providing Care for All; WHO: Geneva, Switzerland, 2020. [Google Scholar]
  5. Obaid, O.I.; Mohammed, M.A.; Ghani, M.K.A.; Mostafa, A.; Taha, F. Evaluating the performance of machine learning techniques in the classification of Wisconsin Breast Cancer. Int. J. Eng. Technol. 2018, 7, 160–166. [Google Scholar]
  6. Al-Fuqaha, A.; Guizani, M.; Mohammadi, M.; Aledhari, M.; Ayyash, M. Internet of things: A survey on enabling technologies, protocols, and applications. IEEE Commun. Surv. Tutor. 2015, 17, 2347–2376. [Google Scholar] [CrossRef]
  7. Mohammed, M.A.; Abdulkareem, K.H.; Mostafa, S.A.; Ghani, M.K.A.; Maashi, M.S.; Garcia-Zapirain, B.; Oleagordia, I.; Alhakami, H.; AL-Dhief, F.T. Voice Pathology Detection and Classification Using Convolutional Neural Network Model. Appl. Sci. 2020, 10, 3723. [Google Scholar] [CrossRef]
  8. Botta, A.; De Donato, W.; Persico, V.; Pescapé, A. Integration of cloud computing and internet of things: A survey. Future Gener. Comput. Syst. 2016, 56, 684–700. [Google Scholar] [CrossRef]
  9. Maskeliunas, R.; Damaševicius, R.; Segal, S. A review of internet of things technologies for ambient assisted living environments. Future Internet 2019, 11, 259. [Google Scholar] [CrossRef] [Green Version]
  10. Połap, D.; Woźniak, M. Introduction to the model of the active assistance system for elder and disabled people. In Communications in Computer and Information Science, Proceedings of the International Conference on Information and Software Technologies, ICIST 2016, Druskininkai, Lithuania, 13–15 October 2016; Springer: Cham, Germany, 2016; Volume 639, pp. 392–403. [Google Scholar] [CrossRef]
  11. Ray Dorsey, E.; Topol, E.J. State of telehealth. N. Engl. J. Med. 2016, 375, 154–161. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Wilson, L.S.; Maeder, A.J. Recent directions in telemedicine: Review of trends in research and practice. Healthc. Inform. Res. 2015, 21, 213–222. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Ayeni, B.; Sowunmi, O.Y.; Misra, S.; Maskeliūnas, R.; Damaševičius, R.; Ahuja, R. A web based system for the discovery of blood banks and donors in emergencies. In Advances in Intelligent Systems and Computing, Proceedings of the International Conference on Intelligent Systems Design and Applications, ISDA 2019, Pretoria, South Africa, 3–5 December 2019; Springer: Cham, Germany, 2019; Volume 1181, pp. 592–600. [Google Scholar] [CrossRef]
  14. Brezulianu, A.; Geman, O.; Zbancioc, M.D.; Hagan, M.; Aghion, C.; Hemanth, D.J.; Son, L.H. IoT Based Heart Activity Monitoring Using Inductive Sensors. Sensors 2019, 19, 3284. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Hemanth, D.J.; Anitha, J.; Naaji, A.; Geman, O.; Popescu, D.E.; Hoang Son, L. A Modified Deep Convolutional Neural Network for Abnormal Brain Image Classification. IEEE Access 2019, 7, 4275–4283. [Google Scholar] [CrossRef]
  16. Almeida, J.S.; Rebouças Filho, P.P.; Carneiro, T.; Wei, W.; Damaševičius, R.; Maskeliūnas, R.; de Albuquerque, V.H.C. Detecting parkinson’s disease with sustained phonation and speech signals using machine learning techniques. Pattern Recognit. Lett. 2019, 125, 55–62. [Google Scholar] [CrossRef] [Green Version]
  17. Sahlol, A.T.; Elaziz, M.A.; Jamal, A.T.; Damaševičius, R.; Hassan, O.F. A novel method for detection of tuberculosis in chest radiographs using artificial ecosystem-based optimisation of deep neural network features. Symmetry 2020, 12, 1146. [Google Scholar] [CrossRef]
  18. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE Int. Jt. Conf. Neural Networks, Seoul, Korea, 17–19 May 2004; Volume 2, pp. 985–990. [Google Scholar]
  19. Ting, W.-C.; Chang, H.-R.; Chang, C.-C.; Lu, C.-J. Developing a Novel Machine Learning-Based Classification Scheme for Predicting SPCs in Colorectal Cancer Survivors. Appl. Sci. 2020, 10, 1355. [Google Scholar] [CrossRef] [Green Version]
  20. Xia, J.; Chen, H.; Li, Q.; Zhou, M.; Chen, L.; Cai, Z.; Fang, Y.; Zhou, H. Ultrasound-based differentiation of malignant and benign thyroid Nodules: An extreme learning machine approach. Comput. Methods Programs Biomed. 2017, 147, 37–49. [Google Scholar] [CrossRef]
  21. Wang, Y.; Wang, A.; Ai, Q.; Sun, H. An adaptive kernel-based weighted extreme learning machine approach for effective detection of parkinson’s disease. Biomed. Signal Process. Control 2017, 38, 400–410. [Google Scholar] [CrossRef]
  22. Khan, M.A.; Ashraf, I.; Alhaisoni, M.; Damaševičius, R.; Scherer, R.; Rehman, A.; Bukhari, S.A.C. Multimodal brain tumor classification using deep learning and robust feature selection: A machine learning application for radiologists. Diagnostics 2020, 10, 565. [Google Scholar] [CrossRef] [PubMed]
  23. Khan, M.A.; Kadry, S.; Parwekar, P.; Damaševičius, R.; Mehmood, A.; Khan, J.A.; Naqvi, S.R. Human Gait Analysis for Osteoarthritis Prediction: A framework of Deep Learning and Kernel Extreme Learning Machine. Complex Intell. Syst. 2021. [Google Scholar] [CrossRef]
  24. Khan, M.A.; Kadry, S.; Zhang, Y.-D.; Akram, T.; Sharif, M.; Rehman, A.; Saba, T. Prediction of COVID-19-Pneumonia based on Selected Deep Features and One Class Kernel Extreme Learning Machine. Comput. Electr. Eng. 2021, 90, 106960. [Google Scholar] [CrossRef] [PubMed]
  25. Elhoseny, M.; Mohammed, M.A.; Mostafa, S.A.; Abdulkareem, K.H.; Maashi, M.S.; Garcia-Zapirain, B.; Mutlag, A.A.; Maashi, M.S. A new multi-agent feature wrapper machine learning approach for heart disease diagnosis. Comput. Mater. Contin. 2021, 67, 51–71. [Google Scholar] [CrossRef]
  26. Gupta, N.; Ahuja, N.; Malhotra, S.; Bala, A.; Kaur, G. Intelligent heart disease prediction in cloud environment through ensembling. Expert Syst. 2017, 34, 1–14. [Google Scholar] [CrossRef]
  27. Saba, T.; Khan, S.U.; Islam, N.; Abbas, N.; Rehman, A.; Javaid, N.; Anjum, A. Cloud-based decision support system for the detection and classification of malignant cells in breast cancer using breast cytology images. Microsc. Res. Tech. 2019, 82, 775–785. [Google Scholar] [CrossRef] [PubMed]
  28. Gonçalves, C.B.; Leles, A.C.Q.; Oliveira, L.E.; Guimaraes, G.; Cunha, J.R.; Fernandes, H. Machine Learning and Infrared Thermography for Breast Cancer Detection. Proceedings 2019, 27, 45. [Google Scholar] [CrossRef] [Green Version]
  29. Rodriguez-Ruiz, A.; Lång, K.; Gubern-Merida, A.; Broeders, M.; Gennaro, G.; Clauser, P.; Helbich, T.H.; Chevalier, M.; Tan, T.; Mertelmeier, T.; et al. Stand-Alone Artificial Intelligence for Breast Cancer Detection in Mammography: Comparison with 101 Radiologists. JNCI J. Natl. Cancer Inst. 2019, 111, 916–922. [Google Scholar] [CrossRef]
  30. Ragab, D.A.; Sharkas, M.; Marshall, S.; Ren, J. Breast cancer detection using deep convolutional neural networks and support vector machines. PeerJ 2019, 2019, 1–23. [Google Scholar] [CrossRef] [PubMed]
  31. Kashif, M.; Malik, K.R.; Jabbar, S.; Chaudhry, J. Application of machine learning and image processing for detection of breast cancer. In Innovation in Health Informatics; Elsevier: Amsterdam, The Netherlands, 2020; pp. 145–162. [Google Scholar] [CrossRef]
  32. Hamed, G.; Marey, M.A.E.R.; Amin, S.E.S.; Tolba, M.F. Deep Learning in Breast Cancer Detection and Classification. In Advances in Intelligent Systems and Computing, Proceedings of the International Conference on Artificial Intelligence and Computer Vision, AICV 2020, Cairo, Egypt, 8–10 April 2020; Hassanien, A.E., Azar, A., Gaber, T., Oliva, D., Tolba, F., Eds.; Springer: Cham, Germany, 2020; Volume 1153. [Google Scholar] [CrossRef]
  33. Ak, M.F. A Comparative Analysis of Breast Cancer Detection and Diagnosis Using Data Visualization and Machine Learning Applications. Healthcare 2020, 8, 111. [Google Scholar] [CrossRef] [PubMed]
  34. Jeyanathan, J.S.; Shenbagavalli, A.; Venkatraman, B.; Menaka, M.; Anitha, J.; de Albuquerque, V.H.C. Analysis of Transform-Based Features on Lateral View Breast Thermograms. Circuits Syst. Signal Process. 2019, 38, 5734–5754. [Google Scholar] [CrossRef]
  35. Abdar, M.; Zomorodi-Moghadam, M.; Zhou, X.; Gururajan, R.; Tao, X.; Barua, P.D.; Gururajan, R. A new nested ensemble technique for automated diagnosis of breast cancer. Pattern Recognit. Lett. 2020, 132, 123–131. [Google Scholar] [CrossRef]
  36. Dhahri, H.; Al Maghayreh, E.; Mahmood, A.; Elkilani, W.; Faisal Nagi, M. Automated breast cancer diagnosis based on machine learning algorithms. J. Healthc. Eng. 2019, 2019, 1–11. [Google Scholar] [CrossRef] [PubMed]
  37. Khan, S.; Islam, N.; Jan, Z.; Din, I.U.; Rodrigues, J.J.P.C. A novel deep learning based framework for the detection and classification of breast cancer using transfer learning. Pattern Recognit. Lett. 2019, 125, 1–6. [Google Scholar] [CrossRef]
  38. McKinney, S.M.; Sieniek, M.; Godbole, V.; Godwin, J.; Antropova, N.; Ashrafian, H.; Shetty, S. International evaluation of an AI system for breast cancer screening. Nature 2020, 577, 89–94. [Google Scholar] [CrossRef]
  39. Memon, M.H.; Li, J.P.; Haq, A.U.; Memon, M.H.; Zhou, W.; Lacuesta, R. Breast cancer detection in the IOT health environment using modified recursive feature selection. Wirel. Commun. Mob. Comput. 2019, 2019, 1–19. [Google Scholar] [CrossRef] [Green Version]
  40. Ronoud, S.; Asadi, S. An evolutionary deep belief network extreme learning-based for breast cancer diagnosis. Soft Comput. 2019, 23, 13139–13159. [Google Scholar] [CrossRef]
  41. Ting, F.F.; Tan, Y.J.; Sim, K.S. Convolutional neural network improvement for breast cancer classification. Expert Syst. Appl. 2019, 120, 103–115. [Google Scholar] [CrossRef]
  42. Vijayarajeswari, R.; Parthasarathy, P.; Vivekanandan, S.; Basha, A.A. Classification of mammogram for early detection of breast cancer using SVM classifier and hough transform. Meas. J. Int. Meas. Confed. 2019, 146, 800–805. [Google Scholar] [CrossRef]
  43. Wu, N.; Phang, J.; Park, J.; Shen, Y.; Huang, Z.; Zorin, M.; Geras, K.J. Deep neural networks improve radiologists’ performance in breast cancer screening. IEEE Trans. Med. Imaging 2020, 39, 1184–1194. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Assiri, A.S.; Nazir, S.; Velastin, S.A. Breast Tumor Classification Using an Ensemble Machine Learning Method. J. Imaging 2020, 6, 39. [Google Scholar] [CrossRef]
  45. Li, L.N.; Ouyang, J.H.; Chen, H.L.; Liu, D.Y. A computer aided diagnosis system for thyroid disease using extreme learning machine. J. Med. Syst. 2012, 36, 3327–3337. [Google Scholar] [CrossRef] [PubMed]
  46. Sartakhti, J.S.; Zangooei, M.H.; Mozafari, K. Hepatitis disease diagnosis using a novel hybrid method based on support vector machine and simulated annealing (SVM-SA). Comput. Methods Programs Biomed. 2012, 108, 570–579. [Google Scholar] [CrossRef] [PubMed]
  47. Kumari, V.A.; Chitra, R. Classification of Diabetes Disease Using Support Vector Machine. Int. J. Eng. Res. Appl. 2013, 3, 1797–1801. [Google Scholar]
  48. Kaya, Y.; Uyar, M. A hybrid decision support system based on rough set and extreme learning machine for diagnosis of hepatitis disease. Appl. Soft Comput. J. 2013, 13, 3429–3438. [Google Scholar] [CrossRef]
  49. Wang, Z.; Yu, G.; Kang, Y.; Zhao, Y.; Qu, Q. Breast tumor detection in digital mammography based on extreme learning machine. Neurocomputing 2014, 128, 175–184. [Google Scholar] [CrossRef]
  50. Zheng, B.; Yoon, S.W.; Lam, S.S. Breast cancer diagnosis based on feature extraction using a hybrid of K-means and support vector machine algorithms. Expert Syst. Appl. 2014, 41 Pt 1, 1476–1482. [Google Scholar] [CrossRef]
  51. Prashanth, R.; Dutta Roy, S.; Mandal, P.K.; Ghosh, S. High-Accuracy Detection of Early Parkinson’s Disease through Multimodal Features and Machine Learning. Int. J. Med. Inform. 2016, 90, 13–21. [Google Scholar] [CrossRef] [PubMed]
  52. Chen, H.-L.; Wang, G.; Ma, C.; Cai, Z.-N.; Liu, W.-B.; Wang, S.-J. An efficient hybrid kernel extreme learning machine approach for early diagnosis of Parkinson׳s disease. Neurocomputing 2016, 184, 131–144. [Google Scholar] [CrossRef] [Green Version]
  53. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef] [PubMed]
  54. Liu, S.; Zheng, H.; Feng, Y.; Li, W. Prostate cancer diagnosis using deep learning with 3D multiparametric MRI. In Medical imaging 2017: Computer-Aided Diagnosis; International Society for Optics and Photonics: Bellingham, WA USA, 2017; Volume 10134, p. 1013428. [Google Scholar]
  55. Chen, M.; Hao, Y.; Hwang, K.; Wang, L.; Wang, L. Disease Prediction by Machine Learning Over Big Data From Healthcare Communities. IEEE Access 2017, 5, 8869–8879. [Google Scholar] [CrossRef]
  56. Khalid, S.; Khalil, T.; Nasreen, S. A survey of feature selection and feature extraction techniques in machine learning. In Proceedings of the 2014 Science and Information Conference (SAI), London, UK, 27–29 August 2014; pp. 372–378. [Google Scholar]
  57. Spencer, R.; Thabtah, F.; Abdelhamid, N.; Thompson, M. Exploring feature selection and classification methods for predicting heart disease. Digit. Health 2020, 6, 205520762091477. [Google Scholar] [CrossRef] [Green Version]
  58. Gabryel, M.; Damaševičius, R. The image classification with different types of image features. In Lecture Notes in Computer Science, Proceedings of the Artificial Intelligence and Soft Computing. ICAISC 2017, Zakopane, Poland, 11–15 June 2017; Springer: Cham, Germany, 2017; Volume 10245, pp. 497–506. [Google Scholar] [CrossRef]
  59. Jia, P.; Dai, J.; Pan, Y.; Zhu, M. Novel algorithm for attribute reduction based on mutual-information gain ratio. J. Zhejiang Univ. 2005, 40, 1041–1044. [Google Scholar]
  60. Ding, S.; Zhao, H.; Zhang, Y.; Xu, X.; Nie, R. Extreme learning machine: Algorithm, theory and applications. Artif. Intell. Rev. 2015, 44, 103–115. [Google Scholar] [CrossRef]
  61. Jiang, L.; Cai, Z.; Wang, D.; Jiang, S. Survey of Improving K-Nearest-Neighbor for Classification. In Proceedings of the Fourth International Conference on Fuzzy Systems and Knowledge Discovery (FSKD 2007), Haikou, China, 24–27 August 2007; pp. 679–683. [Google Scholar]
  62. Al-Aidaroos, K.M.; Bakar, A.A.; Othman, Z. Medical data classification with naive bayes approach. Inf. Technol. J. 2012, 11, 1166–1174. [Google Scholar] [CrossRef] [Green Version]
  63. Gallant, S.I. Perceptron-based learning algorithms. IEEE Trans. Neural Netw. 1990, 1, 179–191. [Google Scholar] [CrossRef]
  64. Hastie, T.; Rosset, S.; Zhu, J.; Zou, H. Multi-class AdaBoost. Stat. Interface 2009, 2, 349–360. [Google Scholar] [CrossRef] [Green Version]
  65. Chang, C.; Lin, C. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  66. Frank, E.; Hall, M.A.; Witten, I.H. The WEKA Workbench. Online Appendix for “Data Mining: Practical Machine Learning Tools and Techniques”, 4th ed.; Morgan Kaufmann: Burlington, MA, USA, 2016. [Google Scholar]
  67. PyCharm: The Python IDE for Professional Developers by JetBrains. Available online: https://www.jetbrains.com/pycharm/ (accessed on 18 October 2020).
  68. Wolberg, W.H.; Street, W.N.; Mangasarian, O.L. Breast Cancer Wisconsin (Diagnostic) Data Set. UCI Machine Learning Repository. 1992. Available online: https://archive.ics.uci.edu/ml/machine-learning-databases/breast-cancer-wisconsin/ (accessed on 3 February 2020).
  69. Gnana Sheela, K.; Deepa, S.N. Review on Methods to Fix Number of Hidden Neurons in Neural Networks. Math. Probl. Eng. 2013, 2013, 425740. [Google Scholar] [CrossRef] [Green Version]
  70. Kumar, P.M.; Lokesh, S.; Varatharajan, R.; Chandra Babu, G.; Parthasarathy, P. Cloud and IoT based disease prediction and diagnosis system for healthcare using Fuzzy neural classifier. Futur. Gener. Comput. Syst. 2018, 86, 527–534. [Google Scholar] [CrossRef]
  71. Kim-Soon, N.; Abdulmaged, A.I.; Mostafa, S.A.; Mohammed, M.A.; Musbah, F.A.; Ali, R.R.; Geman, O. A framework for analyzing the relationships between cancer patient satisfaction, nurse care, patient attitude, and nurse attitude in healthcare systems. J. Ambient Intell. Hum. Comput. 2021, 1–18. [Google Scholar] [CrossRef]
  72. Abdulkareem, K.H.; Mohammed, M.A.; Salim, A.; Arif, M.; Geman, O.; Gupta, D.; Khanna, A. Realizing an Effective COVID-19 Diagnosis System Based on Machine Learning and IOT in Smart Hospital Environment. IEEE Internet Things J. 2021. [Google Scholar] [CrossRef]
  73. Awan, M.J.; Rahim, M.S.M.; Salim, N.; Mohammed, M.A.; Garcia-Zapirain, B.; Abdulkareem, K.H. Efficient Detection of Knee Anterior Cruciate Ligament from Magnetic Resonance Imaging Using Deep Learning Approach. Diagnostics 2021, 11, 105. [Google Scholar] [CrossRef]
  74. Baltres, A.; Al Masry, Z.; Zemouri, R.; Valmary-Degano, S.; Arnould, L.; Zerhouni, N.; Devalland, C. Prediction of Oncotype DX recurrence score using deep multi-layer perceptrons in estrogen receptor-positive, HER2-negative breast cancer. Breast Cancer 2020, 27, 1007–1016. [Google Scholar] [CrossRef] [PubMed]
  75. Zemouri, R.; Zerhouni, N.; Racoceanu, D. Deep Learning in the Biomedical Applications: Recent and Future Status. Appl. Sci. 2019, 9, 1526. [Google Scholar] [CrossRef] [Green Version]
  76. Zemouri, R.; Devalland, C.; Valmary-Degano, S.; Zerhouni, N. Neural network: A future in pathology? Ann. Pathol. 2019, 39, 119–129. [Google Scholar] [CrossRef] [PubMed]
  77. Zemouri, R.; Omri, N.; Devalland, C.; Arnould, L.; Morello, B.; Zerhouni, N.; Fnaiech, F. Breast cancer diagnosis based on joint variable selection and constructive deep neural network. In Proceedings of the 2018 IEEE 4th Middle East Conference on Biomedical Engineering (MECBME), Tunis, Tunisia, 28–30 March 2018; pp. 159–164. [Google Scholar]
  78. Zemouri, R.; Omri, N.; Morello, B.; Devalland, C.; Arnould, L.; Zerhouni, N.; Fnaiech, F. Constructive deep neural network for breast cancer diagnosis. IFAC-Paper 2018, 51, 98–103. [Google Scholar] [CrossRef]
Figure 1. Components of the proposed architecture.
Figure 1. Components of the proposed architecture.
Diagnostics 11 00241 g001
Figure 2. Extreme learning machine.
Figure 2. Extreme learning machine.
Diagnostics 11 00241 g002
Figure 3. Experimental steps.
Figure 3. Experimental steps.
Diagnostics 11 00241 g003
Figure 4. Graphical representation of the ELM performance (standalone environment).
Figure 4. Graphical representation of the ELM performance (standalone environment).
Diagnostics 11 00241 g004
Figure 5. Graphical representation of the model performance (standalone environment).
Figure 5. Graphical representation of the model performance (standalone environment).
Diagnostics 11 00241 g005
Figure 6. Graphical representation of the ELM model performance (cloud computing environment).
Figure 6. Graphical representation of the ELM model performance (cloud computing environment).
Diagnostics 11 00241 g006
Figure 7. Comparison of the accuracy achieved when using standalone and cloud computing environments.
Figure 7. Comparison of the accuracy achieved when using standalone and cloud computing environments.
Diagnostics 11 00241 g007
Table 1. Summary of related work on disease diagnostics.
Table 1. Summary of related work on disease diagnostics.
AuthorsDiseaseYearDatasetClassifierAccuracy (Highest)
Li et al. [45]Thyroid2012Thyroid database from UCI repositoryPCA-ELMPCA-ELM = 98.1%
Sartakhti et al. [46]Hepatitis disease2012hepatitis B dataset UCI RepositorySVM-SASVM-SA = 96.2%
Kumari et al. [47]Diabetes2013Pima Indian diabetes dataset,SVMSVM = 78%
Kaya et al. [48]Hepatitis disease2013hepatitis B dataset from UCI RepositoryRough Set
ELM
Test/Train split
80/20 = 100%
Wang et al. [49]Breast Cancer2014482 mammographsELM, SVMELM = 83%
Zheng et al. [50]Breast Cancer2014Breast Cancer Wisconsin Dataset (BCWD)K-SVMK-SVM = 97.38%
Prashanth et al. [51]Parkinson’s Disease2016PPMI databaseNaïve Bayes, LR, Boosted Tree
Random Forest, SVM
Random Forest = 96.18%
Chen et al. [52]Parkinson’s disease2016PD dataset from UCI repositoryELM
K-ELM
Accuracy = 96.47%
Esteva et al. [53]Skin Cancer2017129,450 clinical images,Deep CNNCNN = 72.1%
Liu et al. [54]Prostate Cancer2017341 casesXMasNet (Based on CNN)XMasNet = 84%
Chen et al. [55]Disease Prediction201731,919 hospitalizedCNN-MDRPCNN-MDRP = 94.8%
Table 2. Four cases concerning the prediction.
Table 2. Four cases concerning the prediction.
CaseDefinition
True Positive (TP)A model forecasts the positive class correctly.
True Negative (TN)A model forecasts the negative class correctly.
False Positive (FP)A model forecasts the positive class correctly.
False Negative (FN)A model forecasts the negative class incorrectly.
Table 3. Evaluation metrics of classification performance.
Table 3. Evaluation metrics of classification performance.
FormulaExpected Value
A c c u r a c y = ( T P + F N ) T o t a l   n u m b e r   o f   s a m p l e s High
P r e c i s i o n = T P / ( T P + F P ) High
R e c a l l = T P / ( T P + F N ) High
K a p p a = P 0 P a 1 P a Value = 1 implies perfect agreement, and
Value < 1 implies a less perfect agreement
F S c o r e = 2 × R e c a l l × P r e c i s i o n ( R e c a l l + P r e c i s i o n ) Best Value is 1, and Worst Value is 0
NoteP0 is the observed agreement, and Pa is the expected agreement: y j is the anticipated output, and y ^ j is the predicted value.
Table 4. Dataset description.
Table 4. Dataset description.
S. No.Attribute NameDescription
1.IdId Number
2.DiagnosisThe diagnosis of breast tissues (M = malignant, B = Benign)
3.Radius_MeanMean of distances from the center to points on the perimeter
4.Texture_MeanStandard deviation of grayscale values
5.Perimeter_MeanMean size of the core tumor
6.Area_MeanMean area of the core tumor
7.Smoothness_MeanMean of local variation in radius lengths
8.Compactness_MeanMean of perimeter2/area − 1
9.Concavity_MeanMean of severity of concave portion of the contour
10.Concave points_meanMean for number of concave portions of the contour
11.Symmetry_mean
12.Fractal_dimension_meanMean for coastline approximation − 1
13.Radius_seStandard error for the mean of distances from the center to the points on the perimeter
14.Texture_seStandard error for standard deviation for grayscale values
15.Perimeter_se
16.Area_se
17.Smoothness_seStandard error for local variation in radius lengths
18.Compactness_seStandard error for perimeter2/area − 1
19.Concavity_seStandard error for severity of concave portions of the contour
20.Concave points_seStandard error for the number of concave portions of the contour
21.Symmetry_se
22.Fractal_dimension_seStandard error for coastline approximation − 1
23.Radius_worst“worst” or largest mean value for the mean of distances from the center to points on perimeter
24.Texture_worst“worst” or largest mean value for standard deviation of grayscale values
25.Perimeter_worst
26.Area_worst
27.Smoothness_worst“worst” or largest mean value for local variation in radius length
28.Compactness_worst“worst” or largest mean value for perimeter2/area − 1
29.Concavity_worst“worst” or largest mean value for severity of concave portions of the contour
30.Concave points_worst“worst” or largest mean value for number of concave portions of the contour
31.Symmetry_worst
32.Fractal_dimension_worst“worst” or largest mean value for coastline approximation − 1
Table 5. Evaluation metrics for extreme learning machine (ELM) with different hidden layer nodes in the standalone environment.
Table 5. Evaluation metrics for extreme learning machine (ELM) with different hidden layer nodes in the standalone environment.
Nodes in the Hidden Layer50100150200250
Accuracy0.93410.94510.95600.96920.9648
Kappa0.83020.79170.78480.60460.4379
Precision0.89470.89290.88510.79120.7294
Recall0.98550.98681.01.00.9118
F-score0.93790.93750.93900.88340.8105
Table 6. Performance of different machine learning models deployed in a standalone environment.
Table 6. Performance of different machine learning models deployed in a standalone environment.
AdaBoostKNNNBPerceptronSVMELM
Accuracy0.92980.90640.84800.83040.92980.9692
Kappa0.84600.79130.67680.66140.84470.6046
Precision0.93750.92110.80000.97650.94640.7912
Recall0.95450.93750.97960.75450.94641.000
F-score0.94590.92920.88070.85130.94640.8834
Table 7. Performance analysis of ELM deployed on a cloud computing environment with different numbers of hidden layer nodes. Best accuracy values are shown in bold.
Table 7. Performance analysis of ELM deployed on a cloud computing environment with different numbers of hidden layer nodes. Best accuracy values are shown in bold.
vCPU = 4 RAM = 16 GBELM (50)ELM (100)ELM (150)ELM (200)ELM (250)
Accuracy0.93630.94730.95160.96920.9648
Kappa0.79170.74280.72730.51710.5471
Precision0.89290.87500.87060.79520.8125
Recall0.98680.95890.97370.90410.9873
F-score0.93750.91500.91930.84620.8914
vCPU = 8 RAM = 32 GB
Accuracy0.94290.95160.96700.97800.9758
Kappa0.75670.73810.62970.52140.5190
Precision0.87640.84710.79760.76920.7857
Recall1.00001.00000.97100.97220.9167
F-score0.93410.91720.87580.85890.8462
vCPU = 16 RAM = 64 GB
Accuracy0.93630.95380.96920.97580.9824
Kappa0.87860.71620.47360.50150.3889
Precision0.92590.88370.70450.71430.7528
Recall1.00000.96200.96880.95240.9054
F-score0.96150.92120.81580.81630.8221
vCPU = 36 RAM = 60 GB
Accuracy0.93850.95820.97140.98020.9868
Kappa0.80640.67690.63410.47340.4302
Precision0.85070.80490.78950.80680.7326
Recall0.98280.98510.93750.91030.9130
F-score0.91200.88590.85720.85540.8129
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lahoura, V.; Singh, H.; Aggarwal, A.; Sharma, B.; Mohammed, M.A.; Damaševičius, R.; Kadry, S.; Cengiz, K. Cloud Computing-Based Framework for Breast Cancer Diagnosis Using Extreme Learning Machine. Diagnostics 2021, 11, 241. https://doi.org/10.3390/diagnostics11020241

AMA Style

Lahoura V, Singh H, Aggarwal A, Sharma B, Mohammed MA, Damaševičius R, Kadry S, Cengiz K. Cloud Computing-Based Framework for Breast Cancer Diagnosis Using Extreme Learning Machine. Diagnostics. 2021; 11(2):241. https://doi.org/10.3390/diagnostics11020241

Chicago/Turabian Style

Lahoura, Vivek, Harpreet Singh, Ashutosh Aggarwal, Bhisham Sharma, Mazin Abed Mohammed, Robertas Damaševičius, Seifedine Kadry, and Korhan Cengiz. 2021. "Cloud Computing-Based Framework for Breast Cancer Diagnosis Using Extreme Learning Machine" Diagnostics 11, no. 2: 241. https://doi.org/10.3390/diagnostics11020241

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop