Next Article in Journal
Telemedicine during COVID-19 Crisis and in Post-Pandemic/Post-Vaccine World—Historical Overview, Current Utilization, and Innovative Practices to Increase Utilization
Next Article in Special Issue
Auxiliary Segmentation Method of Osteosarcoma in MRI Images Based on Denoising and Local Enhancement
Previous Article in Journal
Quantitative MRI of Pancreatic Cystic Lesions: A New Diagnostic Approach
Previous Article in Special Issue
Intelligent Deep-Learning-Enabled Decision-Making Medical System for Pancreatic Tumor Classification on CT Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Deep Stacked Sparse Autoencoder Based Osteosarcoma Detection and Classification Model

by
Bahjat Fakieh
1,
Abdullah S. AL-Malaise AL-Ghamdi
1,2,3 and
Mahmoud Ragab
3,4,5,6,*
1
Information Systems Department, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia
2
Information Systems Department, HECI School, Dar Alhekma University, Jeddah 22246, Saudi Arabia
3
Center of Excellence in Smart Environment Research, King Abdulaziz University, Jeddah 21589, Saudi Arabia
4
Information Technology Department, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia
5
Department of Mathematics, Faculty of Science, Al-Azhar University, Naser City, Cairo 11884, Egypt
6
Centre for Artificial Intelligence in Precision Medicines, King Abdulaziz University, Jeddah 21589, Saudi Arabia
*
Author to whom correspondence should be addressed.
Healthcare 2022, 10(6), 1040; https://doi.org/10.3390/healthcare10061040
Submission received: 7 May 2022 / Revised: 30 May 2022 / Accepted: 30 May 2022 / Published: 2 June 2022
(This article belongs to the Special Issue Advances of Decision-Making Medical System in Healthcare)

Abstract

:
Osteosarcoma is a kind of bone cancer which generally starts to develop in the lengthy bones in the legs and arms. Because of an increase in occurrence of cancer and patient-specific treatment options, the detection and classification of cancer becomes a difficult process. The manual recognition of osteosarcoma necessitates expert knowledge and is time consuming. An earlier identification of osteosarcoma can reduce the death rate. With the development of new technologies, automated detection models can be exploited for medical image classification, thereby decreasing the expert’s reliance and resulting in timely identification. In recent times, an amount of Computer-Aided Detection (CAD) systems are available in the literature for the segmentation and detection of osteosarcoma using medicinal images. In this view, this research work develops a wind driven optimization with deep transfer learning enabled osteosarcoma detection and classification (WDODTL-ODC) method. The presented WDODTL-ODC model intends to determine the presence of osteosarcoma in the biomedical images. To accomplish this, the osteosarcoma model involves Gaussian filtering (GF) based on pre-processing and contrast enhancement techniques. In addition, deep transfer learning using a SqueezNet model is utilized as a featured extractor. At last, the Wind Driven Optimization (WDO) algorithm with a deep-stacked sparse auto-encoder (DSSAE) is employed for the classification process. The simulation outcome demonstrated that the WDODTL-ODC technique outperformed the existing models in the detection of osteosarcoma on biomedical images.

1. Introduction

Osteosarcoma is considered to be an aggressive bone malignancy which occurs often in the extremes of adolescents and children with a terminal diagnosis [1]. The occurrence of osteosarcoma is assumed to be the most frequent amongst all of the basic malicious tumors across the globe, achieving three cases per million people per year, including a female male ratio of 1:1.5. Meanwhile, the chance of living with this tumor is lesser. The five-year survival rate of patients affected by osteosarcoma was below 20% before the 1980s [2]. This type of tumor consists of a higher extent of malignity and is vulnerable to the lungs and metastases; as such, its medication seems to be very complex. Initial identification and medication could improve the survival rate of patients and decrease the chances of amputation [3]. In clinical prognoses, MRI images cannot pose significant biological and radiation hazards to the tissue in the course of inspection and are very apparent in tissue elements, such as the blood vessels and tumors [4]. This is why we usually utilize MRI images for detecting osteosarcomas. In the recent identification procedure of osteosarcoma MRI images, patients having osteosarcoma would produce a huge volume of image data, since the ratio of useful images is less. Each osteosarcoma patient will produce 600 to 700 images, but the fact is just 10 to 20 images are useful for medication, as incomplete statistics are reported [5]. Nowadays, preliminary processing and screening of raw pictures are conducted by physicians manually [6]. This procedure includes more material sources and labor forces. Meanwhile, there is no standard interpretation the diagnostic features of osteosarcoma.
The computer-aided diagnosis (CAD) of osteosarcoma metastasizes a prominent research focus in the literature, as it is helpful to the physician in detecting the nodule in a lung of patients at the prior level [7]. There are several methodologies that have been suggested in recent decades. An effective ML tool is the Convolutional Neural Network (CNN), as it can be trained on labeled data or images which are passed into the network for attaining output. It is mainly used for image classification. It derives the features and learns essential features in the convolutional layers, then categorizes the output as the presence or absence of osteosarcoma metastasizes by the fully connected layers [8]. This learning technique, known as a supervised learning algorithm, can be employed for training the CNN model to accomplish effective results by the use of high-quality images in the dataset [9,10].
This study develops a wind driven optimization with deep transfer learning enabled osteosarcoma detection and classification (WDODTL-ODC) method. The presented WDODTL-ODC model intends to determine the presence of osteosarcoma in the biomedical images. To accomplish this, the osteosarcoma model involves Gaussian filtering (GF) based on pre-processing and contrast enhancement techniques, and is followed by deep transfer learning using SqueezNet model utilized as a feature extractor. At last, the WDO algorithm with deep-stacked sparse auto-encoder (DSSAE) is employed for the classification process. The simulation outcome of the WDODTL-ODC technique is tested using benchmark biomedical images.

2. Related Works

Shen et al. [11] suggest that an osteosarcoma-supported segmentation methodology depends on a directed combined bilateral network (OSGABN) that improvises the segmentation precision of the model and hugely diminishes the variable scale, efficiently easing the issues which are discussed above. The rapid bilateral segmentation network (FaBiNet) is utilized for the division of images. It is a high accuracy method that has a detailed branch, which takes a lightweight semantic branch and low-level data, and which takes high level semantic contexts. Varalakshmi et al. [12] recommend an original route for classifying distinct osteosarcoma forms having a higher precision by utilizing histology image (Microscopic bone image). The examination of bone utilizing histology is a prolonged and tiresome procedure. During this article, the eXtreme Gradient Boosting (XGBoost) system is utilized for classifying osteosarcoma.
The researchers in [13] suggest an original CNN structure made up of Concatenation of numerous Networks, termed as C-Net, for categorizing biomedical images. The method inculcates numerous CNNs involving Inner, Outer, and Middle. The initial 2 portion of the structure consists of 6 networks which act as feature extractors for feeding into the Inner network to categorize the images with regard to benignancy and malignancy. Asito et al. [14] suggest a computer aided prognosis system depends on CNNs for the recognition of osteosarcoma on bone radiographs. The CNN must denote areas of the image which can consist of tumors. For indicating such areas on the image, we suggest dividing image into windows and categorizing them individually with the help of CNN. Methods for pre-processing, like labeling and window exclusion, are recommended. In the suggested system comparison is made for 2 CNNs.
Asmaria et al. [15] mainly focus to categorize the cell viability of osteosarcoma’s dataset with hematoxylin and eosin (H&E) stained. The CNN structure incorporated 6 convolution layers, fully linked layers, and max pooling layers to feature extraction. Data augmentation is utilized for boosting the executions. Sharma et al. [16] discover the better appropriate edge detection system once the 2 feature sets one without hog and another with hog is ready. For testing the effectiveness of such feature sets, the Random Forest, 2 ML methods, and support vector machine (SVM) and, are used.
Rajagoal et al. [17] reveals several methods of medical image processing and DL and implies them to identify and categorize tumors as malignant or benign. Approaches utilized inculcates image pre-processing utilizing filtering methodologies, K-means edge detection, and segmentation for identifying cancerous areas in Computer Tomography (CT) images for enchondroma, osteochondroma, and Parosteal osteosarcoma, forms of bone cancer. Once the process of segmentation of the tumor is done, categorization of cancerous cells and benign is made by using a DL method related CNN classifier. Few works have been available in the literature for osteosarcoma classification and design of automated classification models need to be explored more. In addition, the existing models do not focus on the hyper-parameter selection process, which mainly influences the performance of the classification model. Particularly, the hyper-parameters such as epoch count, batch size, and learning rate selection are essential to attain effectual outcome. Since the trial and error method for hyper-parameter tuning is a tedious and erroneous process, metaheuristic algorithms can be applied. Therefore, in this work, we employ WDO algorithm for the parameter selection of the DSSAE model.

3. The Proposed Model

In this study, a novel WDODTL-ODC approach was established to determine the presence of osteosarcoma in biomedical images. The WDODTL-ODC technique comprises GF-based pre-processing and contrast enhancement techniques. In addition, deep transfer learning using SqueezNet model is utilized as feature extractor. At last, the WDO algorithm with DSSAE is employed for classification process. Figure 1 depicts the overall process of WDODTL-ODC technique.

3.1. Image Pre-Processing

At the primary level, the WDODTL-ODC technique comprises GF-based pre-processing and contrast enhancement techniques. GF is a method that decreases pixel difference by weighted average for image smoothing from different applications. But, the low--pass filter could not preserve image detail, for example, the textures and edges. Then, linear translation—variant function f defines the abovementioned filtering procedure as follows [18]:
f ( p ) = q K p , q ( Q ) P q
In Equation (1), K p , q signifies the pixel q centred at pixel p in filter kernel K , and Q and P denote guidance and input images, correspondingly. For instance, the kernel of Bilateral Filter (BF) is defined as:
K p , q ( Q ) = 1 n   exp   ( p q 2 σ s 2 ) · exp   ( P p Q q 2 σ r 2 )
The exponential distribution function is utilized in Equation (2) to compute the effect of spatial distance by exp   ( p q 2 σ r 2 ) , and exp   ( P p Q q 2 σ r 2   ) describes the influence of pixel intensity range. If Q and P are identical, Equation (2) is simplified as a single image smoothing form.

3.2. Feature Extractor

Next to image pre-processing, deep transfer learning using SqueezNet model is utilized as feature extractor. SqueezeNet is an efficient and lightweight CNN model [19]. With SqueezeNet, we achieve a 50× reduction in model size compared to AlexNet, while meeting or exceeding the top-1 and top-5 accuracy of AlexNet. The SqueezetNet model achieves effective performance with less number of parameters. There are two important parts of CNN, namely feature extraction and classification. The extracted features were utilized for the accurate classification of images. Specifically, these 2 parts of CNN achieve the initial function of CNNs. The feature extracting part of CNN contains convolutional and sampling layers. The filter of convolutional layer was utilized for diminishing the noise in the images; afterward, the feature of images was improved. The convolutional procedure was complete amongst the upper layer feature vector and the presentation layer convolutional kernel layer. The activation function of CNN lastly finalizes the convolutional procedure computations. The efficacy of training a NN was obvious when utilizing the cost function that represents the proportion amongst the trained instance and the reached output. Figure 2 illustrates the structure of SqueezeNet technique.
Z = 1 m [ x   l n   a ( 1 x ) ln ( 1 β ) ]
whereas m represents the amount of trained data, x signifies the predictable value, and β denotes the actual value in the resultant layer.
The activation function roles a vital part in the classifier procedure with transmission kernel size and weighted the resultant of CNN method. The ReLU activation function was between the generally utilized activation function. It can be approximately utilized from every CNN technique for setting every negative value equivalent to 0. These zero settings inhibit many nodes in participate from the learning procedure. Another function, namely LReLU and ELU that offer smaller negative values was rarely utilized from classifier approaches. ReLU activation function displays optimum outcomes than LReLU activation function from the classifier which is utilized in our classifier method. The provided formula mathematically represents the ReLU activation function.
R e L U ( x ) = m a x ( 0 ;   x )
It is mostly employed in an embedded setting; it includes different methodologies of compression model. For instance, 3 × 3 convolutional kernel in the presented model is substituted with 1 × 1 convolutional kernel. With this method, the parameter count for single convolutional function is minimized by the factor of 9. Additionally, the 3 × 3 convolutional kernel is decreased as well as down-sampling is delayed in the network layer. Consequently, the presented model decreases the number of trained variables and the computation efforts. Therefore, it is feasible to set up SqueezeNet in memory-limited hardware device. In comparison to current AI models, we have observed that SqueezeNet has the minimum parameter count; therefore, it is the right selection for robot vacuum applications. Unfortunately, the model size (viz., 6.1 MB) is greater when compared to the memory space presented in robot vacuums.

3.3. Image Classification

At the final stage, the WDO algorithm with DSSAE is employed for classification process. The DDSAE model is applied in this work as it learns non-linear transformations with a non-linear activation function and multiple layers. Besides, it is effective in eLearning different layers with an auto-encoder (AE) instead of learning one huge transformation with PCA. The essential component of DSSAE is the AE [20], viz., is a standard NN that learns to map the input Y to output Z . The AE is classified into an encoding part ( W Y , B Y ) that map input Y to code C , and a decoding part ( W Z , B Z ) that map code C to recreated dataset Z ,
  Y   e n c o d e r   C   d e c o d e r   Z  
Given that output, Z is equivalent to the input Y .
Whereby the encoded part is weighted W Y and bias B Y , and the decoded part is with weight W Z and bias B Z .
C = g L S ( W Y Y + B Y )
Z = g L S ( W Z C + B Z )
From the equation, the output Z is an approximation of input Y , and g L S indicates the log sigmoid function:
g L S ( x ) = 1 1 + exp ( x )
The SAE is variant of A E . In order to minimalize the errors among the input vector Y as well as output Z , the loss function of AE is assumed as follows:
I A E ( W Y ,   W Z , B Y , B Z ) = 1 N S Z Y 2
In Equation (9), N s indicates the number of training instances. From Equations (6) and (7), the output Z is formulated by using following equation
Z = g A E ( Y | W Y ,   W Z , B Y , B Z )
In Equation (10), g A E indicates the abstract of AE; Thus, Equation (9) is written by
I A E ( W Y ,   W Z , B Y , B Z ) = 1 N S g A E ( Y | W Y ,   W Z , B Y , B Z ) Y 2  
Actually, L 2 regularization term Γ w of the weight ( W Y ,   W Z ) and Γ s regularization terms of the sparsity constraints are determined to evade trivial or overcomplete mapping.
l S A E ( W Y , W Z , B Y , B Z ) = 1 N S g A E ( Y | W Y , W Z , B Y , B Z ) Y 2 + c s × Γ s + c w × Γ w
In Equation (12), c s indicates the sparsity regulation factor, c w represent the weighted regulation factors. The sparsity regularization term is determined by:
Γ s = j = 1 | C | g K L ( ρ , ρ ^ j ) = j = 1 | C | ρ   log   ρ ρ ^ j + ( 1 ρ ) log   1 ρ 1 ρ ^ j
In Equation (14), g K L represent the Kullback-Leibler divergence function, | C | denotes the element count of internal code output C. ρ ^ indicates the j t h average activation value on each N S trained sample and ρ denotes chosen value, called sparsity proportion factor. The Γ w weight regularization term is determined by
Γ w = 1 2 × W Y W Z 2 2 .  
The training process is fixed to scaled conjugate gradient descent (SCGD) technique. It can be utilized SAE as structure block and generate the last DSSAE classification with subsequent 3 functions: (i) It can contain PZM layer, input layer, vectorization layer, and pre-processing layer; (ii) It can be stack 4 SAEs with several amounts of hidden neurons; (iii) It can be attached softmax layer at the end of AI method.

3.4. Parameter Optimization

To tune the parameters related to the DSSAE model, the WDO algorithm has been exploited. Zikri Bayraktar, in 2010, first developed the concept of WDO [21]. It is inspired by the natural wind movement concept that serves as a stabilizer for equalizing the air pressure inequality. Wind blow from a higher-to lower-pressure region, along with a velocity viz. directly proportionate to pressure gradient (the high the pressure variance, the strong the wind blows) [22]. The major concept behind the proposed algorithm is Newton’s 2nd law of motion:
ρ a = F i
In Equation (15), the acceleration vector can be expressed by the term a , the air density for a component with smaller amount is represented as ρ , and the force acting on the mass is denoted by F j . Temperature, Air pressure, and density are interrelated with the formula based on the ideal gas law formulated in the following
P = ρ R T
In Equation (16) ,   R , and T are correspondingly represented as pressure, universal gas constant, and temperature.
In Equation (15), the important force that causes the wind blows in a direction to diverge in that direction is classified into frictional force ( F F ) , pressure gradient force ( F P G ) , coriolis force ( F C ) and gravitational force (F).
F P G = P δ V
F C = 2 Ω × u
F G = ρ δ g
F F = ρ α u ,
From the above equations, the pressure gradient is represented as P , δ V indicates a small volume of air, Ω signifies the rotation of earth, u denotes the wind velocity vector and g represents the gravitational acceleration. The fusion of Equations (15) and (17)–(20) generate an expression given in the following:
ρ u Δ t = P δ V 2 Ω × u + ρ δ V g ρ α u .  
The unit step of time, Δ t is equivalent to 1. The fusion of Equations (16) and (21) generated an equation represented in the Equation (22):
u n e w = g x o l d + ( 1 α ) u o l d + | 1 P   max   P o l d | R T ( x   max   x o l d ) c u o l d o t h e r   d i m P o l d
Let u n e w and u o l d be the upgraded velocity and the present velocity, correspondingly; x o l d and x max   indicates the existing position and the maximum pressure position of the air pack, correspondingly; P max   and P o l d indicates the maximal pressure and the pressure at the existing position, correspondingly; T indicates the temperature; and R ,   α , and c denote constant.
In Equation (22) the pressure value is generally higher. Thus, the velocity estimation is also higher. This causes the efficiency level of the WDO to reduce. Equation (22) is reordered into the subsequent formula:
u n e w = g x o l d + ( 1 α ) u o l d + | 1 1 k | R T ( x   max   x o l d ) c u o l d o t h e r   d i m k .  
In Equation (23), k signifies the ranking among each air parcel ( k = 1 , 2 , ,   20 ) . The subsequent formula is utilized for updating the location of air pack:
x n e w = x o l d + u n e w × Δ t .  
The WDO grows a fitness function (FF) for achieving higher classifier efficiency. It defines a positive integer for denoting the best efficiency of candidate solution. During this study, the minimized classification error rate was regarded as FF as existing in Equation (25).
f i t n e s s ( x i ) = C l a s s i f i e r E r r o r R a t e ( x i ) = n u m b e r   o f   m i s c l a s s i f i e d   s a m p l e s T o t a l   n u m b e r   o f   s a m p l e s × 100  

4. Results and Discussion

The performance validation of the WDODTL-ODC model is tested using a benchmark dataset [23] comprising 1144 images under three classes. It includes 345 images under viable tumor (VT), 263 images under non-Viable Tumor (NVT), and 536 images under Non-Tumor (NT) class. A few sample images are depicted in Figure 3.
Figure 4 demonstrates the confusion matrices produced by the WDODTL-ODC model on distinct sizes of data. With 80% of TR data, the WDODTL-ODC model has categorized 261, 213, and 426 samples under VT, NVT, and NT classes respectively. Meanwhile, with 20% of TS data, the WDODTL-ODC approach has categorized 83, 45, and 100 samples under VT, NVT, and NT classes correspondingly. Eventually, with 70% of TR data, the WDODTL-ODC system has categorized 231, 173, and 385 samples under VT, NVT, and NT classes correspondingly.
Table 1 and Figure 5 represent the detailed classifier results of the WDODTL-ODC model with TR data of 80% and TS data of 20%. The results implied that the WDODTL-ODC model has effectually categorized the images with maximum outcome. For instance, the WDODTL-ODC model has identified samples under VT class with a c c u y , p r e c n , r e c a l , F s c o r e , MCC, and G m e a n of 99.45%, 98.49%, 99.62%, 99.05%, 98.67%, and 99.50% respectively. Along with that, the WDODTL-ODC approach has identified samples under NVT class with a c c u y , p r e c n , r e c a l , F s c o r e , MCC, and G m e a n of 98.91%, 97.71%, 97.71%, 97.70%, 96.99%, and 98.49% correspondingly. Moreover, the WDODTL-ODC algorithm has identified samples under NT class with a c c u y , p r e c n , r e c a l , F s c o r e , MCC, and G m e a n of 98.36%, 98.61%, 97.93%, 98.27%, 96.71%, and 98.34% correspondingly.
Table 2 and Figure 6 demonstrate the detailed classifier outcomes of the WDODTL-ODC algorithm with TR data of 80% and TS data of 20%. The results outperformed that the WDODTL-ODC algorithm has effectually categorized the images with maximal outcomes.
For instance, the WDODTL-ODC system has identified samples under VT class with a c c u y , p r e c n , r e c a l , F s c o r e , MCC, and G m e a n of 99%, 98.30%, 98.30%, 98.30%, 97.59%, and 98.79% correspondingly. Besides, the WDODTL-ODC system has identified samples under NVT class with a c c u y , p r e c n , r e c a l , F s c o r e , MCC, and G m e a n of 99.25%, 99.43%, 97.19%, 98.30%, 97.83%, and 98.51% correspondingly. Furthermore, the WDODTL-ODC approach has identified samples under NT class with a c c u y , p r e c n , r e c a l , F s c o r e , MCC, and G m e a n of 99%, 98.47%, 99.48%, 98.97%, 98%, and 99.01% correspondingly.
Figure 7 offers the accuracy and loss graph analysis of the WDODTL-ODC technique on distinct set of TR/TS datasets. The outcomes outperformed that the accuracy value tends to improve and loss value tends to lesser with enhance in epoch count. It is also observed that the training loss is lower and validation accuracy is high on distinct sets of TR/TS datasets.
Figure 8 demonstrates the classifier results of the ODCNN-RFIC technique on distinct sets of TR/TS datasets. Figure 8a,c establish the precision recall analysis of the ODCNN-RFIC model under 80:20 and 70:30 datasets. By observing the figure, it can be noticed that the ODCNN-RFIC model has accomplished maximal precision-recall performance under all classes. Lastly, Figure 8b,d illustrate the ROC examination of the ODCNN-RFIC model under 80:20 and 70:30 datasets. The figure indicated that the MFODBN-MDC model has obtained higher ROC under VT, NVT, and NT classes correspondingly.
A detailed comparative analysis of the results offered by the WDODTL-ODC model with recent models is provided in Table 3 [24,25]. Figure 9 reports a brief a c c u y examination of the WDODTL-ODC model with existing models. The figure implied that the handcrafted feature, EfficientNet-B0, and VGG-16 models have resulted to least classification a c c u y of 95.50%, 95.20%, and 95.11% respectively. At the same time, the Xception, EfficientNet-B0-Handcrafted, Xception-Handcrafted, and ResNet-50 models have reached slightly enhanced a c c u y values of 97.22%, 97.74%, 97.10%, and 97.09% respectively. Though the MobileNet-v2 model has shown reasonable performance with a c c u y of 98.24%, the WDODTL-ODC model has accomplished maximum a c c u y of 99.22%.
Figure 10 defines a brief p r e c n inspection of the WDODTL-ODC approach with existing techniques. The figure implied that the handcrafted feature, EfficientNet-B0, and VGG-16 models have resulted in minimal classification p r e c n of 98.34%, 95.47%, and 97.61% respectively. At the same time, the Xception, EfficientNet-B0-Handcrafted, Xception-Handcrafted, and ResNet-50 models have reached somewhat improved p r e c n values of 94.84%, 98.04%, 94.85%, and 98.32% correspondingly. But, the MobileNet-v2 system has shown reasonable performance with p r e c n of 97.39%, the WDODTL-ODC algorithm has been able higher p r e c n of 98.86%.
Figure 11 illustrates a brief r e c a l investigation of the WDODTL-ODC system with existing techniques. The figure implied that the handcrafted feature, EfficientNet-B0, and VGG-16 models have resulted in least classification r e c a l of 98.17%, 97.37%, and 97.30% correspondingly. Also, the Xception, EfficientNet-B0-Handcrafted, Xception-Handcrafted, and ResNet-50 models have attained somewhat enhanced r e c a l values of 95.85%, 98.39%, 96.64%, and 94.16% correspondingly. Eventually, the MobileNet-v2 algorithm has revealed reasonable performance with r e c a l of 98.33%, the WDODTL-ODC methodology has been able to maximal r e c a l of 98.69%.
Figure 12 showcases a brief F s c o r e analysis of the WDODTL-ODC algorithm with existing models. The figure exposed that the handcrafted feature, EfficientNet-B0, and VGG-16 approaches have resulted in least classification F s c o r e of 98.51%, 95.97%, and 95.30% correspondingly. Likewise, the Xception, EfficientNet-B0-Handcrafted, Xception-Handcrafted, and ResNet-50 models have reached slightly higher F s c o r e values of 95.71%, 95.01%, 96.63%, and 96.71% correspondingly. But the MobileNet-v2 approach has outperformed reasonable performance with F s c o r e of 97.98%, the WDODTL-ODC algorithm has accomplished superior F s c o r e of 98.77%. The detailed results and discussion confirmed the effectual outcomes of the WDODTL-ODC model over recent models in terms of different evaluation measures. Therefore, the proposed model can be employed for effectual osteosarcoma classification in real time investigation. The enhanced performance of the proposed model is due to the characteristics of SqueezeNet and optimal parameter tuning process using the WDO algorithm.

5. Conclusions

In this study, a novel WDODTL-ODC technique has been developed to determine the presence of osteosarcoma in biomedical images. The WDODTL-ODC technique comprises GF based pre-processing and contrast enhancement techniques. In addition, deep transfer learning using SqueezNet model is utilized as feature extractor. At last, the WDO algorithm with DSSAE is employed for classification process. The simulation outcome of the WDODTL-ODC technique is tested using benchmark biomedical images. The simulation outcome demonstrated that the WDODTL-ODC technique outperformed the existing models in the detection of osteosarcoma on biomedical images. Thus, the WDODTL-ODC technique has been exploited for proper identification of osteosarcoma in biomedical images. In future, hybrid DL models can be applied to improve the detection efficiency of the WDODTL-ODC model.

Author Contributions

Funding acquisition, B.F.; Investigation, B.F. and M.R.; Methodology, M.R.; Project administration, A.S.A.-M.A.-G.; Software, B.F.; Supervision, A.S.A.-M.A.-G.; Validation, M.R.; Visualization, A.S.A.-M.A.-G.; Writing—original draft, B.F.; Writing—review & editing, M.R. All authors have read and agreed to the published version of the manuscript.

Funding

This project was funded by the Deanship of Scientific Research (DSR) at King Abdulaziz University, Jeddah, Saudi Arabia, under grant no. (G: 148-611-1441).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable to this article as no datasets were generated during the current study.

Acknowledgments

This project was funded by the Deanship of Scientific Research (DSR) at King Abdulaziz University, Jeddah, Saudi Arabia, under grant no. (G: 148-611-1441). The authors, therefore, acknowledge with thanks DSR for technical and financial support.

Conflicts of Interest

The authors declare that they have no conflict of interest. The manuscript was written through contributions of all authors. All authors have given approval to the final version of the manuscript.

References

  1. Anisuzzaman, D.M.; Barzekar, H.; Tong, L.; Luo, J.; Yu, Z. A deep learning study on osteosarcoma detection from histological images. Biomed. Signal Processing Control. 2021, 69, 102931. [Google Scholar] [CrossRef]
  2. Han, Z.; Yi, J.; Yang, Y.; Li, D.; Peng, C.; Long, S.; Peng, X.; Shen, Y.; Liu, B.; Qiao, L. SERS and MALDI-TOF MS based plasma exosome profiling for rapid detection of osteosarcoma. Analyst 2021, 146, 6496–6505. [Google Scholar] [CrossRef]
  3. Makielski, K.M.; Donnelly, A.J.; Khammanivong, A.; Scott, M.C.; Ortiz, A.R.; Galvan, D.C.; Tomiyasu, H.; Amaya, C.; Ward, K.A.; Montoya, A.; et al. Development of an exosomal gene signature to detect residual disease in dogs with osteosarcoma using a novel xenograft platform and machine learning. Lab. Investig. 2021, 101, 1585–1596. [Google Scholar] [CrossRef]
  4. Tang, H.; Sun, N.; Shen, S. Improving generalization of deep learning models for diagnostic pathology by increasing variability in training data: Experiments on osteosarcoma subtypes. J. Pathol. Inform. 2021, 12, 30. [Google Scholar] [CrossRef] [PubMed]
  5. Badashah, S.J.; Basha, S.S.; Ahamed, S.R.; Subba Rao, S.P.V. Fractional-Harris hawks optimization-based generative adversarial network for osteosarcoma detection using Renyi entropy-hybrid fusion. Int. J. Intell. Syst. 2021, 36, 6007–6031. [Google Scholar] [CrossRef]
  6. Mahore, S.; Bhole, K.; Rathod, S. Machine Learning approach to classify and predict different Osteosarcoma types. In Proceedings of the 2021 8th International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 26–27 August 2021; pp. 641–645. [Google Scholar]
  7. Pan, L.; Wang, H.; Wang, L.; Ji, B.; Liu, M.; Chongcheawchamnan, M.; Yuan, J.; Peng, S. Noise-reducing attention cross fusion learning transformer for histological image classification of osteosarcoma. Biomed. Signal Processing Control. 2022, 77, 103824. [Google Scholar] [CrossRef]
  8. Chen, Y.; Liu, R.; Wang, W.; Wang, C.; Zhang, N.; Shao, X.; He, Q.; Ying, M. Advances in targeted therapy for osteosarcoma based on molecular classification. Pharmacol. Res. 2021, 169, 105684. [Google Scholar] [CrossRef] [PubMed]
  9. Pereira, H.M.; Leite Duarte, M.E.; Ribeiro Damasceno, I.; de Oliveira Moura Santos, L.A.; Nogueira-Barbosa, M.H. Machine learning-based CT radiomics features for the prediction of pulmonary metastasis in osteosarcoma. Br. J. Radiol. 2021, 94, 20201391. [Google Scholar] [CrossRef]
  10. Wu, J.; Yang, S.; Gou, F.; Zhou, Z.; Xie, P.; Xu, N.; Dai, Z. Intelligent Segmentation Medical Assistance System for MRI Images of Osteosarcoma in Developing Countries. Comput. Math. Methods Med. 2022, 2022, 7703583. [Google Scholar] [CrossRef] [PubMed]
  11. Shen, Y.; Gou, F.; Dai, Z. Osteosarcoma MRI Image-Assisted Segmentation System Base on Guided Aggregated Bilateral Network. Mathematics 2022, 10, 1090. [Google Scholar] [CrossRef]
  12. Varalakshmi, P.; Priyamvadan, A.V.; Rajakumar, B.R. Predicting Osteosarcoma using eXtreme Gradient Boosting Model. In Proceedings of the 2022 International Conference on Advances in Computing, Communication and Applied Informatics (ACCAI), Chennai, India, 28–29 January 2022; pp. 1–6. [Google Scholar]
  13. Barzekar, H.; Yu, Z. C-Net: A reliable convolutional neural network for biomedical image classification. Expert Syst. Appl. 2022, 187, 116003. [Google Scholar] [CrossRef]
  14. Asito, L.Y.; Pereira, H.M.; Nogueira-Barbosa, M.H.; Tinós, R. Detection of Osteosarcoma on Bone Radiographs Using Convolutional Neural Networks. In Proceedings of the 2021 Congresso Brasileiro de Inteligência Computacional, Joinville, Brasil, 3–6 October 2021. [Google Scholar]
  15. Asmaria, T.; Mayasari, D.A.; Heryanto, M.A.; Kurniatie, M.; Wati, R.; Aurellia, S. Osteosarcoma Classification using Convolutional Neural Network. In Proceedings of the 2021 International Conference on Computer, Control, Informatics and Its Applications, virtual, 5–6 October 2021; pp. 26–30. [Google Scholar]
  16. Sharma, A.; Yadav, D.P.; Garg, H.; Kumar, M.; Sharma, B.; Koundal, D. Bone Cancer Detection Using Feature Extraction Based Machine Learning Model. Comput. Math. Methods Med. 2021, 2021, 7433186. [Google Scholar] [CrossRef]
  17. Rajagopal, S.; Kanimozhi, S.; Chakrabarti, A.; Velev, D.G. Convolution Neural Network Based Bone Cancer Detection. SPAST Abstr. 2021, 1. [Google Scholar]
  18. Cheng, S.W.; Lin, Y.T.; Peng, Y.T. A Fast Two-Stage Bilateral Filter Using Constant Time O (1) Histogram Generation. Sensors 2022, 22, 926. [Google Scholar] [CrossRef] [PubMed]
  19. Huang, Q. Weight-Quantized SqueezeNet for Resource-Constrained Robot Vacuums for Indoor Obstacle Classification. AI 2022, 3, 180–193. [Google Scholar] [CrossRef]
  20. Wang, S.H.; Satapathy, S.C.; Zhou, Q.; Zhang, X.; Zhang, Y.D. Secondary Pulmonary Tuberculosis Identification Via pseudo-Zernike Moment and Deep Stacked Sparse Autoencoder. J. Grid Comput. 2022, 20, 1–16. [Google Scholar] [CrossRef]
  21. Bayraktar, Z.; Komurcu, M.; Werner, D.H. Wind Driven Optimization (WDO): A novel nature-inspired optimization algorithm and its application to electromagnetics. In Proceedings of the 2010 IEEE Antennas and Propagation Society International Symposium, Toronto, ON, Canada, 11–17 July 2010; pp. 1–4. [Google Scholar]
  22. Ramli, N.F.; Kamari, N.A.M.; Abd Halim, S.; Zulkifley, M.A.; Sahri, M.S.M.; Musirin, I. A Non-Convex Economic Dispatch Problem with Point-Valve Effect Using a Wind-Driven Optimisation Approach. J. Electr. Eng. Technol. 2022, 17, 85–95. [Google Scholar] [CrossRef]
  23. Leavey, P.; Sengupta, A.; Rakheja, D.; Daescu, O.; Arunachalam, H.B.; Mishra, R. Osteosarcoma data from UT Southwestern/UT Dallas for Viable and Necrotic Tumor Assessment [Data set]. Cancer Imaging Arch. 2019, 14. [Google Scholar] [CrossRef]
  24. Bansal, P.; Gehlot, K.; Singhal, A. Automatic Detection of Osteosarcoma Based on Integrated Features and Feature Selection Using Binary Arithmetic Optimization Algorithm. Multimed. Tools Appl. 2022, 81, 8807–8834. [Google Scholar] [CrossRef]
  25. Loraksa, C.; Mongkolsomlit, S.; Nimsuk, N.; Uscharapong, M.; Kiatisevi, P. Effectiveness of Learning Systems from Common Image File Types to Detect Osteosarcoma Based on Convolutional Neural Networks (CNNs) Models. J. Imaging 2021, 8, 2. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Overall process of WDODTL-ODC technique.
Figure 1. Overall process of WDODTL-ODC technique.
Healthcare 10 01040 g001
Figure 2. Structure of SqueezeNet Model.
Figure 2. Structure of SqueezeNet Model.
Healthcare 10 01040 g002
Figure 3. Sample Images.
Figure 3. Sample Images.
Healthcare 10 01040 g003
Figure 4. Confusion matrices of WDODTL-ODC technique (a) 80% of TR data, (b) 20% of TS data, (c) 70% of TR data, and (d) 30% TS data.
Figure 4. Confusion matrices of WDODTL-ODC technique (a) 80% of TR data, (b) 20% of TS data, (c) 70% of TR data, and (d) 30% TS data.
Healthcare 10 01040 g004
Figure 5. Result analysis of WDODTL-ODC technique under 80% of TR and 20% of TS data.
Figure 5. Result analysis of WDODTL-ODC technique under 80% of TR and 20% of TS data.
Healthcare 10 01040 g005
Figure 6. Result analysis of WDODTL-ODC technique under 70% of TR and 30% of TS data.
Figure 6. Result analysis of WDODTL-ODC technique under 70% of TR and 30% of TS data.
Healthcare 10 01040 g006
Figure 7. Classification analysis of WDODTL-ODC technique (a) 80:20 of accuracy, (b) 80:20 of loss, (c) 70:30 of accuracy, and (d) 70:30 of loss.
Figure 7. Classification analysis of WDODTL-ODC technique (a) 80:20 of accuracy, (b) 80:20 of loss, (c) 70:30 of accuracy, and (d) 70:30 of loss.
Healthcare 10 01040 g007
Figure 8. Classification analysis of WDODTL-ODC technique (a) 80:20 of precision-recall, (b) 80:20 of ROC, (c) 70:30 of precision-recall, and (d) 70:30 of ROC.
Figure 8. Classification analysis of WDODTL-ODC technique (a) 80:20 of precision-recall, (b) 80:20 of ROC, (c) 70:30 of precision-recall, and (d) 70:30 of ROC.
Healthcare 10 01040 g008
Figure 9. A c c u y analysis of WDODTL-ODC technique with existing algorithms.
Figure 9. A c c u y analysis of WDODTL-ODC technique with existing algorithms.
Healthcare 10 01040 g009
Figure 10. P r e c n analysis of WDODTL-ODC approach with existing algorithms.
Figure 10. P r e c n analysis of WDODTL-ODC approach with existing algorithms.
Healthcare 10 01040 g010
Figure 11. R e c a l analysis of WDODTL-ODC approach with existing algorithms.
Figure 11. R e c a l analysis of WDODTL-ODC approach with existing algorithms.
Healthcare 10 01040 g011
Figure 12. F s c o r e analysis of WDODTL-ODC approach with existing algorithms.
Figure 12. F s c o r e analysis of WDODTL-ODC approach with existing algorithms.
Healthcare 10 01040 g012
Table 1. Result analysis of WDODTL-ODC technique with various measures under 80% of TR and 20% of TS data.
Table 1. Result analysis of WDODTL-ODC technique with various measures under 80% of TR and 20% of TS data.
Class LabelsAccuracyPrecisionRecallF-ScoreMCCGeometric Mean
Training Phase (80%)
Viable Tumor99.4598.4999.6299.0598.6799.50
Non-Viable Tumor98.9197.7197.7197.7196.9998.49
Non-Tumor98.3698.6197.9398.2796.7198.34
Average98.9198.2798.4298.3497.4698.78
Testing Phase (20%)
Viable Tumor100.00100.00100.00100.00100.00100.00
Non-Viable Tumor99.5697.83100.0098.9098.6499.73
Non-Tumor99.56100.0099.0199.5099.1299.50
Average99.7199.2899.6799.4799.2599.74
Table 2. Result analysis of WDODTL-ODC technique with various measures under 70% of TR and 30% of TS data.
Table 2. Result analysis of WDODTL-ODC technique with various measures under 70% of TR and 30% of TS data.
Class LabelsAccuracyPrecisionRecallF-ScoreMCCGeometric Mean
Training Phase (70%)
Viable Tumor99.0098.3098.3098.3097.5998.79
Non-Viable Tumor99.2599.4397.1998.3097.8398.51
Non-Tumor99.0098.4799.4898.9798.0099.01
Average99.0898.7398.3298.5297.8198.77
Testing Phase (30%)
Viable Tumor99.4299.0999.0999.0998.6699.33
Non-Viable Tumor99.1398.8197.6598.2297.6598.63
Non-Tumor99.1398.6799.3399.0098.2399.15
Average99.2298.8698.6998.7798.1899.04
Table 3. Comparative analysis of WDODTL-ODC technique with existing approaches.
Table 3. Comparative analysis of WDODTL-ODC technique with existing approaches.
MethodsAccuracyPrecisionRecallF-Score
Handcrafted Feature95.5098.3498.1798.51
EfficientNet-B095.2095.4797.3795.97
Xception97.2294.8495.8595.71
EfficientNet-B0-Handcrafted97.7498.0498.3995.01
Xception-Handcrafted97.1094.8596.6496.63
VGG-1695.1197.6197.3095.30
ResNet-5097.0998.3294.1696.71
MobileNet-V298.2497.3998.3397.98
WDODTL-ODC99.2298.8698.6998.77
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Fakieh, B.; AL-Ghamdi, A.S.A.-M.; Ragab, M. Optimal Deep Stacked Sparse Autoencoder Based Osteosarcoma Detection and Classification Model. Healthcare 2022, 10, 1040. https://doi.org/10.3390/healthcare10061040

AMA Style

Fakieh B, AL-Ghamdi ASA-M, Ragab M. Optimal Deep Stacked Sparse Autoencoder Based Osteosarcoma Detection and Classification Model. Healthcare. 2022; 10(6):1040. https://doi.org/10.3390/healthcare10061040

Chicago/Turabian Style

Fakieh, Bahjat, Abdullah S. AL-Malaise AL-Ghamdi, and Mahmoud Ragab. 2022. "Optimal Deep Stacked Sparse Autoencoder Based Osteosarcoma Detection and Classification Model" Healthcare 10, no. 6: 1040. https://doi.org/10.3390/healthcare10061040

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop