Next Article in Journal
HydrAd: A Helper-Dependent Adenovirus Targeting Multiple Immune Pathways for Cancer Immunotherapy
Next Article in Special Issue
Cancerous Tumor Controlled Treatment Using Search Heuristic (GA)-Based Sliding Mode and Synergetic Controller
Previous Article in Journal
Chemoresistant Cancer Cell Lines Are Characterized by Migratory, Amino Acid Metabolism, Protein Catabolism and IFN1 Signalling Perturbations
Previous Article in Special Issue
Lightweight Deep Learning Model for Assessment of Substitution Voicing and Speech after Laryngeal Carcinoma Surgery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Chaotic Sparrow Search Algorithm with Deep Transfer Learning Enabled Breast Cancer Classification on Histopathological Images

1
Big Data and Machine Learning Laboratory, South Ural State University, 454080 Chelyabinsk, Russia
2
Department of Computer Science and Information System, College of Applied Sciences, AlMaarefa University, Riyadh 11597, Saudi Arabia
3
Department of Computer Science and Engineering, Sejong University, Seoul 05006, Korea
4
Artificial Intelligence Education, Hankuk University of Foreign Studies, Dongdaemun-gu, Seoul 02450, Korea
*
Authors to whom correspondence should be addressed.
Cancers 2022, 14(11), 2770; https://doi.org/10.3390/cancers14112770
Submission received: 18 April 2022 / Revised: 30 May 2022 / Accepted: 30 May 2022 / Published: 2 June 2022

Abstract

:

Simple Summary

Cancer is considered the most significant public health issue which severely threatens people’s health. The occurrence and mortality rate of breast cancer have been growing consistently. Initial precise diagnostics act as primary factors in improving the endurance rate of patients. Even though there are several means to identify breast cancer, histopathological diagnosis is now considered the gold standard in the diagnosis of cancer. However, the difficulty of histopathological image and the rapid rise in workload render this process time-consuming, and the outcomes might be subjected to pathologists’ subjectivity. Hence, the development of a precise and automatic histopathological image analysis method is essential for the field. Recently, the deep learning method for breast cancer pathological image classification has made significant progress, which has become mainstream in this field. Therefore, in this work, we focused on the design of metaheuristics with deep learning based breast cancer classification process. The proposed model is found to be an effective tool to assist physicians in the decision making process.

Abstract

Breast cancer is the major cause behind the death of women worldwide and is responsible for several deaths each year. Even though there are several means to identify breast cancer, histopathological diagnosis is now considered the gold standard in the diagnosis of cancer. However, the difficulty of histopathological image and the rapid rise in workload render this process time-consuming, and the outcomes might be subjected to pathologists’ subjectivity. Hence, the development of a precise and automatic histopathological image analysis method is essential for the field. Recently, the deep learning method for breast cancer pathological image classification has made significant progress, which has become mainstream in this field. This study introduces a novel chaotic sparrow search algorithm with a deep transfer learning-enabled breast cancer classification (CSSADTL-BCC) model on histopathological images. The presented CSSADTL-BCC model mainly focused on the recognition and classification of breast cancer. To accomplish this, the CSSADTL-BCC model primarily applies the Gaussian filtering (GF) approach to eradicate the occurrence of noise. In addition, a MixNet-based feature extraction model is employed to generate a useful set of feature vectors. Moreover, a stacked gated recurrent unit (SGRU) classification approach is exploited to allot class labels. Furthermore, CSSA is applied to optimally modify the hyperparameters involved in the SGRU model. None of the earlier works have utilized the hyperparameter-tuned SGRU model for breast cancer classification on HIs. The design of the CSSA for optimal hyperparameter tuning of the SGRU model demonstrates the novelty of the work. The performance validation of the CSSADTL-BCC model is tested by a benchmark dataset, and the results reported the superior execution of the CSSADTL-BCC model over recent state-of-the-art approaches.

1. Introduction

Cancer is considered the most significant public health issue which severely threatens people’s health. The occurrence and mortality rate of breast cancer (BC) have been growing consistently. Initial precise diagnostics act as primary factors in improving the endurance rate of patients [1]. A mammogram is the starting stage of initial prognosis; hence, it becomes hard to detect cancer in the denser breasts of teenage women. X-ray radiation warns radiologists of the patient’s health [2]. The golden standard for BC prognosis is only pathological examination. Pathological examinations generally attain tumor samples via excision, puncture, etc. [3]. Hematoxylin combines deoxyribonucleic acid (DNA), and eosin combines proteins. The precise prognosis of BC demands proficient histopathologists, and it needs more time and endeavor to finish this work. Moreover, the prognosis outcomes of distinct histopathologists are dissimilar and heavily based on histopathologists’ earlier experience [4].
Recently, BC prognosis is dependent on the histopathological image, and this is confronted by three major difficulties. At first, there is a shortcoming of proficient histopathologists across the globe, particularly in quite a few undeveloped regions and small hospitals [5]. Next, the prognosis of histopathologists is subjective, and evaluation is not performed on an objective basis. Whether prognosis is right or not is wholly based on the histopathologists’ earlier knowledge [6]. Lastly, the prognosis of BC depends on the histopathological image, which is time consuming, highly complex, and labor-intensive, and it is considered ineffective during the era of big data. Despite such issues, an objective and effective BC prognosis technique is essential for mitigating the pressure of the workload of histopathologists [7]. The speedy advancement of computer-aided diagnosis (CAD) was slowly employed in the clinical domain. The CAD system will not act as a substitute for the physician; however, it can be utilized as a “second reader” in assisting the physician in recognizing diseases [8]. However, there are false-positive areas identified by the computer that will consume time for the physician in evaluating the outcomes induced by the computer, again leading to a decline in effectiveness and preciseness. Thus, methods for improving the sensitiveness of computer-aided tumor identification methodologies while greatly minimizing the incorrect positive identification rate and enhancing the efficiency of the identification technique constitute a potential research area [9].
Currently, deep learning (DL) methods have become popular in computer vision (CV), particularly in biomedical image processing. These methods were able to investigate complex and enhanced characteristics from images automatically. At the same time, these methods greatly require the attention of several authors in using such techniques to categorize BC histopathology images [10]. In particular terms, convolutional neural networks (CNNs) are broadly utilized in image-based works because of their capabilities to efficiently distribute variables over several layers inside a DL method.
This study introduces a novel chaotic sparrow search algorithm with a deep transfer learning-enabled breast cancer classification (CSSADTL-BCC) model applied on histopathological images. The presented CSSADTL-BCC model applies the Gaussian filtering (GF) approach to eradicate the occurrence of noise. In addition, a MixNet-based feature extraction model was employed to generate a useful set of feature vectors. Furthermore, a CSSA with a stacked gated recurrent unit (SGRU) classification approach was exploited to allot class labels. The CSSADTL-BCC model does not exist in the literature to the best of our knowledge. The design of the CSSA for optimal hyperparameter tuning of the SGRU model demonstrates the novelty of the work. The performance validation of the CSSADTL-BCC model was verified using benchmark data collection, and the outcomes were inspected under different evaluation measures.
The remaining sections of the paper are planned as follows. Section 2 indicates the existing works related to BC classification. Next, Section 3 elaborates the proposed model, and Section 4 offers the performance validation. At last, Section 5 draws the conclusions.

2. Literature Review

In [11], the authors proposed a real time data augmentation-related transfer learning method to resolve existing limitations. Two popular and well-established image classification methods, such as Xception and InceptionV3 frameworks, have been trained on a freely accessible BC histopathological image data named BreakHis. Alom et al. [12] presented a technique for classifying BC using the Inception Recurrent Residual Convolution Neural Network (IRRCNN) framework. The proposed method is an effective DCNN system that integrates the strength of the Recurrent Convolution Neural Network (RCNN), Inception Network (Inception-v4), and the Residual Network (ResNet). The experiment result illustrates better performance against RCNN, Inception Network, and ResNet for object-detection tasks.
Vo et al. [13] presented a technique that employs the DL method with a convolution layer for extracting the visual feature for BC classification. It has been found that the DL model extracts the most useful feature when compared to the handcrafted feature extraction approach. In [14], the authors proposed a BC histopathological image categorization related to deep feature fusion and enhanced routing (FE-BkCapsNet) to exploit CapsNet and CNN models. Firstly, a new architecture with two channels could simultaneously extract capsule and convolutional features and incorporate spatial and sematic features into the new capsule to obtain a discriminative dataset.
The researchers in [15] proposed a patch-based DL method named Pa-DBN-BC for classifying and detecting BC on histopathology images with the Deep Belief Network (DBN). The feature is extracted by supervised finetuning and unsupervised pre-training phases. The network extracts feature automatically from image patches. Logistic regression is utilized for classifying the patches from histopathology images. In [16], the authors proposed a robust and novel technique based convolution-LSTM (CLSTM) learning method, the pre-processing method with the optimized SVM classifier, and the marker-controlled watershed segmentation algorithm (MWSA) for automatically identifying BC. Saxena et al. [17] presented a hybrid ML method for solving class imbalance problems. The presented method uses the kernelized weighted ELM and pre-trained ResNet50 for CAD of BC using histopathology.
Several automated breast cancer classification models are available in the literature. However, the models still contains a challenging problem. Because of the continual deepening of models, the number of parameters of DL models also increases quickly, which results in model overfitting. At the same time, different hyperparameters have a significant impact on the efficiency of the CNN model, particularly in terms of the learning rate. Modifying the learning rate parameter for obtaining better performance is also required. Therefore, in this study, we employ the CSSA technique for the hyperparameter tuning of the SGRU model.

3. The Proposed Model

In this study, a new CSSADTL-BCC model was developed to classify BC on histopathological images. The presented CSSADTL-BCC model mainly focused on the recognition and classification of BC. At the primary stage, the CSSADTL-BCC model employed the GF technique to eradicate the occurrence of noise. It was then followed by using a MixNet-based feature extraction model employed to produce a useful set of feature vectors. Then, the CSSA-SGRU classifier was exploited to allot class labels. Figure 1 illustrates the overall process of the CSSADTL-BCC technique.

3.1. Image Pre-Processing

At the primary stage, the CSSADTL-BCC model employed the GF technique to eradicate the occurrence of noise. GF is a bandpass filter, viz., efficiently implemented in machine vision and image processing applications [18]. A two-dimensional Gabor purpose was oriented by sinusoidal grates controlled by two dimensional Gaussian envelopes. In the two-dimensional coordinate a , b model, the GF comprising an imaginary and real one is illustrated by the following:
G δ , θ , ψ , σ , γ a , b = exp a 2 + γ 2 b 2 2 σ 2 × exp j 2 π a δ + ψ
where they are described as follows.
a = a   cos   θ + b   sin   θ
b = a   sin   θ + b   cos   θ
Now θ implies the orientation separation angle of the Gabor kernel, and   δ signifies the wavelength of sinusoidal features. Notably, it is essential to consider θ from the range 0 o ,   180 o as symmetry generates another redundant direction. ψ denotes the stage offset, σ indicates the standard derivation of the Gaussian envelope, and γ represents the ratio of spatial features for identifying the ellipticity of the Gabor role. ψ = 0 and ψ = π / 2 return the real and imaginary parts of GF. Variable 0 can be determined as 6 and spatial frequency bandwidth b w is given by the following.
σ = δ p i l n 2 2     2 b w + 1 2 b w 1

3.2. MixNet-Based Feature Extractor

Next, for image pre-processing, a MixNet-based feature extraction model is employed to generate a useful set of feature vectors. A CNN algorithm created by the traditional convolutional operation is difficult to use for mobile terminals due to its complicated calculations and excessive parameters. In order to improve its effectiveness on mobile terminals and to guarantee the accuracy of the model, a sequence of lightweight convolutional operators has been presented. Amongst them, one of the most commonly utilized is a depthwise separable convolution layer. A depthwise separable convolutional layer splits the convolution into pointwise and depthwise convolution. In the initial phase, it convolves a single channel at a time using convolutional kernels at size = 3. In the second phase, it uses a feature map with the 1 × 1 convolutional kernel. Assume that N D k × D k feature view and 1 convolutional sliding step are utilized to convolve a feature map with D F × D F × M dimensions, including the output feature map with dimensions of D F × D F × N . The parameter amount of traditional convolutional operations is provided as follows.
D k × D k × M × N
The parameters involved in the depthwise separable convolutional operation is provided below.
D k × D k × M + 1 × 1 × M × N  
The computation involved in traditional convolutional operation is provided as follows.
D k × D k × M × N × D F × D F
The computation involved in depthwise separable convolutional operation is defined in Equation (8).
D k × D k × M × D F × D F × M × N × D F × D F
The ratio of the two operations is provided as follows.
D k × D k × M × D F × D F × M × N × D F × D F D k × D k × M × N × D F × D F
A depthwise separable convolutional layer uses a similar size 3 × 3 convolutional kernel in the computation method; however, a network with larger convolutional kernels of 5 × 5 or 7 × 7 confirms that a larger convolutional kernel improves the efficiency and accuracy of the model. However, the experiment shows that the case where a larger convolutional kernel is better is rare; simultaneously, a large convolutional kernel minimizes the model’s accuracy. Here, MDConv splits the input channel with M size into C groups, later convolving all the groups with distinct kernel sizes. The standard depthwise separable convolution splits the input channel with M size into M groups and later implements convolutional calculations for all groups with a similar kernel size.

3.3. Image Classification Using SGRU Model

At this stage, the generated feature vectors are passed into the SGRU classifier to allot class labels. SGRU is made up of various GRU units. For time series t , the input series e 1 ,   e 2 ,   ,   e t first enters into hidden layer h 1 1 ,   h 2 1 ,   ,   h t 1 to attain all data from the previous time step. Next, the upper hidden layer takes the output from the lower hidden layers at a similar time step as the input for extracting features [19]. In particular, the upper layer of the hidden layer is h 1 2 ,   h 2 2 ,   ,   h t 2 . For all layers, a hidden layer h t i , as provided in Equation (13), is shown by Equations (10)–(12) to attain the candidate value, update, and reset gates. It should be noted that in Equations (10)–(12), we have included embedding vector e t in the initial layer. Starting from the next layer upward, we employ the hidden state from the current time step in the previous layer, h t i 1 , rather than e t in (10)–(12). Figure 2 depicts the framework of SGRU.
u t i = σ W u i h t 1 i + U u i e t + b u i
r t i = σ W r i h t 1 i + U r i e t + b r i
C ˜ = tanh W c i . r t i × h t 1 i + U c i e t + b c i
h t i = u t i × C ˜ t i + 1 u t i × h t 1 i

3.4. Hyperparameter Optimization

Finally, CSSA is implied to optimally modify the hyperparameters included in the MixNet model. SSA attains the best possible solution by mimicking certain behaviors of sparrows [20]. Firstly, the discoverer–joiner sparrow population models are established, and then the sparrow is arbitrarily chosen as a guard. The joiner snatches food from the discoverer, observes the discoverer, and follows the discoverer for food. The discoverer takes the responsibility to provide foraging direction and areas for the sparrow population. Once the vigilante realizes the threat, the population implements anti-predation behavior immediately. Lastly, with various iterations of the location of the discoverer and joiner, the adoptive position for the entire population can be found. The sparrow population is within the space of N × D , where N indicates the overall amount of sparrows, D represents the spatial dimension. Next, the location of the i-th sparrow in space represents X i = x i 1 ,   x i 2 ,   ,   x i d , i 1 ,   N ,   d 1 ,   D , and x i d characterizes the location of i-th sparrow in d -dimension. The position update equation of the discoverer can be shown in the following Equation (14).
x i d t + 1 = x i d t e x p . i α T R 2 < S T x i d t + Q · L , R 2 S T
In the equation, t signifies the existing amount of iterations; T indicates the maximal amount of iterations; α represents an arbitrary value within [0, 1] ; Q implies an arbitrary value with standard distribution; L indicates a matrix in that element is 1, and its size is 1 × d ; R 2 0 , 1 signifies the warning values; S T 0.5 , 1 denotes the safety values. If R 2 < S T , this implies that the population is not at risk and the discoverer continues searching. If R 2   S T , this implies that the vigilante discovered the predator and instantly delivered an alarm to the others. The sparrow population implements anti-predation behavior immediately any fly to a safer region for food. The position update equation of the joiner can be shown in the following Equation (15).
x i d t + 1 = Q e x p x w o r s t d t x i d t i 2 i > N 2 x b e s t   d t + 1 + 1 D d = 1 D ( r a n d 1 , 1 | x i d t x b e s t   d t + 1 i N 2
Here, x w o r s t d t signifies the global worst place in t t h iteration; x b e s t d t + 1 signifies the global optimal location at the t t h iteration. If i > N 2 , it implies that the i-th joiner has not attained food and that it needs to fly toward another location in order to search for food. If i N 2 , this implies that the i-th joiner is closer to the world’s best location and is arbitrarily foraging around. The vigilant location upgrade equation is provided as follows:
x i d r + 1 = x w o r s t   d t + β x i d t x w o r s t   d t , f i f g x i d t + K x i d t x w o r s t   d t f i f w + e f i = f g
where β signifies the step length control variable that is an arbitrary value subjected to a regular distribution with a variance of 1 and means value of 0 ; K denotes the movement direction of sparrow, and arbitrary values lie within [1, 1]; e indicates a constant with smaller value; f i characterizes the fitness of i-th sparrow; f g signifies the optimum fitness of the existing population; f w denotes the worst fitness of existing population. If f i f g , this implies that the i-th sparrow is at the edge of the population and can be attacked easily by the predator. If f i = f g , this implies that i-th sparrow is within center of the population, and it is aware of danger; it relocates closer to other sparrows in order to reduce the threat of becoming caught.
With the addition of a global optimum sparrow neighborhood in all iterations, the searching ability of SSA can be enhanced. Additionally, this could assist the sparrow group in attaining the best location through the search process. The chaotic local searching technique can be employed in the iteration process of SSA for improving the capability of exploitation and maintaining a better harmony among the core search processes. Moreover, the logical chaotic function is employed to calculate chaotic SSA. This can be obtained as follows.
ρ k + 1 = μ ρ k 1 ρ k ,   k = 1 , 2 ,   ,   N 1
On the other hand, ρ 1 0 , 1 and ρ 1 0.25 ,   0.5 ,   0.75 , and 1 once the control parameter μ is set to 4, and the logistic function is converted to a chaotic state. Therefore, the chaotic local searching function is shown below.
P i = b + ρ i × b a ,   i = 1 , 2 , , N
Here, a ,   b indicates the searching space, and the chaotic function was produced by mapping chaotic parameters ρ i into the chaotic vector P i . Furthermore, chaotic vector P i was linearly integrated with targeted position T P for generating candidate location C L , which is expressed as follows.
C L = 1 S C × T P + S C × P i
S C = T t + 1 / T
The CSSA approach resolves an FF for obtaining higher classification performances. It defines a positive integer for demonstrating the optimal performance of candidate solutions. During this case, the minimized classifier error rate was regarded as FF, as offered in Equation (21).
f i t n e s s x i = C l a s s i f i e r E r r o r R a t e x i = n u m b e r   o f   m i s c l a s s i f i e d   s a m p l e s T o t a l   n u m b e r   o f   s a m p l e s × 100

4. Performance Validation

In this section, the experimental validation of the CSSADTL-BCC model is tested using a benchmark dataset [21], and the details are provided in Table 1. The CSSADTL-BCC model is simulated using the Python 3.6.5 tool. The parameter settings are provided as follows: learning rate—0.01; dropout—0.5; batch size—5; epoch count—50; activation—ReLU. A few sample images are demonstrated in Figure 3.
Figure 4 illustrates the confusion matrices produced by the CSSADTL-BCC model under distinct epochs. With 500 epochs, the CSSADTL-BCC model has identified 65 samples in class A, 205 samples in class F, 81 samples in class PT, 84 samples in class TA, 760 samples in class DC, 93 samples in class LC, 117 samples in class MC, and 96 samples in class PC. Along with that, with 2000 epochs, the CSSADTL-BCC approach has identified 89 samples in class A, 228 samples in class F, 109 samples in class PT, 112 samples in class TA, 779 samples in class DC, 116 samples in class LC, 160 samples in class MC, and 121 samples in class PC.
Table 2 and Figure 5 highlight the overall classification outcomes of the CSSADTL-BCC model under distinct epochs and class labels. The experimental outcomes implied that the CSSADTL-BCC model has resulted in ineffectual outcomes over other models in terms of different measures such as accuracy ( a c c u y ), precision ( p r e c n ), recall ( r e c a l ), specificity ( s p e c y ), F-score ( F s c o r e ), MCC, and G-mean ( G m e a n ). For instance, with 500 epochs, the CSSADTL-BCC model provided the averages of a c c u y , p r e c n , r e c a l , s p e c y , F s c o r e , MCC, and G m e a n at 95.62%, 78.78%, 73.25%, 97.09%, 75.71%, 73.18%, and 84.01%, respectively. Moreover, with 1000 epochs, the CSSADTL-BCC method obtained the averages of a c c u y , p r e c n , r e c a l , s p e c y , F s c o r e , MCC, and G m e a n at 97.10%, 85.21%, 82.09%, 98.16%, 83.52%, 81.84%, and 89.62%, respectively. In addition, with 1500 epochs, the CSSADTL-BCC methodology provided averages of a c c u y , p r e c n , r e c a l , s p e c y , F s c o r e , MCC, and G m e a n at 98.61%, 92.80%, 91.48%, 99.14%, 92.10%, 91.29%, and 95.19%, respectively. At last, with 2000 epochs, the CSSADTL-BCC technique obtained the averages of a c c u y , p r e c n , r e c a l , s p e c y , F s c o r e , MCC, and G m e a n at 98.54%, 92.58%, 90.87%, 99.08%, 91.66%, 90.82%, and 94.84%, respectively.
The training accuracy (TA) and validation accuracy (VA) attained by the CSSADTL-BCC model on test dataset are demonstrated in Figure 6. The experimental outcomes implied that the CSSADTL-BCC model has gained maximum values of TA and VA. In particular, VA appeared to be higher than TA.
The training loss (TL) and validation loss (VL) achieved by the CSSADTL-BCC method on test dataset are established in Figure 7. The experimental outcome inferred that the CSSADTL-BCC model obtained the lowest values of TL and VL. In particular, VL seemed to be lower than TL. Next, a brief precision–recall examination performed on the CSSADTL-BCC method on the test dataset is displayed in Figure 8. By observing the figure, it can be observed that the CSSADTL-BCC approach has established maximal precision–recall performance under all classes.
Figure 9 portrays a clear ROC investigation of the CSSADTL-BCC model on the test dataset. The figure portrayed that the CSSADTL-BCC model has resulted in proficient results with maximum ROC values under distinct class labels.
Figure 10 reports detailed classification accuracy outcomes of the CSSADTL-BCC model under distinct iterations and runs. The figures highlighted that CSSADTL-BCC has showcased effectual classifier results under every epoch.
To highlight the enhanced outcomes of the CSSADTL-BCC model, a brief comparison study with recent models is shown in Table 3 [22]. Figure 11 investigates a detailed a c c u y and F s c o r e analysis of the CSSADTL-BCC with existing models. The results indicated that GLCM-KNN and GLCM-NB models obtained lower values of a c c u y and F s c o r e . At the same time, the GLCM-discrete transform, GLCM-SVM, and Deep learning-IRV2 models have attained moderately closer values of a c c u y and F s c o r e . Next to that, the GLCM-DL and Deep learning INV3 models have resulted in reasonable a c c u y and F s c o r e values. However, the CSSADTL-BCC model has gained an effectual outcome with maximum a c c u y and F s c o r e at 98.61% and 92.80%, respectively.
Figure 12 examines a detailed p r e c n and r e c a l examination of CSSADTL-BCC with existing techniques. The outcomes represented that the GLCM-KNN and GLCM-NB approaches have gained lesser values of p r e c n and r e c a l . Moreover, the GLCM-discrete transform, GLCM-SVM, and Deep learning-IRV2 algorithms have attained moderately closer values of p r e c n and r e c a l . Along with that, the GLCM-DL and Deep learning INV3 approaches have resulted in reasonable p r e c n and r e c a l values. However, the CSSADTL-BCC technique has gained effectual outcomes with maximum values of p r e c n and r e c a l at 92.80% and 91.48%, respectively. After observing the results and discussion, it is apparent that the CSSADTL-BCC model has showcased enhanced outcomes over other methods. The enhanced performance of the CSSADTL-BCC model is due to the effectual hyperparameter tuning process of the SGRU classifier. Thus, the proposed model can be applied to assist physicians in the disease diagnosis process.

5. Conclusions

In this study, a new CSSADTL-BCC method was advanced for classifying BC on histopathological images. The presented CSSADTL-BCC model mainly focused on the recognition and classification of BC. At the primary stage, the CSSADTL-BCC model employed the GF technique to eradicate the occurrence of noise. Moreover, a MixNet-based feature extraction model was employed for producing a useful collection of feature vectors. Then, the SGRU classifier was exploited to allot class labels. Furthermore, CSSA is applied to optimally modify the hyperparameters involved in the MixNet model. The performance validation of the CSSADTL-BCC model can be tested by using a benchmark dataset, and the results reported the superior efficiency of the CSSADTL-BCC method over the current existing approaches with a maximum accuracy of 98.61%. In the future, deep instance segmentation approaches can be included to enhance classification performance. In addition, the classifier’s results can be boosted by designing deep fusion-based ensemble models.

Author Contributions

Conceptualization, K.S.; methodology, K.S., A.K.D.; software, S.K.; validation, S.K., I.C.D.; formal analysis, K.S.; investigation, I.C.D.; resources, G.P.J.; data curation, A.K.D.; writing—original draft preparation, K.S.; writing—review and editing, G.P.J.; visualization, S.K.; supervision, I.C.D.; project administration, I.C.D. funding acquisition, G.P.J., I.C.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Hankuk University of Foreign Studies Research Fund (of 2022) and the Ministry of Science and Higher Education of the Russian Federation (Government Order FENU-2020–0022).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable to this article as no datasets were generated during the current study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Das, A.; Nair, M.S.; Peter, S.D. Computer-aided histopathological image analysis techniques for automated nuclear atypia scoring of breast cancer: A review. J. Digit. Imaging 2020, 33, 1091–1121. [Google Scholar] [CrossRef] [PubMed]
  2. Krithiga, R.; Geetha, P. Breast cancer detection, segmentation and classification on histopathology images analysis: A systematic review. Arch. Comput. Methods Eng. 2021, 28, 2607–2619. [Google Scholar] [CrossRef]
  3. Carvalho, E.D.; Antonio Filho, O.C.; Silva, R.R.; Araujo, F.H.; Diniz, J.O.; Silva, A.C.; Paiva, A.C.; Gattass, M. Breast cancer diagnosis from histopathological images using textural features and CBIR. Artif. Intell. Med. 2020, 105, 101845. [Google Scholar] [CrossRef] [PubMed]
  4. Xie, J.; Liu, R.; Luttrell, J., IV; Zhang, C. Deep learning based analysis of histopathological images of breast cancer. Front. Genet. 2019, 10, 80. [Google Scholar] [CrossRef] [Green Version]
  5. Kaushal, C.; Bhat, S.; Koundal, D.; Singla, A. Recent trends in computer assisted diagnosis (CAD) systems for breast cancer diagnosis using histopathological images. IRBM 2019, 40, 211–227. [Google Scholar] [CrossRef]
  6. Yan, R.; Ren, F.; Wang, Z.; Wang, L.; Zhang, T.; Liu, Y.; Rao, X.; Zheng, C.; Zhang, F. Breast cancer histopathological image classification using a hybrid deep neural network. Methods 2020, 173, 52–60. [Google Scholar] [CrossRef]
  7. Mehra, R. Breast cancer histology images classification: Training from scratch or transfer learning? ICT Express 2018, 4, 247–254. [Google Scholar]
  8. Alkassar, S.; Jebur, B.A.; Abdullah, M.A.; Al-Khalidy, J.H.; Chambers, J.A. Going deeper: Magnification-invariant approach for breast cancer classification using histopathological images. IET Comput. Vis. 2021, 15, 151–164. [Google Scholar] [CrossRef]
  9. Sohail, A.; Khan, A.; Wahab, N.; Zameer, A.; Khan, S. A multi-phase deep CNN based mitosis detection framework for breast cancer histopathological images. Sci. Rep. 2021, 11, 6215. [Google Scholar] [CrossRef]
  10. Ahmad, N.; Asghar, S.; Gillani, S.A. Transfer learning-assisted multi-resolution breast cancer histopathological images classification. Vis. Comput. 2021, 1–20. [Google Scholar] [CrossRef]
  11. Rai, R.; Sisodia, D.S. Real-time data augmentation based transfer learning model for breast cancer diagnosis using histopathological images. In Advances in Biomedical Engineering and Technology; Springer: Singapore, 2021; pp. 473–488. [Google Scholar]
  12. Alom, M.Z.; Yakopcic, C.; Nasrin, M.; Taha, T.M.; Asari, V.K. Breast cancer classification from histopathological images with inception recurrent residual convolutional neural network. J. Digit. Imaging 2019, 32, 605–617. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Vo, D.M.; Nguyen, N.Q.; Lee, S.W. Classification of breast cancer histology images using incremental boosting convolution networks. Inf. Sci. 2019, 482, 123–138. [Google Scholar] [CrossRef]
  14. Wang, P.; Wang, J.; Li, Y.; Li, P.; Li, L.; Jiang, M. Automatic classification of breast cancer histopathological images based on deep feature fusion and enhanced routing. Biomed. Signal Process. Control. 2021, 65, 102341. [Google Scholar] [CrossRef]
  15. Hirra, I.; Ahmad, M.; Hussain, A.; Ashraf, M.U.; Saeed, I.A.; Qadri, S.F.; Alghamdi, A.M.; Alfakeeh, A.S. Breast cancer classification from histopathological images using patch-based deep learning modeling. IEEE Access 2021, 9, 24273–24287. [Google Scholar] [CrossRef]
  16. Demir, F. DeepBreastNet: A novel and robust approach for automated breast cancer detection from histopathological images. Biocybern. Biomed. Eng. 2021, 41, 1123–1139. [Google Scholar] [CrossRef]
  17. Saxena, S.; Shukla, S.; Gyanchandani, M. Breast cancer histopathology image classification using kernelized weighted extreme learning machine. Int. J. Imaging Syst. Technol. 2021, 31, 168–179. [Google Scholar] [CrossRef]
  18. Wang, Y.; Yan, J.; Yang, Z.; Zhao, Y.; Liu, T. Optimizing GIS partial discharge pattern recognition in the ubiquitous power internet of things context: A MixNet deep learning model. Int. J. Electr. Power Energy Syst. 2021, 125, 106484. [Google Scholar] [CrossRef]
  19. Al Wazrah, A.; Alhumoud, S. Sentiment Analysis Using Stacked Gated Recurrent Unit for Arabic Tweets. IEEE Access 2021, 9, 137176–137187. [Google Scholar] [CrossRef]
  20. Yuan, J.; Zhao, Z.; Liu, Y.; He, B.; Wang, L.; Xie, B.; Gao, Y. DMPPT control of photovoltaic microgrid based on improved sparrow search algorithm. IEEE Access 2021, 9, 16623–16629. [Google Scholar] [CrossRef]
  21. Spanhol, F.; Oliveira, L.S.; Petitjean, C.; Heutte, L. A Dataset for Breast Cancer Histopathological Image Classification. IEEE Trans. Biomed. Eng. (TBME) 2016, 63, 1455–1462. [Google Scholar] [CrossRef]
  22. Reshma, V.K.; Arya, N.; Ahmad, S.S.; Wattar, I.; Mekala, S.; Joshi, S.; Krah, D. Detection of Breast Cancer Using Histopathological Image Classification Dataset with Deep Learning Techniques. BioMed Res. Int. 2022. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The overall process of the CSSADTL-BCC technique.
Figure 1. The overall process of the CSSADTL-BCC technique.
Cancers 14 02770 g001
Figure 2. Framework of SGRU model.
Figure 2. Framework of SGRU model.
Cancers 14 02770 g002
Figure 3. Sample images: (a) benign (b) malignant.
Figure 3. Sample images: (a) benign (b) malignant.
Cancers 14 02770 g003
Figure 4. Confusion matrix of CSSADTL-BCC technique under various epochs: (a) 500 epochs, (b) 1000 epochs, (c) 1500 epochs, and (d) 2000 epochs.
Figure 4. Confusion matrix of CSSADTL-BCC technique under various epochs: (a) 500 epochs, (b) 1000 epochs, (c) 1500 epochs, and (d) 2000 epochs.
Cancers 14 02770 g004
Figure 5. Result analysis of CSSADTL-BCC technique with distinct epochs.
Figure 5. Result analysis of CSSADTL-BCC technique with distinct epochs.
Cancers 14 02770 g005
Figure 6. TA and VA analysis of CSSADTL-BCC technique.
Figure 6. TA and VA analysis of CSSADTL-BCC technique.
Cancers 14 02770 g006
Figure 7. TL and VL analysis of CSSADTL-BCC technique.
Figure 7. TL and VL analysis of CSSADTL-BCC technique.
Cancers 14 02770 g007
Figure 8. Precision–recall curve analysis of CSSADTL-BCC technique.
Figure 8. Precision–recall curve analysis of CSSADTL-BCC technique.
Cancers 14 02770 g008
Figure 9. ROC curve analysis of the CSSADTL-BCC technique.
Figure 9. ROC curve analysis of the CSSADTL-BCC technique.
Cancers 14 02770 g009
Figure 10. Classification accuracy of CSSADTL-BCC technique under distinct iterations: (a) 500 epochs, (b) 1000 epochs, (c) 1500 epochs, and (d) 2000 epochs.
Figure 10. Classification accuracy of CSSADTL-BCC technique under distinct iterations: (a) 500 epochs, (b) 1000 epochs, (c) 1500 epochs, and (d) 2000 epochs.
Cancers 14 02770 g010
Figure 11. A c c u y and F s c o r e analysis of CSSADTL-BCC technique with existing algorithms.
Figure 11. A c c u y and F s c o r e analysis of CSSADTL-BCC technique with existing algorithms.
Cancers 14 02770 g011
Figure 12. R e c a l and P r e c n analysis of the CSSADTL-BCC technique with existing algorithms.
Figure 12. R e c a l and P r e c n analysis of the CSSADTL-BCC technique with existing algorithms.
Cancers 14 02770 g012
Table 1. Dataset details.
Table 1. Dataset details.
CategoryClass NamesLabelsNo. of ImagesTotal
BenignAdenosisA106588
FibroadenomaF237
Phyllodes TumorPT115
Tubular AdenomaTA130
MalignantCarcinomaDC7881232
Lobular CarcinomaLC137
Mucinous CarcinomaMC169
Papillary CarcinomaPC138
Total Number of Images1820
Table 2. Result analysis of CSSADTL-BCC technique with various measures and epochs.
Table 2. Result analysis of CSSADTL-BCC technique with various measures and epochs.
Class LabelsAccuracyPrecisionRecallSpecificityF-ScoreMCCG-Mean
Epoch-500
A96.4373.0361.3298.6066.6765.0777.76
F95.7782.0086.5097.1684.1981.7991.67
PT97.2583.5170.4399.0676.4275.2783.53
TA95.5570.5964.6297.9367.4765.1679.55
DC92.4287.3696.4589.3491.6885.1092.83
LC96.2178.8167.8898.5172.9471.1481.78
MC94.8473.5869.2397.4671.3468.5482.14
PC96.4881.3669.5798.6975.0073.3882.86
Average95.6278.7873.2597.0975.7173.1884.01
Epoch-1000
A97.2580.4369.8198.9574.7573.5183.11
F97.2087.5091.5698.0489.4887.9094.75
PT98.1389.3280.0099.3584.4083.5689.15
TA96.7677.5276.9298.2877.2275.4886.95
DC95.8293.2097.4694.5795.2991.6196.01
LC97.1484.5575.9198.8780.0078.6086.63
MC96.9883.1484.6298.2483.8782.2191.18
PC97.5386.0580.4398.9383.1581.8789.20
Average97.1085.2182.0998.1683.5281.8489.62
Epoch-1500
A98.4689.8083.0299.4286.2785.5390.85
F98.6894.1995.7899.1294.9894.2297.43
PT99.2393.9193.9199.5993.9193.5096.71
TA98.4190.4086.9299.2988.6387.7992.90
DC98.1397.1298.6097.7797.8696.2198.19
LC98.6893.8088.3299.5290.9890.3193.76
MC98.5289.8994.6798.9192.2291.4496.77
PC98.7993.2890.5899.4691.9191.2794.92
Average98.6192.8091.4899.1492.1091.2995.19
Epoch-2000
A98.5790.8283.9699.4787.2586.5791.39
F98.6893.8396.2099.0595.0094.2597.62
PT99.1892.3794.7899.4793.5693.1397.10
TA98.3089.6086.1599.2387.8486.9592.46
DC98.0296.6598.8697.3897.7496.0098.12
LC98.4694.3184.6799.5889.2388.5591.83
MC98.6891.4394.6799.0993.0292.3196.86
PC98.4691.6787.6899.3589.6388.8293.33
Average98.5492.5890.8799.0891.6690.8294.84
Table 3. Comparative analysis of the CSSADTL-BCC technique with existing algorithms.
Table 3. Comparative analysis of the CSSADTL-BCC technique with existing algorithms.
MethodsAccuracy Precision RecallF-Score
GLCM-KNN Model76.1762.4083.6082.22
GLCM-NB Model78.4582.1683.4586.97
GLCM-Discrete transform85.0083.5681.6684.69
GLCM-SVM Model85.0087.3287.6181.62
GLCM-DL Model92.4486.8980.2487.92
Deep Learning-INV394.7187.5787.0781.86
Deep Learning-IRV288.1281.7081.4486.42
CSSADTL-BCC98.6192.8091.4892.10
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shankar, K.; Dutta, A.K.; Kumar, S.; Joshi, G.P.; Doo, I.C. Chaotic Sparrow Search Algorithm with Deep Transfer Learning Enabled Breast Cancer Classification on Histopathological Images. Cancers 2022, 14, 2770. https://doi.org/10.3390/cancers14112770

AMA Style

Shankar K, Dutta AK, Kumar S, Joshi GP, Doo IC. Chaotic Sparrow Search Algorithm with Deep Transfer Learning Enabled Breast Cancer Classification on Histopathological Images. Cancers. 2022; 14(11):2770. https://doi.org/10.3390/cancers14112770

Chicago/Turabian Style

Shankar, K., Ashit Kumar Dutta, Sachin Kumar, Gyanendra Prasad Joshi, and Ill Chul Doo. 2022. "Chaotic Sparrow Search Algorithm with Deep Transfer Learning Enabled Breast Cancer Classification on Histopathological Images" Cancers 14, no. 11: 2770. https://doi.org/10.3390/cancers14112770

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop