Next Article in Journal
Development and Validation of a Nomograph Model for Post-Operative Central Nervous System Infection after Craniocerebral Surgery
Next Article in Special Issue
Deep Learning Framework with Multi-Head Dilated Encoders for Enhanced Segmentation of Cervical Cancer on Multiparametric Magnetic Resonance Imaging
Previous Article in Journal
Overview of Long-Term Outcome in Adults with Systemic Right Ventricle and Transposition of the Great Arteries: A Review
Previous Article in Special Issue
ResRandSVM: Hybrid Approach for Acute Lymphocytic Leukemia Classification in Blood Smear Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ExtRanFS: An Automated Lung Cancer Malignancy Detection System Using Extremely Randomized Feature Selector

Department of Computer Science, University of Kerala, Thiruvananthapuram 695581, India
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(13), 2206; https://doi.org/10.3390/diagnostics13132206
Submission received: 31 May 2023 / Revised: 22 June 2023 / Accepted: 25 June 2023 / Published: 29 June 2023
(This article belongs to the Special Issue Deep Learning for Early Detection of Cancer)

Abstract

:
Lung cancer is an abnormality where the body’s cells multiply uncontrollably. The disease can be deadly if not detected in the initial stage. To address this issue, an automated lung cancer malignancy detection (ExtRanFS) framework is developed using transfer learning. We used the IQ-OTH/NCCD dataset gathered from the Iraq Hospital in 2019, encompassing CT scans of patients suffering from various lung cancers and healthy subjects. The annotated dataset consists of CT slices from 110 patients, of which 40 were diagnosed with malignant tumors and 15 with benign tumors. Fifty-five patients were determined to be in good health. All CT images are in DICOM format with a 1mm slice thickness, consisting of 80 to 200 slices at various sides and angles. The proposed system utilized a convolution-based pre-trained VGG16 model as the feature extractor and an Extremely Randomized Tree Classifier as the feature selector. The selected features are fed to the Multi-Layer Perceptron (MLP) Classifier for detecting whether the lung cancer is benign, malignant, or normal. The accuracy, sensitivity, and F1-Score of the proposed framework are 99.09%, 98.33%, and 98.33%, respectively. To evaluate the proposed model, a comparison is performed with other pre-trained models as feature extractors and also with the existing state-of-the-art methodologies as classifiers. From the experimental results, it is evident that the proposed framework outperformed other existing methodologies. This work would be beneficial to both the practitioners and the patients in identifying whether the tumor is benign, malignant, or normal.

1. Introduction

Lung cancer is considered the second most common cancer worldwide, and it usually begins in the lungs and spreads to nearby tissues or fluids easily. Normal cell division and growth are adequate for repairing the body’s cells. Still, when the growth is abnormal and uncontrollable, tumors tend to develop. Lung tumors may not have specific symptoms initially and are broadly divided into benign (non-cancerous) and malignant (cancerous). When a patient is diagnosed with a lung tumor, the next step is to identify whether the tumor is benign or malignant. Benign tumors are non-cancerous masses that grow slowly. The cells within the benign tumor always lie within the tumor boundary and will not spread or invade nearby tissues. Sometimes benign tumors can evolve into enormous masses of tissues with well-defined boundaries and are not harmful. However, malignant tumors are cancerous tumors that can spread to any body part through the circulatory system or the lymph. These tumors invade nearby cells easily and will not have well-defined boundaries. Owing to the rapid proliferation, these malignant tumors may eventually return even after surgery. The differentiation of benign and malignant tumors is clearly illustrated in the annotated Figure 1. Figure 1a is an annotated benign tumor, Figure 1b represents a lung CT with a malignant tumor, and Figure 1c is a normal lung CT image. There are various imaging modalities for lung cancer detection. Computed Tomography (CT), PET (Positron Emission Tomography), and MRI (Magnetic Resonance Imaging) are widely used for lung abnormality detection.
Sometimes even an experienced physician may miss a lung tumor because of an error in observation, low image quality, or wrong patient position. At times, the size of the cancer tissue and the location can also lead to a misdiagnosis. Hence, Computer Aided Detection (CAD) can be effectively utilized to lessen manual and technical errors in diagnosis and can be employed for the early detection of lung cancer. Artificial intelligence is a broad domain with unlimited applications, including clinical decision-making [1]. Various Machine Learning and Deep Learning algorithms can effectively automate disease or abnormality detection, localization, and quantification [2,3]. Initially Niki et al. [4] experimented with lung CT images to detect whether the tumor was cancerous in 2001 by utilizing CAD techniques, whereas Aberle et al. [5] analyzed how low-contrast CT images can be used in lung cancer diagnosis and thereby reducing mortality. Awai et al. [6], and Mozaffary et al. [7] utilized CAD for identifying small lung nodules from computed tomography images. Sahiner et al. [8] identified that CAD has better performance when compared to radiologists’ visualization and can eliminate manual misdiagnosis errors too.
As part of disease diagnosis, various pre-processing techniques can be applied to medical imaging. Studies reveal that pre-processing enhances the efficiency of the model. Researchers proposed various methodologies for lung cancer diagnosis using machine learning techniques. In the case of machine learning, relevant features must be manually extracted with the help of domain expertise, but sometimes inaccurate feature selection may impact the classifier. Chaganti et al. [9] suggested that in image classification, the traditional machine learning approach is superseded by deep learning because of its computational efficiency. Deep learning enables the extraction of fundamental and complicated features automatically rather than manual feature extraction. Deep learning enables us to quantify the tumor size, thereby measuring the tumor size and depth. Our proposed method employs lung CT imaging modality for malignancy detection.
In our proposed model, we choose transfer learning as a feature extractor which enables us to utilize the information learned from one dataset to another and works well even with inadequate datasets. We have utilized pre-trained VGG16 as a feature extractor by freezing all convolution layers. To reduce the computational complexity, an extremely Randomized Tree Classifier is employed as a feature selector and is fed to the Multi-Layer Perceptron (MLP) Classifier for classification. The model could improve performance when compared to other machine learning techniques.

Related Works

The advancements in artificial intelligence and machine learning can be used in various stages of lung cancer. When a patient has no symptoms, machine learning can be used for the prediction and risk assessment of lung cancer based on multiple factors such as family history, smoking history, etc. If a patient is symptomatic, these techniques can be used for disease detection, localization, and segmentation. Another use case of machine learning is in assessing disease progression and treatment response prediction. Danjuma et al. [10] conducted a study on the life expectancy prediction of lung cancer patients by utilizing Naive Bayes, Decision Tree, and Artificial Neural Network algorithms. Zehra et al. [11] compared various machine learning algorithms such as SVM, KNN, and Logistic Regression and out of which the SVM Classifier achieved the best accuracy. Radhika et al. [12] employed a comparative study on various machine learning algorithms for lung cancer detection.
Many works were conducted, including pre-processing, contrast enhancement, and segmentation. Image pre-processing and contrast improvement are vital steps in digital image processing. Segmentation techniques enable one to separate the foreground from the background. In machine learning algorithms, manual feature extraction has to be performed for image classification. Tiwari et al. [13] suggested the mask3 FCM and TWEDLNN algorithms for lung cancer diagnosis. The work involves lung segmentation by utilizing OTSU thresholding, followed by contrast enhancement. Features are extracted from the thresholded images and are fed to the classifier, and the model achieved an accuracy of 96.00%. Surendar et al. [14] proposed a hybrid deep learning model utilizing image pre-processing with non-local means filter followed by segmentation using thresholding. The features are extracted from the segmented image by a grey-level matrix accompanied by feature selection with a binary grasshopper optimizer. The selected features are classified using a hybrid deep neural network model with a sine-cosine crow search algorithm.
The significant contributions of the work are as follows:
  • Developed a framework for predicting lung cancer malignancy at an early stage. Sometimes even an expert radiologist may miss a relatively small lung tumor tissue which can be life-threatening.
  • The implication of various tree splitting criteria in ExtraTreeClassifier as feature selector is compared.
  • A comparative study is performed on various CNN models as feature extractors, with further consideration of the performance of the proposed framework with existing systems.
  • A comparison is performed with other state-of-the-art machine learning classifiers.

2. Materials and Methods

The pipeline of the proposed framework is shown in Figure 2. We have used a publicly available IQ-OTH/NCCD dataset consisting of CT images of lung cancer patients with benign, malignant, and normal cases. Initially, the images are pre-processed, followed by extracting relevant features. To reduce the computational complexity, feature selection is performed using ‘ExtraTreeClassifier’. The selected features are fed as neurons to the input layer of the Multi-Layer-perceptron (MLP) Classifier for classifying tumors into either benign, malignant, or normal cases. The models were trained and tested in Python 3.8.10 on NVIDIA Tesla V100-PCIE Graphics Processing Units (GPU) deployed on a high-performance computing cluster with 1 Teraflop.

2.1. Dataset

We utilized the lung cancer dataset from the Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases (IQ-OTH/NCCD) from Kaggle [15], which was published in 2019 and comprises CT scans of individuals who have been diagnosed with lung cancer. The collection includes images of 110 people with lung cancer in various stages, including benign, malignant, and normal. These 110 patients vary in gender, age, education, etc. Out of these, 40 patients were diagnosed with a malignancy in lung tumors, 15 patients had lung tumors that were not cancerous (benign), and 55 patients had normal lungs. Each patient has roughly 80–200 slices of a CT scan, each with a 1mm slice thickness. These patients come in various ages and genders; they do not all belong to the same category. The dataset comprises 1097 CT images of two different lung cancer types and a normal category. The first category is benign lung cancer containing 120 CT images, and the second category includes 561 malignant lung CT images. The third category is normal lung CT containing 416 images. Tumors are typically classified into either benign or malignant. Benign tumors are non-cancerous, whereas malignant tumors are cancerous. Here, 80% of the data are used for training, and 20% of the data are used for testing.

2.2. Proposed Methodology

This paper proposes a Deep Transfer learning Based Lung Cancer Detection System consisting of three tasks such as feature extraction, feature selection, and classification. The detailed architecture of the proposed methodology is illustrated in Figure 3. Initially, the Lung CT images are acquired from the ‘IQ-OTH/NCCD’ dataset. Let IOri be the initial image which is resized to 224 × 224 × 3 dimension and is represented by IRsize as in Equation (1). Keras provides the VGG16 model pre-trained on the huge ImageNet dataset.
I R s i z e = R e s i z e ( I Ori )
The resized images are inputted into the VGG16 model, with pre-trained weights on ImageNet Dataset.
The model has five blocks of 3 × 3 convolution layers with five max-pooling layers in between. The feature extraction is performed by the convolution and pooling layers of VGG16 by leaving behind the fully connected layer. Hence the Include_Top is set as False to exclude the final classification layer of VGG16. The pre-trained model outputs a three-dimensional feature stack after feature extraction.
An extremely randomized tree classifier is applied on the feature stack as a feature selector. As a result of feature selection, the training and testing of the classifier take less computational time. The model performs better because feature selection reduces execution time, eliminates duplicate and irrelevant features, and removes misleading characteristics. The model training becomes faster with fewer features, which eliminates overfitting. The model becomes less intricate, more economical, and more generic. The features selected by the extremely randomized tree classifier are given to the input layer of the Multi-Layer Perceptron (MLP) Classifier. The classifier has two fully connected layers with 100 and 70 neurons, respectively, with ReLu activation. The last output layer consists of three neurons with Softmax activation. The model is trained using the selected features, and the model predicts the testing data efficiently. The overview of ExtRanFS framework is shown in Algorithm 1.
Algorithm 1 ExtRanFS: Proposed Framework
  Input: Lung CT Image (IOri)
  Output: Classified Lung CT Image (IOut)
1:
procedure PROC1
2:
    Read each (IOri) in the Dataset
3:
     Resize   ( I Ori )   into (224,224,3)
4:
     Extract   Features   from   ( I Ori )   Using Pre-trained VGG16 to F
5:
end procedure
6:
procedure PROC2
7:
       F s Select the Best Features using ExtraTreeClassifier
8:
      Apply Label Encoding
9:
      Divide the Dataset into Training and Testing
10:
    Input Selected Features to Multi-Layer Perceptron (MLP) Classifier
11:
    Predict the lung cancer malignancy into the classes benign, malignant, or normal
12:
end procedure

2.2.1. VGG16 as Feature Extractor

The proposed methodology is implemented in a Python environment using the IQ-OTHNCCD dataset. Transfer learning is a strategy that allows the knowledge gained in one task to be applied to another, improving the performance of the second work. Transfer learning eliminates the need for an extensive dataset and training the model from scratch. A transfer learning approach has a lot to offer. When compared to a transfer-learned model, normal convolutional neural networks require days to train. Since the transfer learned model was trained on a huge dataset, it outperforms the model that was created from scratch and can already recognize key features with less training data. In our case, we have an inadequate dataset with three classes: benign, malignant, and normal. To deal with the insufficient dataset, it is highly recommended to use the features identified by another model trained on a larger dataset. This technique is known as transfer learning, where the knowledge gained on one model is re-applied to another problem. Transfer learning can be used as a feature extractor or for fine-tuning. Here we used VGG16 for feature extraction only.
Filters are used in the convolution layer to extract features from images. The kernel size and stride are the two most crucial variables. The pooling layer’s purpose is to lower the network’s spatial size so that there are fewer parameters and computations to be made. The fully connected layers in pre-trained models are replaced with the classifier as per our requirement. The VGG16, a type of convolutional neural network, was unveiled as a part of the ILSVRC competition. It has 16 layers and is trained on an ImageNet dataset with 1000 classes. The detailed architecture of the proposed framework is illustrated in Figure 4. Initially, we imported VGG16 from keras.applications with an input image size. The input images in the dataset are in the dimension of 512 × 512 × 3 RGB images. To use the dataset for the VGG16 model, the images are resized to 224 × 224 × 3. The Include = Top is set to False, which signifies omitting the topmost classification layer, and the weights are imported from ImageNet classification from which VGG16 is pre-trained. We have used VGG16 only for feature extraction, which contains five blocks of convolution layers. The first convolution block consists of two 3 × 3 Convolution layers with 64 filters, stride 1, and ReLu activation. Each convolution block is followed by a Max-pooling layer with a 2 × 2 window and stride 2. The second convolution block has two 3 × 3 Convolution layers with 128 filters, stride 1, and ReLu activation. The third block of convolution has three 3 × 3 Convolution layers with 256 filters, stride 1, and ReLu activation. The fourth and fifth blocks of convolution have three 3 × 3 Convolution layers, 512 filters, stride 1, and ReLu activation.
The detailed illustration of feature maps generated at each convolution block is projected as Figure 5. The pre-trained model’s convolution and max-pooling layers gather features from the image and output a 3-Dimensional feature stack. The first five convolution layers of the pre-trained VGG16 extracted visual features necessary for classification. From this, it is evident that the initial layers extract more prominent and finer details when compared to the last convolution layers. From the initial convolution layers to the final layers, the model can abstract the image details into broader ideas to lead to categorizing the image. Before utilizing these features in machine learning classifiers, they will be flattened to 1-Dimension. The extracted flattened features are given to the Extremely Randomized Classifier for feature selection. The original VGG16 architecture has three fully connected layers with 4096 neurons at two dense layers and 1000 neurons in the output layer. The selected features are fed to the input layer of the MLP Classifier with 4323 neurons. We removed the fully connected classifier of VGG16, and instead of the last three fully connected layers (classifier part) of the VGG16 model, we used an MLP Classifier with two dense layers as fully connected layers with 100 and 70 neurons, respectively. The output layer has a fully connected layer with three neurons with Softmax Activation.

2.2.2. Extremely Randomized Ensemble Classifier as Feature Selector

An extremely Randomized Trees Classifier is an ensemble learning algorithm that resembles the Random Forest Classifier, also known as Extra Trees Classifier. This method accomplishes classification by aggregating the output of numerous decision trees that make up a forest. The decision trees in Extra Tree Classifier are constructed from the training data, and the splitting criteria are decided mainly by Gini Index or Entropy. Gini Index and Entropy are used for assessing the information gain and are considered as a measure of the impurity of a node. When splitting the node, if a node has multiple classes, the class is highly impure. Entropy is calculated using Equation (2), and Gini Index is calculated in Equation (2). Pc(i) is the probability of class c(i) in a node, and n is the number of classes.
E n t r o p y = i = 1 n P ( c i l o g 2 ( P ( c i ) ) )
G i n i = 1 i = 1 n P 2 c ( i )
If the Gini Index is 0, it means that when two classes are split, just one class shows on one side while the other class is represented on the other. This division is thought to be perfect and is considered a pure classification. The Gini Index calculation is simple when compared to the Entropy calculation. In Entropy, in addition to probability, the log of the probabilities has to be computed and hence will not be computationally efficient. The Extra Tree Classifier fits several extra trees on different samples of the dataset and utilizes averaging ensemble technique to increase the accuracy and thereby control overfitting. The ExtraTreeClassifier is an ensemble technique similar to the Random Forest. The bootstrapping technique is used in the Random Forest Classifier, and the sub-samples are created from the input using replacement. Instead of using bootstrapping, Extra Trees use the entire training sample, which minimizes bias. Random Forest employs an optimal split when dividing trees, whereas Extra Trees uses a randomized split. Random Forest and Extra Trees chose the best feature after deciding on the split point. Extra Trees, when compared to the Random Forest, adds randomness by selecting each node’s split point at random, which unquestionably lowers variance.
Extra Trees are faster than Random Forest. Hence this technique is more computationally efficient and takes less time to execute. The classifier employs averaging to improve the ensemble classifier accuracy and will also lessen over-fitting by fitting several randomized decision trees. We can adjust the number of trees and the tree-splitting criteria. The number of trees is generally in the range of 10 to 100, and as the number of trees increases, the performance improves, but obviously, the computation time will be more. It is obvious that beyond a certain no of trees, the results are not improving. Either Gini or Entropy can be used as the tree-splitting criterion. Entropy and the Gini Index are used to evaluate information gain and are seen as indicators of a node’s impurity. If a node contains several classes after splitting, the class is quite impure. When contrasted with the Entropy computation, the Gini Index calculation is simpler. Entropy is not computationally efficient since, in addition to probability, the log of the probabilities also needs to be computed.

2.2.3. Classification

The features selected from the Extremely Randomized Tree Classifier are inputted into the Multi-Layer Perceptron(MLP) Classifier. The input is flattened in the sequential model to produce a single input-single output model. The input layer consists of 4323 neurons comprising the features selected by ExtraTreeClassifier. The next two dense layers make up the fully connected hidden layers, and the final output layer, which is the final dense layer, is accountable for deciding the image class. The classifier has two fully connected dense layers of 100 and 70 neurons, respectively. Each dense layer is activated by ReLu activation. The model is optimized using the Adam optimizer, and loss is calculated using sparse categorical cross-entropy. The last output layer had three neurons and was activated by Softmax activation. The Batch size is fixed at 32, and the number of epochs is 15. The proposed MLP Classifier classified the input image into either of the three classes with an accuracy of 99.09%.

2.2.4. Evaluation

The assessment of the model is measured in terms of quality metrics such as accuracy, precision, sensitivity, specificity, and F1-Score. The performance of the classifier is evaluated using the confusion matrix. The accuracy and loss of the model, along with the Receiver Operating Characteristics (ROC) Curve, is also illustrated in Section 3.

3. Results and Discussion

We discuss the experiments conducted by various pre-trained models as feature extractors. The performance of the MLP Classifier is compared with other state-of-the-art methodologies as classifiers. On NVIDIA Tesla V100-PCIE Graphics Processing Units (GPU) set up on a high-performance computing cluster with 1 Teraflop, all models were trained and tested in Python 3.8.10.

3.1. Feature Extraction by VGG16

The proposed methodology is implemented in a Python environment using the IQ-OTHNCCD dataset. In this study, transfer learning is applied as a feature extractor to deal with insufficient datasets. We utilized VGG16 as a feature extractor by excluding the top classification layers. Figure 5 represents the visualization of feature maps extracted from each convolution block of the VGG16 feature extractor.
To compare the feature extraction with other state-of-the-art methodologies, the experiment was performed with different pre-trained models as feature extractors, such as InceptionV3, Xception, and MobileNetV2. Google unveiled the Inception V1 architecture, commonly known as GoogleNet, in 2015. Later, Inception-V2, a variation of Inception V1 with Batch normalization, was developed. The Inception-V3 with BN auxiliary is an improved version of the Inception-V2. The main design of Xception consists of three flows: entry, middle, and exit, each with Convolution and Separable Convolution block along with Residual connection. MobileNetV2 is another CNN developed by Google with residual skip connections between bottleneck layers. The results are discussed in Section 3.3.

3.2. Feature Selection by ExtraTreeClassifier

The features extracted by VGG16 are fed to the ExtraTree Classifier, which is a model-based feature selection technique. The Extremely Random Tree Classifier is an ensemble technique that sorts the features with the most votes and generates numerous trees built randomly from the training dataset. The decision trees are fitted to the complete dataset rather than utilizing a bootstrap technique, and the nodes are divided randomly. The splitting is based on randomness, thereby lowering variance. The model’s accuracy will be reduced when irrelevant features are processed and will take more time for computation. The execution time for the classification in this study has been significantly shortened by feature selection employing extra trees classifiers. In ExtraTreeClassifier, the number of trees in the forest can be in the range of 10 to 100, and based on the number of trees, the features selected will be different. We tried other tree-splitting criteria, and the number of features chosen depends on the number of trees in the forest as in Table 1. The highest performance is achieved when the tree splitting criteria is Gini and the trees in the forest are set as 50.
We experimented with different pre-trained models as feature extractors, and the features selected by ExtraTreeClassifier are projected in Table 2. The number of trees in the ExtraTreeClassifier is 50, and the tree splitting criteria is chosen as ‘Gini’.

3.3. Classification

The features extracted by VGG16 are selected by utilizing ExtraTreeClassifier and are fed to the Multi-Layer Perceptron (MLP) model as the classifier. The MLP Classifier has an input layer, two hidden layers, and an output layer. The input layer consists of 4323 neurons comprising the features selected by ExtraTreeClassifier. The first and second hidden layers consist of 100 and 70 neurons, respectively, with ReLu activation. The output layer has three neurons with Softmax activation. The VGG16+MLP achieved an accuracy of 99.09%. Accuracy, precision, recall, and F1-Score are standard evaluation metrics that we have used to assess the classification performance. The ratio of accurate predictions to all the input images is called accuracy and is calculated using Equation (4).
A c c u r a c y = T P + T N T P + T N + F P + F N
The terms recall, or True Positive Rate (TPR) are also used to describe sensitivity. Equation (5) calculates it, giving an idea of the percentage of positive samples that were accurately predicted as positive.
S e n s i t i v i t y = T P T P + F N
The percentage of all negative samples that were correctly predicted to be negative is what is known as the “True Negative Rate” (TNR) or Specificity and is computed using Equation (6).
S p e c i f i c i t y = T N T N + F P
Equation (7) can be used to measure precision, which indicates what percentage of positive predictions came true.
P r e c i s i o n = T P T P + F P
An evaluation is performed with different adaptive optimizers in the classifier’s performance as in Table 3, and the best accuracy is achieved with the Adam optimizer.
A comparison is performed with other existing pre-trained models as feature extractors, and the results are illustrated in Table 4.
From Table 4, it is evident that the proposed framework ExtRanFS outperformed all other models regarding quality metrics such as accuracy, precision, recall, and F1-Score. Similarly, we have compared with other existing state-of-the-art feature selectors with and without ExtraTreeClassifier as feature selectors.

Computational Complexity

An analysis of the mathematical possibilities for effective computer learning is called ‘Computational Complexity of Machine Learning Algorithms’. The total number of trainable parameters accounts for the computational complexity of the model. Table 5 projects the number of trainable parameters and the run time with and without feature selection. With feature selection, the number of trainable parameters and the run time is drastically reduced compared to the run time without feature selection.
When evaluating classification performance, the confusion matrix compares real and predicted values. We obtain a 3 × 3 confusion matrix since there are three classes. The confusion matrix of the classifier upon feature extraction from various pre-trained models is illustrated in Figure 6. Figure 6a represents the confusion matrix of MLP upon feature extraction from VGG16. There is one wrong classification in the benign class and one in the normal case.
Figure 6b illustrates the confusion matrix of MLP upon feature extraction from Xception and has eight wrong predictions in the benign class, along with one wrong prediction in the malignant case. Figure 6c represents the confusion matrix of MLP classifier upon feature extraction from MobileNetV2 and has five wrong predictions in the benign class and one in the normal class. Figure 6d represents the confusion matrix of MLP upon feature extraction from InceptionV3 with six wrong predictions in the benign class and four in the malignant class. From this, it is evident that the MLP classifier with VGG16 as feature extractor outperformed well when compared to other pre-trained models as feature extractors.
We experimented with different state-of-the-art classifiers upon feature extraction from VGG16. Figure 7a is the confusion matrix of the MLP Classifier with two misclassifications, one in benign and the other in normal cases. Figure 7b has four wrong predictions in the benign category, whereas Figure 7c has a total of eleven misclassifications. Figure 7d shows the confusion matrix of the VGG16+KNN Classifier with fifteen wrong predictions. The Decision Tree Classifier had seven wrong classifications Figure 7e. In this scenario also, the MLP classifier with VGG16 outperformed well.
The accuracy and loss curves are used to evaluate the model’s progress. These are prominently used quality metrices in machine learning. When there is a gap in accuracy between training and testing, overfitting is evident. The overfitting increases as the gap widens. Figure 8 shows the accuracy and loss of the model. The figure shows that the model is not overfitted. According to the ‘Model Accuracy’ graph, training does not require as many epochs as is suggested by the flattening of the curve after the first few rapid rises in accuracy.
Figure 8a shows the accuracy and loss curve of the MLP classifier upon feature extraction from VGG16. From the accuracy curve, it is evident that the model attained around 97.00% accuracy nearly after the second epoch. Since there is no drastic gap between training and testing accuracy, we can say that the model is not overfitted. Figure 8b projects the accuracy loss curves of MLP classifier when pre-trained from Xception for feature extraction. Similarly, Figure 8c illustrates the accuracy loss curve of the MLP classifier upon feature extraction from MobileNetV2. Figure 8d shows the accuracy-loss curve of the MLP classifier on feature extraction from pre-trained InceptionV3.
The Receiver Operating Characteristics (ROC) of the classifier is constructed by plotting Sensitivity against Specificity. Classifiers perform better when their curves are closer to the top-left corner. From Figure 9, it is evident that the classifier performed well with different pre-trained models as feature extractors.

4. Discussion

Researchers proposed various deep architectures for the detection of lung cancer since deep learning eliminates the necessity of feature extraction. Deep learning works as a black box that performs both feature extraction and classification. Sharaf et al. [16] utilized ExtraTreeClassifier and metaheuristics to classify email. Diana et al. [17] performed leukocyte classification by extracting features using ResNet50, feature selection by ExtraTreeClassifier, followed by SVM Classifier and attained an accuracy of 90.76%. From CT scans, Asuntha et al. [18] identified textural, geometric, and intensity features. Using Convolutional Neural Networks (CNN), these optimized features were employed for categorization and obtained an accuracy of 95.62%. The model could effectively reduce computational complexity. Song et al. [19] experimented with three different deep neural networks for classification and out of which CNN outperformed other models with an accuracy of 84.15%. According to Dutta et al. [20], Convolutional Neural Networks can be used for medical image classification either by training from scratch or by utilizing transfer learning. Sajja et al. [21] used the LIDC dataset for lung cancer detection using GoogleNet and achieved a testing accuracy of 99.00%. In contrast, Rahul et al. [22] used pre-trained models for feature extraction in the survival prediction of adenocarcinoma patients and resulted in 90.00% accuracy.
Gao et al. [23] used pre-trained VGG16 on the LUNA16 dataset for lung nodule detection and obtained an accuracy of 96.86%. In contrast, Chon et al. [24] performed various operations such as segmentation, nodule detection, and classification and achieved an AUC of 0.83. Razeq et al. [25] concentrated on how CNNs classified lung nodules according to the size of the input lung CT, whereas Abdulla et al. [26] performed lung cancer detection using a deep neural network. Chang et al. [27] suggested a two-staged Convolutional Network for lung cancer detection by integrating noise removal. Satti et al. [28] proposed a filter-based noise removal technique for eliminating the commonly occurring impulse noise in medical images. Kaviya et al. [29] conducted a detailed literature review on how artificial intelligence can be effectively utilized in detecting lung cancer.
The authors of [30,31,32,33] used Artificial Neural Networks to classify lung nodules or lung cancer, but the significant limitations were the manual handcrafted feature selection technique. Manual feature selection requires domain expertise, and choosing inappropriate or irrelevant features may result in building a less robust model. The authors of [34] employed LeNet-5 architecture with 10-fold cross-validation and attained a validation accuracy of 97.04%. Wang et al. [35] utilized a transfer learning-based strategy to classify lung cancer, and the model achieved an accuracy of 85.71%. We have evaluated different machine learning models as classifiers, such as Support Vector Machine (SVM), Random Forest (RF), K-Nearest Neighbor (KNN), and Decision Tree (DT), and the results are projected in Table 6. We utilized LinearSVC from sklearn.svm and the loss was calculated using a squared hinge. The one-vs-rest pattern provides the multiclass classification with a penalty selected as l2, and the tolerance for the stopping criterion is set as 1 × 10 5 . With 50 trees in the forest, the Random Forest meta-estimator fits a variety of decision tree classifiers to different dataset subsamples. The number of neighbors in the K-Nearest Classifier is set as 7, whereas the tree splitting criteria in DecisionTreeClassifier is set as Gini.
Table 7 compares our proposed framework with other existing state-of-the-art models on the same dataset ‘IQ-OTHNCCD’. The results demonstrate that our proposed classifier obtained the best accuracy of 99.09% when pre-trained with VGG16 as a feature extractor.
AlexNet CNNs were used by Al-Yasriy et al. [36] to identify and categorize lung cancer. The model that was developed had a high accuracy of 93.54%. Nishio et al. [39] tried to classify lung nodules by implementing a transfer-learning-based deep CNN; the authors concluded that larger input images and transfer learning enhanced the Computer Aided Diagnosis(CAD). Humayun et al. [37] incorporated data augmentation on the ‘IQ-OTH/NCCD dataset’, and the classification on pre-trained VGG16 achieved an accuracy of 98.83% for the training data. Still, for testing data, the accuracy was only 86.45% indicating an overfitting model. Kareem et al. [40] utilized the same ‘IQ-OTH/NCCD dataset’ for lung cancer classification by utilizing image enhancement, segmentation, and feature extraction. The features were classified using the SVM classifier, and the authors achieved an accuracy of 89.88%. AL-Huseiny et al. [38] performed various pre-processing involving texture analysis, morphological operation, and ROI Extraction. They used pre-trained GoogleNet architecture for lung cancer classification and obtained an accuracy of 94.38%.

5. Conclusions

Transfer learning is the process of using a model that has already been learned to solve a new problem. It is currently highly popular in deep learning since it can train deep neural networks using a minimal amount of data. Transfer learning eliminates the requirement for initial training from scratch. The proposed methodology classified lung cancers into three groups. For the classification challenge, we employed the VGG16 pre-trained model as a feature extractor. The fully connected dense layers of VGG16 are replaced with an MLP Classifier. The proposed classifier outperformed other existing state-of-the-art methodologies with the best accuracy of 99.09%.

Author Contributions

Conceptualization, N.V.R.; methodology, N.V.R.; software, N.V.R.; validation, N.V.R. and V.C.S.S.; formal analysis, N.V.R.; investigation, N.V.R.; resources, N.V.R.; data curation, N.V.R.; writing—original draft preparation, N.V.R.; writing—review and editing, V.C.S.S.; visualization, N.V.R.; supervision, V.C.S.S.; project administration, V.C.S.S.; funding acquisition, N.V.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

https://www.kaggle.com/datasets/waseemnagahhenes/lung-cancer-dataset-iq-othnccd (accessed on 3 January 2023). The source code is available as ‘Automated Lung Cancer Malignancy Detection System—Source Code’ (available since 25 June 2023)—at http://www.mirworks.in/downloads.php.

Acknowledgments

The authors would like to extend gratitude to all researchers affiliated with the Machine Intelligence Research (MIR) Laboratory(Kerala University) for their support during this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Christie, J.R.; Lang, P.; Zelko, L.M.; Palma, D.A.; Abdelrazek, M.; Mattonen, S.A. Artificial Intelligence in Lung Cancer: Bridging the Gap Between Computational Power and Clinical Decision-Making. Can. Assoc. Radiol. J. 2021, 72, 86–97. [Google Scholar] [CrossRef] [PubMed]
  2. Rajaraman, S.; Ganesan, P.; Antani, S. Deep learning model calibration for improving performance in class-imbalanced medical image classification tasks. PLoS ONE 2022, 17, e0262838. [Google Scholar] [CrossRef] [PubMed]
  3. Wang, S.; Yang, D.M.; Rong, R.; Zhan, X.; Fujimoto, J.; Liu, H.; Minna, J.; Wistuba, I.I.; Xie, Y.; Xiao, G. Artificial intelligence in lung cancer pathology image analysis. Cancers 2019, 11, 1673. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Niki, N.; Kawata, Y.; Kubo, M. A CAD system for lung cancer based on CT image. Int. Congr. Ser. 2001, 1230, 631–638. [Google Scholar] [CrossRef]
  5. de Koning, H.J.; van der Aalst, C.M.; de Jong, P.A.; Scholten, E.T.; Nackaerts, K.; Heuvelmans, M.A.; Lammers, J.W.J.; Weenink, C.; Yousaf-Khan, U.; Horeweg, N.; et al. Reduced Lung-Cancer Mortality with Volume CT Screening in a Randomized Trial. N. Engl. J. Med. 2020, 382, 503–513. [Google Scholar] [CrossRef]
  6. Kazuo, A.; Awai, K.; Murao, K.; Ozawa, A.; Komi, M.; Hayakawa, H.; Hori, S.; Nishimura, Y. Pulmonary nodules at chest CT: Effect of computer-aided diagnosis on radiologists’ detection performance. Radiology 2004, 230, 347–352. [Google Scholar]
  7. Mozaffary, A.; Trabzonlu, T.A.; Lombardi, P.; Seyal, A.R.; Agrawal, R.; Yaghmai, V. Integration of fully automated computer-aided pulmonary nodule detection into CT pulmonary angiography studies in the emergency department: Effect on workflow and diagnostic accuracy. Emerg. Radiol. 2019, 26, 609–614. [Google Scholar] [CrossRef]
  8. Sahiner, B.; Chan, H.P.; Hadjiiski, L.M.; Cascade, P.N.; Kazerooni, E.A.; Chughtai, A.R.; Poopat, C.; Song, T.; Frank, L.; Stojanovska, J.; et al. Effect of CAD on radiologists’ detection of lung nodules on thoracic CT scans: Analysis of an observer performance study by nodule size. Acad. Radiol. 2009, 16, 1518–1530. [Google Scholar] [CrossRef] [Green Version]
  9. Chaganti, S.Y.; Nanda, I.; Pandi, K.R.; Prudhvith, T.G.N.R.S.N.; Kumar, N. Image Classification using SVM and CNN. In Proceedings of the 2020 International Conference on Computer Science, Engineering and Applications (ICCSEA), Gunupur, India, 13–14 March 2020; pp. 1–5. [Google Scholar]
  10. Danjuma, K.J. Performance Evaluation of Machine Learning Algorithms in Post-operative Life Expectancy in the Lung Cancer Patients. arXiv 2015, arXiv:1504.04646. [Google Scholar]
  11. Karhan, Z.; Tunç, T. Lung Cancer Detection and Classification with Classification Algorithms. IOSR J. Comput. Eng. (IOSR-JCE) 2016, 18, 71–77. [Google Scholar]
  12. Radhika, P.R.; Nair, R.A.; Veena, G.A. Comparative Study of Lung Cancer Detection using Machine Learning Algorithms. In Proceedings of the 2019 IEEE International Conference on Electrical, Computer and Communication Technologies (ICECCT), Coimbatore, India, 20–22 February 2019; pp. 1–4. [Google Scholar]
  13. Tiwari, L.; Raja, R.; Awasthi, V.; Miri, R.; Sinha, G.R.; Alkinani, M.H.; Polat, K. Detection of lung nodule and cancer using novel Mask-3 FCM and TWEDLNN algorithms. Measurement 2021, 172, 108882. [Google Scholar] [CrossRef]
  14. Surendar, P.; Ponni Bala, M. Diagnosis of lung cancer using hybrid deep neural network with adaptive sine cosine crow search algorithm. J. Comput. Sci. 2021, 53, 101374. [Google Scholar]
  15. Hamdalla, F. The IQ-OTH/NCCD Lung Cancer Dataset. Kaggle. 2020. Available online: https://www.kaggle.com/datasets/hamdallak/the-iqothnccd-lung-cancer-dataset (accessed on 3 January 2023).
  16. Sharaff, A.; Gupta, H. Extra-Tree Classifier with Metaheuristics Approach for Email Classification. In Advances in Computer Communication and Computational Sciences; Advances in Intelligent Systems and Computing; Springer: Singapore, 2019; pp. 189–197. [Google Scholar]
  17. Baby, D.; Devaraj, S.J.; Hemanth, J. Leukocyte classification based on feature selection using extra trees classifier: A transfer learning approach. Turk. J. Electr. Eng. Comput. Sci. 2021, 29, 2742–2757. [Google Scholar] [CrossRef]
  18. Asuntha, A.; Srinivasan, A. Deep learning for lung Cancer detection and classification. Multimed. Tools Appl. 2020, 79, 7731–7762. [Google Scholar] [CrossRef]
  19. Song, Q.; Zhao, L.; Luo, X.; Dou, X. Using deep learning for classification of lung nodules on computed tomography images. J. Healthc. Eng. 2017, 2017, 8314740. [Google Scholar] [CrossRef] [Green Version]
  20. Dutta, P.; Upadhyay, P.; De, M.; Khalkar, R.G. Medical Image Analysis using Deep Convolutional Neural Networks: CNN Architectures and Transfer Learning. In Proceedings of the 2020 International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 26–28 February 2020. [Google Scholar]
  21. Sajja, T.; Devarapalli, R.; Kalluri, H. Lung Cancer Detection Based on CT Scan Images by Using Deep Transfer Learning. Trait. Signal 2019, 36, 339–344. [Google Scholar] [CrossRef]
  22. Rahul, P.; Paul, R.; Hawkins, S.H.; Balagurunathan, Y.; Schabath, M.; Gillies, R.J.; Hall, L.O.; Goldgof, D.B. Deep Feature Transfer Learning in Combination with Traditional Features Predicts Survival Among Patients with Lung Adenocarcinoma. Tomography 2016, 2, 388–395. [Google Scholar]
  23. Gao, J.; Jiang, Q.; Zhou, B.; Chen, D. Lung nodule detection using convolutional neural networks with transfer learning on CT images. Comb. Chem. High Throughput Screen. 2021, 24, 814–824. [Google Scholar] [CrossRef]
  24. Chon, A.; Balachandar, N.; Lu, P. Deep Convolutional Neural Networks for Lung Cancer Detection; Standford University: Stanford, CA, USA, 2017; pp. 1–9. [Google Scholar]
  25. Razeq, K.; Abdel-Razeq, H.; Attiga, F.; Mansour, A. Cancer care in Jordan. Hematol. Stem Cell Ther. 2015, 8, 64–70. [Google Scholar] [CrossRef] [Green Version]
  26. Ibrahim, N.; Abdullah, A.A.; Shaharum, S.M. Lung Cancer Cell Classification method Using Artificial Neural Network. Inf. Eng. Lett. 2012, 2, 48. [Google Scholar]
  27. Chang, Y.; Yan, L.; Chen, M.; Fang, H.; Zhong, S. Two-Stage Convolutional Neural Network for Medical Noise Removal via Image Decomposition. IEEE Trans. Instrum. Meas. 2019, 69, 2707–2721. [Google Scholar] [CrossRef]
  28. Satti, P.; Sharma, N.; Garg, B. Min-Max Average Pooling Based Filter for Impulse Noise Removal. IEEE Signal Process. Lett. 2020, 27, 1475–1479. [Google Scholar] [CrossRef]
  29. Kaviya, S.; Sathyakumar, K.; Munoz, M.; Singh, J.; Hussain, N.; Babu, B.A.; Babu, B. Automated Lung Cancer Detection Using Artificial Intelligence (AI) Deep Convolutional Neural Networks: A Narrative Literature Review. Cureus 2020, 12, e10017. [Google Scholar]
  30. Khehrah, N.; Farid, M.S.; Bilal, S.; Khan, M.H. Lung Nodule Detection in CT Images Using Statistical and Shape-Based Features. J. Imaging 2020, 6, 6. [Google Scholar] [CrossRef] [Green Version]
  31. Xie, Y.; Zhang, J.; Xia, Y.; Fulham, M.; Zhang, Y. Fusing texture, shape and deep model-learned information at decision level for automated classification of lung nodules on chest CT. Inf. Fusion 2018, 42, 102–110. [Google Scholar] [CrossRef]
  32. Nasser, I.M.; Abu-Naser, S.S. Lung Cancer Detection Using Artificial Neural Network. Int. J. Eng. Inf. Syst. (IJEAIS) 2019, 3, 17–23. [Google Scholar]
  33. Miah, M.B.A.; Yousuf, M.A. Detection of lung cancer from CT image using image processing and neural network. In Proceedings of the 2015 International Conference on Electrical Engineering and Information Communication Technology (ICEEICT), Savar, Bangladesh, 21–23 May 2015. [Google Scholar]
  34. Zhang, S.; Sun, F.; Wang, N.; Zhang, C.; Yu, Q.; Zhang, M.; Babyn, P.; Zhong, H. Computer-Aided Diagnosis (CAD) of Pulmonary Nodule of Thoracic CT Image Using Transfer Learning. J. Digit. Imaging 2019, 32, 995–1007. [Google Scholar] [CrossRef]
  35. Wang, S.; Dong, L.; Wang, X.; Wang, X. Classification of pathological types of lung cancer from CT images by deep residual neural networks with transfer learning strategy. Open Med. 2020, 15, 190–197. [Google Scholar] [CrossRef]
  36. Al-Yasriy, H.F.; AL-Husieny, M.S.; Mohsen, F.Y.; Khalil, E.A.; Hassan, Z.S. Diagnosis of Lung Cancer Based on CT Scans Using CNN. In IOP Conference Series: Materials Science and Engineering; IOP Publishing: Bristol, UK, 2020; Volume 928, p. 022035. [Google Scholar]
  37. Humayun, M.; Sujatha, R.; Almuayqil, S.N.; Jhanjhi, N.Z.A. Transfer Learning Approach with a Convolutional Neural Network for the Classification of Lung Carcinoma. Healthcare 2022, 10, 10061058. [Google Scholar] [CrossRef]
  38. AL-Huseiny, M.S.; Sajit, A.S. Transfer learning with GoogLeNet for detection of lung cancer. Indones. J. Electr. Eng. Comput. Sci. 2021, 22, 1078–1086. [Google Scholar] [CrossRef]
  39. Nishio, M.; Sugiyama, O.; Yakami, M.; Ueno, S.; Kubo, T.; Kuroda, T.; Togashi, K. Computer-aided diagnosis of lung nodule classification between benign nodule, primary lung cancer, and metastatic lung cancer at different image size using deep convolutional neural network with transfer learning. PLoS ONE 2018, 13, e0200721. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Kareem, H.F.; AL-Husieny, M.S.; Mohsen, F.Y.; Khalil, E.A.; Hassan, Z.S. Evaluation of SVM performance in the detection of lung cancer in marked CT scan dataset. Indones. J. Electr. Eng. Comput. Sci. 2021, 21, 1731–1738. [Google Scholar] [CrossRef]
Figure 1. Annotation of malignancy in lung tumor.
Figure 1. Annotation of malignancy in lung tumor.
Diagnostics 13 02206 g001
Figure 2. Pipeline of the proposed methodology.
Figure 2. Pipeline of the proposed methodology.
Diagnostics 13 02206 g002
Figure 3. Architecture of ExtRanFS framework.
Figure 3. Architecture of ExtRanFS framework.
Diagnostics 13 02206 g003
Figure 4. Detailed architecture of the proposed framework.
Figure 4. Detailed architecture of the proposed framework.
Diagnostics 13 02206 g004
Figure 5. Visualization of the feature maps extracted from each block in the VGG16 Model.
Figure 5. Visualization of the feature maps extracted from each block in the VGG16 Model.
Diagnostics 13 02206 g005
Figure 6. Confusion matrix of the classifier upon feature extraction from different pre-trained models.
Figure 6. Confusion matrix of the classifier upon feature extraction from different pre-trained models.
Diagnostics 13 02206 g006
Figure 7. Confusion matrix of various classifiers upon feature extraction from VGG16.
Figure 7. Confusion matrix of various classifiers upon feature extraction from VGG16.
Diagnostics 13 02206 g007
Figure 8. Accuracy and loss of the MLP Classifier upon feature extraction from various pre-trained models.
Figure 8. Accuracy and loss of the MLP Classifier upon feature extraction from various pre-trained models.
Diagnostics 13 02206 g008
Figure 9. ROC curve of the classifier upon various pre-trained models as feature extractors.
Figure 9. ROC curve of the classifier upon various pre-trained models as feature extractors.
Diagnostics 13 02206 g009
Table 1. Features selected by different tree splitting criteria.
Table 1. Features selected by different tree splitting criteria.
Trees in ForestSelected Features (Gini)Selected Features (Entropy)
1012851013
2022872009
3030742950
4037223690
5043234104
6048764562
7052255143
8063185940
9068966231
10072006921
Table 2. Feature extraction and feature selection.
Table 2. Feature extraction and feature selection.
Pre-Trained ModelFeatures ExtractedFeatures Selected
VGG1625,0084323
Xception100,3526140
MobileNetV262,7205191
InceptionV351,2005949
Table 3. Comparison of ExtRanFS framework (VGG16+MLP) with various adaptive optimizers.
Table 3. Comparison of ExtRanFS framework (VGG16+MLP) with various adaptive optimizers.
OptimzerAccuracy (%)Precision (%)Recall (%)F1 Score (%)
Adam99.0998.6698.6698.66
Adagrad94.0091.6688.3389.66
Adadelta65.0052.3351.0051.66
RMSprop66.0071.3368.3369.33
Table 4. Comparison of ExtRanFS framework with other state-of-the-art methodologies with feature selection.
Table 4. Comparison of ExtRanFS framework with other state-of-the-art methodologies with feature selection.
Classification ModelAccuracy (%)Precision (%)Recall (%)F1 Score (%)
VGG16+MLP (Proposed)99.0998.6698.6698.66
Xception+MLP96.0096.6790.3393.00
MobileNetV2+MLP97.0096.6794.0095.00
InceptionV3+MLP94.0090.3389.6689.66
Table 5. Comparison of ExtRanFS framework with other state-of-the-art classifiers with and without feature selection in terms of run time.
Table 5. Comparison of ExtRanFS framework with other state-of-the-art classifiers with and without feature selection in terms of run time.
Classification ModelWith Feature SelectionWithout Feature Selection
Trainable ParametersRun Time (s)Trainable ParametersRun Time (s)
VGG16+MLP (Proposed)445,1833002,516,183660
Xception+MLP835,58360010,042,583780
MobileNetV2+MLP762,1831806,279,383300
InceptionV3+MLP867,4833605,127,383660
Table 6. Comparison of ExtRanFS Framework with other state-of-the-art Classifiers.
Table 6. Comparison of ExtRanFS Framework with other state-of-the-art Classifiers.
Classification ModelAccuracy (%)Precision (%)Recall (%)F1 Score (%)
VGG16+MLP (Proposed)99.0998.3398.3398.33
VGG16+SVM98.6396.6697.3397.00
VGG16+RF95.9089.0096.3391.66
VGG16+KNN93.1882.3388.6684.66
VGG16+DT93.6391.0092.3391.33
Table 7. Comparison of ExtRanFS framework with other state-of-the-art Classifiers.
Table 7. Comparison of ExtRanFS framework with other state-of-the-art Classifiers.
AuthorArchitectureAccuracy (%)Sensitivity (%)Specificity (%)
Al-Yasriy et al. [36]AlexNet93.5495.7195.00
Humayun et al. [37]VGG1698.83Not SpecifiedNot Specified
AL-Huseiny et al. [38]GoogLeNet94.3895.0893.70
ExtRanFS (Proposed)VGG16 +MLP99.0998.6698.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

V. R., N.; Chandra S. S., V. ExtRanFS: An Automated Lung Cancer Malignancy Detection System Using Extremely Randomized Feature Selector. Diagnostics 2023, 13, 2206. https://doi.org/10.3390/diagnostics13132206

AMA Style

V. R. N, Chandra S. S. V. ExtRanFS: An Automated Lung Cancer Malignancy Detection System Using Extremely Randomized Feature Selector. Diagnostics. 2023; 13(13):2206. https://doi.org/10.3390/diagnostics13132206

Chicago/Turabian Style

V. R., Nitha, and Vinod Chandra S. S. 2023. "ExtRanFS: An Automated Lung Cancer Malignancy Detection System Using Extremely Randomized Feature Selector" Diagnostics 13, no. 13: 2206. https://doi.org/10.3390/diagnostics13132206

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop