Next Article in Journal
Automated Quantification of Pneumonia Infected Volume in Lung CT Images: A Comparison with Subjective Assessment of Radiologists
Next Article in Special Issue
Machine Learning Diffuse Optical Tomography Using Extreme Gradient Boosting and Genetic Programming
Previous Article in Journal
The Role of Fascial Tissue Layer in Electric Signal Transmission from the Forearm Musculature to the Cutaneous Layer as a Possibility for Increased Signal Strength in Myoelectric Forearm Exoprosthesis Development
Previous Article in Special Issue
Multi-Layered Non-Local Bayes Model for Lung Cancer Early Diagnosis Prediction with the Internet of Medical Things
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization System Based on Convolutional Neural Network and Internet of Medical Things for Early Diagnosis of Lung Cancer

by
Yossra Hussain Ali
1,
Varghese Sabu Chooralil
2,
Karthikeyan Balasubramanian
3,
Rajasekhar Reddy Manyam
4,
Sekar Kidambi Raju
3,*,
Ahmed T. Sadiq
1 and
Alaa K. Farhan
1
1
Department of Computer Sciences, University of Technology, Bagdad 110066, Iraq
2
Department of Computer Sciences & Engineering, Rajagiri School of Engineering & Technology, Kochi 682039, Kerala, India
3
School of Computing, SASTRA Deemed University, Thanjavur 613401, Tamil Nadu, India
4
Amrita School of Computing, Amrita Vishwa Vidyapeetham, Amaravati Campus, Amaravati 522503, Andhra Pradesh, India
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(3), 320; https://doi.org/10.3390/bioengineering10030320
Submission received: 28 November 2022 / Revised: 22 February 2023 / Accepted: 24 February 2023 / Published: 2 March 2023
(This article belongs to the Special Issue Artificial Intelligence in Advanced Medical Imaging)

Abstract

:
Recently, deep learning and the Internet of Things (IoT) have been widely used in the healthcare monitoring system for decision making. Disease prediction is one of the emerging applications in current practices. In the method described in this paper, lung cancer prediction is implemented using deep learning and IoT, which is a challenging task in computer-aided diagnosis (CAD). Because lung cancer is a dangerous medical disease that must be identified at a higher detection rate, disease-related information is obtained from IoT medical devices and transmitted to the server. The medical data are then processed and classified into two categories, benign and malignant, using a multi-layer CNN (ML-CNN) model. In addition, a particle swarm optimization method is used to improve the learning ability (loss and accuracy). This step uses medical data (CT scan and sensor information) based on the Internet of Medical Things (IoMT). For this purpose, sensor information and image information from IoMT devices and sensors are gathered, and then classification actions are taken. The performance of the proposed technique is compared with well-known existing methods, such as the Support Vector Machine (SVM), probabilistic neural network (PNN), and conventional CNN, in terms of accuracy, precision, sensitivity, specificity, F-score, and computation time. For this purpose, two lung datasets were tested to evaluate the performance: Lung Image Database Consortium (LIDC) and Linear Imaging and Self-Scanning Sensor (LISS) datasets. Compared to alternative methods, the trial outcomes showed that the suggested technique has the potential to help the radiologist make an accurate and efficient early lung cancer diagnosis. The performance of the proposed ML-CNN was analyzed using Python, where the accuracy (2.5–10.5%) was high when compared to the number of instances, precision (2.3–9.5%) was high when compared to the number of instances, sensitivity (2.4–12.5%) was high when compared to several instances, the F-score (2–30%) was high when compared to the number of cases, the error rate (0.7–11.5%) was low compared to the number of cases, and the computation time (170 ms to 400 ms) was low compared to how many cases were computed for the proposed work, including previous known methods. The proposed ML-CNN architecture shows that this technique outperforms previous works.

1. Introduction

The development of Internet of Things (IoT)-based medical services and sensor systems has created numerous new study topics. IoT, cloud services, and intelligent systems (SIs) have all been covered by the current literature, changing traditional health services into intelligent healthcare. Health practitioners can enhance their practice by using essential technologies such as artificial intelligence (AI) and SIs [1]. The confluence of IoT and AI in the healthcare industry provides a wide variety of options. In light of this, the present study proposes a new AI- and IoT-based illness diagnostic paradigm for intelligent health systems. Due to the enormous cost of health insurance and the prevalence of multiple diseases, the medical center must be converted to a user system as soon as possible [2]. Researchers discovered that the survival rate is much higher when cancer is detected early. However, the early detection of such conditions is a difficult task. According to the correct classification, around half of cancer patients are in the middle or late stages of the disease [3]. In general, medical-image-processing-based disease diagnosis follows the simple procedure described below:
  • Image Preprocessing: In this step, image quality is improved by removing undesirable distortions in the image and correcting the geometric transformation of the image (rotation, scaling, and translation). Some of the essential preprocessing tasks are as follows: denoising, contrast enhancement, normalization, and so on [4].
  • Feature Extraction: This step converts the given input data into a set of features in a process known as feature extraction. In medical image processing, feature extraction starts with the initial set of reliable data, and feature extraction is performed on borrowed values known as features [5]. This makes the classification method much more satisfactory for prediction. Most medical image classifiers are based on the result of feature extraction in CAD.
  • Feature Selection: This step reduces the feature space extracted from the whole image, enhancing the detection rate and reducing the execution and response time. The reduced feature space is obtained by neglecting the irrelevant, noisy, and redundant features and choosing the feature subset that can achieve good performance with the subject of all metrics. In this stage, the dimensionality is reduced [6].
  • Classification: In this stage, the selected features are applied to classify each sample to denote the desired class. However, classification is a big issue in medical image analysis tasks. For automatic classification, optimal sets of features are used, and this step needs to be further improved for the better clinical analysis of the human body for diseases [7]. Computer-aided diagnostic systems have proven efficient in finding and diagnosing problems more rapidly and accurately. Unfortunately, this diagnostic technique is prone to several flaws and ambiguities, which might lead to an inaccurate prognosis. Hence, IoT sensor information is fused into the classification model. Machine learning (ML) models are currently used to indicate lung disease in computerized tomography (CT) images. The pipeline architecture is given in Figure 1. The following are the paper’s significant contributions [8]:
  • Improved CNN Model: The CNN’s efficiency is optimized to boost the lung cancer diagnostic performance, and its computation is based on various inputs. The CNN calculates the improved range of values based on the extracted features [9].
  • Hyperparameter Tuning: The learning parameters for the CNN must be generated on demand. It is also regarded as an optimization problem in which hyperparameters are tuned using the particle swarm optimization (PSO) algorithm [10,11,12].
  • Significant Performance Gains of Diagnosis Complexity: The optimized CNN architecture employs fewer convolutional layers by considering relevant information from the convolutional, fully connected, and max-pool layers. The proposed optimized CNN can reach minimal complexity as a result of the primary structural view of the CNN [13,14].
  • Consideration of IoT Sensor and Medical Images: A combination of sensor information and medical images is used to diagnose lung cancer from pictures [15,16].
Figure 1 shows the pipeline for IoMT-based lung disease diagnosis together with the design of the research model. The scientific effort has been given a defined road map.

Motivation for Lung Cancer Research

People affected by the disease daily come from many regions of the country and lose their identity, pleasures, and gifted quality of life. These facts inspired us to identify and analyze the disease’s underlying causes and address the need for the early disease detection of lung cancer. This motivated us to research lung cancer as soon as possible. Other factors are summarized below:
(i) Earlier diagnosis can improve the patient’s survival and quality of life.
(ii) For people who have been diagnosed with lung cancer, research can provide a better and longer future.
(iii) It can also, in the end, improve the number of survivors.
In advanced medical images, the internal organs and tissues of the body are seen using one or more specialized diagnostic techniques, referred to as “advanced imaging.” Without any human oversight, the CNN automatically finds essential characteristics. A multi-layer perceptron that can automatically learn the distribution law of data from a massive dataset has been created using the convolutional neural network (CNN).
The convolutional neural network (CNN) uses a significant quantity of data to identify the complicated representation of visual input. Because humans can perceive colors, it is essentially inspired by the human visual system. Convolutional networks recognize images as volumes, which represent three-dimensional objects. Red, Green, and Blue (RGB) encoding and the mixing of these three colors create the color spectrum in digital pictures. Each layer produces a dot product of the input pixels before moving on to the following layer. Along with the fundamental CNN structure, the CNN contains the following levels of activities.
In particle swarm optimization (PSO), the potential solutions are particles. At the optimal level, particles traverse the problem space and are linked with fitness, which is the best option. Each particle also maintains a tracking of all of its coordinates. Pbest is the new name for this fitness value. A particle’s value is the best overall value and is referred to as the best if it accepts the whole population as its neighbor.
The elements of the study project are (2) primary knowledge, which aids the reader in understanding the fundamental idea; (3) related works, which piece together the main principles; (4) system model, intended for the essential features; (5) experimental results and discussion, where the results are extensively discussed; (6) the conclusion, which discusses the results, viewpoints, and references for a simple comprehension of the work.
The rest of the paper is structured as follows: Section 2 discusses the preliminary knowledge about lung cancer diagnostic methods. Section 3 describes the related work, which elaborates on the current state-of-the-art methods for the classification of lung cancer. Section 4 discusses the proposed system model for lung cancer diagnosis. Section 5 discusses the experimental results for the proposed as well as existing methods. Section 6 concludes the paper and also explains future work.

2. Primary Knowledge

Current classification methods can be divided into two categories: supervised or unsupervised classification methods. In SVM, the Gaussian kernel is used; when there is no prior knowledge of the data, it is utilized to execute the transformation. Random Forest is used throughout the cited studies. Because of its efficiency and short processing time, SVM is the best technique for finding the best model and has received growing interest. An SVM algorithm for lung cancer disease does not support small-scale datasets and also requires a long processing time [17].
The automatic segmentation of components present inside an image, such as the region of interest in diagnostic images, has improved because of new methodologies combining transfer learning and techniques such as fine-tuning [18]. Deep learning and artificially intelligent techniques are applied to this data collection to develop statistics. Professionals then analyze the data to help customers lead healthier lives and avoid illness. Patients’ data are collected via health records (HRs), IoT sensor systems, personal and mobile phones, internet data, and social networks, which are emphasized among recent innovations. The PH uses artificial intelligence (AI) approaches to enhance illness development, illness prognosis, self-regulation and the self, and clinical intervention procedures using the set of data generated [19]. Recent studies are progressively looking into IoT devices in the healthcare profession. Healthcare diagnosis is routinely aided. Lung cancer patients will have a better chance of surviving if they can be diagnosed early. Deep learning and machine learning are frequently employed in the screening and effective identification of lung disease. This study’s major goal is to examine the function of deep learning in identifying and diagnosing lung disease. As a result, we employed a convolutional neural network (CNN), a type of deep learning model, to diagnose lung disease using radiography imaging [20]. A CT scan of the lungs can be used to identify and arrange the lung spikes and determine how dangerous they are. An improved CNN has high accuracy and temporal complexity characteristics; however, this is not as responsive to the setup as prior frameworks. They have tumors.

3. Related Work

A survey of roughly 36 articles that employed computational methods [21] to forecast various diseases was carried out in this study. The research consists of a variety of neural network models that are used to identify a range of illnesses and identify gaps that can be filled in the future for identifying lung disease in healthcare IoT. Every method was examined stepwise, and the general disadvantages were identified. Furthermore, the review examined the type of information utilized to predict the disease in question, that is, whether they were benchmark or manually obtained data. Lastly, strategies are defined and shown based on several accessible approaches. This will aid research in the future in correctly identifying malignant patients in their initial stages with no defects.
In [22], the authors propose an efficient SVM for lung classification, wherein the SVM’s characteristics are adjusted, and features are extracted using a refined gray wolf optimization algorithm in combination with a genetic algorithm (GWO-GA). The experimental phase is classified into three categories: parameterization testing, feature engineering, and optimal SVM. A benchmarking image dataset of 50 low-dose and archived lung CT scan images was used to monitor the effectiveness of the provided technique. The given work demonstrated the performance on the entire sample image dataset in various ways. Furthermore, it has an accuracy rate of 93.54%, which is much better than the evaluated approaches.
The current study presents a novel prototype for the feature extraction and classification of computed pulmonary tomography based on the Internet of Things. Combining the backpropagation algorithm with Parzen’s probability density, the presented approach employs a Software Development Kit (API) known as the Internet of Medical Things to identify lung images. The method worked well, with more than 98% correct classifications of lung images. The model then moves on to the lung segmentation method, which creates a lung map and uses fine-tuning to detect the pulmonary boundaries on the CT image using the Masks R-CNN network. The hypothesis was proven correct, as the suggested approach outperformed previous efforts in the research, achieving categorization parameters such as 98.34% [23].
Again, for evaluation, this work utilized computed tomography (CT) images of the lung and a probabilistic neural network (PNN) for diagnosis. After the raw lung images are preprocessed, features are extracted using the Gray-Level Co-Occurrence Matrix (GLCM). The chaotic crow search algorithm (CCSA)-based selection of features is presented. The calculation metrics employed were precision, sensitivity, predictive value, and accuracy. According to the investigation, on pictures with unevenly distributed cells, the feature selection based on CCSA effectively reaches 90%.When pulmonary development and degeneration are out of proportion, cancer cases of different kinds arise as a result; cancer cells then grow and expand, eventually creating a tumor. Whenever atmospheric poisons contact lung tissue, they cause harm. The unclear features of the lung cancer nodule make computer-aided automated diagnosis difficult at the moment. A selection approach to extract key characteristics was utilized together with the average accuracy to improve the classification performance. According to the results, the PNN with CCSA-based selection features outperforms the PNN without such features [24].
Researchers proposed a novel Internet of Things (IoT)-based predictive model based on fuzzy-cluster-based enhancement and classification to forecast lung cancer development via ongoing analysis and to enhance health by delivering medical instructions. For successful picture segmentation, the fuzzy clustering approach is utilized, which is based on the energy that can be transferred through extraction. In addition, the Fuzzy C-Means Clustering technique is employed to classify the transition area characteristics from the lung cancer image features. This method uses the Otsu thresholding approach to recover the transition zone from a lung image. In addition, the right-hand-side images and the morphology reduction procedure are applied to improve segmentation accuracy [25].
The morphology cleansing and image region filling procedures were conducted across an edge’s lung image to obtain the image areas. The authors also present a novel continuous classification approach that integrates existing Association Rules (ARM), the conventional Decision Tree (DT) with feature representation, and a CNN. LungNet, a unique combination DNN-network-based framework built using CT scans and healthcare IoT data obtained using a wearable camera, is presented to address the issue. LungNet is made up of a 22-layer CNN model that blends latent factors acquired from CT scan pictures with MIoT data to improve the system’s detection ability. This system, which is controlled from a centralized server and was trained using a balanced dataset of 525,000 images, can classify lung disease into five classifications with high precision (96.81%) and a lower false-positive rate (3.35%) than comparable CNN-based classifications. Furthermore, it accurately differentiates phase 1 and phase 2 lung tumors into 1A, 1B, 2A, and 2B sub-classes with a probability of detection of 10.5% [26].
All of the devices in this system are responsive to VOCs linked to lung cancer. The device’s data had to be normalized before they could be used. Next, to improve classification success, characteristics were collected from the data. Principal Component Analysis was performed to eliminate superfluous characteristics. Finally, k-Nearest Neighborhood and SVM algorithms were used to classify the collected characteristics [27].
This research used computed tomography (CT) images to develop automatized learning with quantitative research utilizing lung identification and therapy. On CT samples acquired from a lung dataset, the proposed model applies a Gaussian filtering (GF)-based preprocessing method. Moreover, they are put through the normalized cuts (Nuts) method, which may spot nodules that existed before the image. In addition, the oriented FAST and rotated BRIEF (ORB) technique is used as a feature extractor. Finally, the sunflower optimization-based wavelet neural network (SFO-WNN) model is used for lung cancer diagnosis. A descriptive analysis was conducted to evaluate the diagnosis results of the MLDS-LCDC model, and the findings were tested based on many factors [28].
The use of automated and optimal computer-aided diagnosis for lung disease is proposed in this paper. The procedure begins with the normalization and noise removal of the input images in a preliminary step. After that, Kapur entropy maximization and classification techniques are used to divide pulmonary areas. In the remainder of this study, 19 GLCM features were collected from the segmented images. The images with the highest priority were then chosen to reduce the program’s complexities. The feature selection method is considered a new optimization approach known as Enhanced Thermal Exchange Optimization (ITEO), which aims to increase reliability and stability. ITEO uses an optimized artificial neural network to classify images into healthy or malignant situations [29].
Magnetic resonance imaging (MRI), computed tomography scans (CT scans), ultrasonic devices, and positron emission tomography (PET) have all been used widely in the past; nevertheless, these methods have significant disadvantages, including cost, as well as the existence of radioactive material. Because of its convenience, accessibility, low price, and convenience of use, sensor synchronization in processing has recently been adopted in advanced screening. In terms of medical use, biotechnologies are becoming common tools for cancer scares, and intelligent sensor advancement has been realized. This article describes the results of studies on the use of biosensing in the treatment of lung cancer. According to the study, the electrochemical biosensor is the most widely utilized for early-stage lung cancer diagnosis [30].
In this study, the researchers present a new deep learning approach for lung cancer detection, which combines a convolutional neural network (CNN) with kernel k-means clustering. The proposed methodology was tested using the Anti-PD-1 Immunotherapy Lung datasets from The Medical Image Archives (https://www.kaggle.com/datasets/nih-chest-xrays/data and https://comport.com/healthcare-it-solutions/medical-image-archives/, accessed on 19 October 2022). The researchers utilized 400 MRI images from these data, which were carefully classified and included 150 routine lung scans and 250 cancer images. The framework was used to analyze the information throughout the first step. The kernel k-means clustering approach flattens neurons in the feature space for each image resulting from the convolution operation in the CNN. The center of each cluster is again determined using this procedure, which provides the prediction classification of each piece of data in the cross-validation. Various data points were used in k-fold cross-validation to assess the effectiveness of the suggested approach. When employing kernel functions with sigma = 0.05 in nine-fold cross-validation, the presented approach achieved the best set of metrics, with 98.85%, 98.32% sensitivity, 99.40% precision, 99.39% specificity, and a 98.86% F1-measure [31].
Employing expanded SegNet and convolutional neural network models, an intense CAD system for cancer detection with breast tomography images was constructed in this study (CNNs). To segment the lung from chest CT images, a dilated SegNet model was used, and a CNN model with batch normalization was built to identify the real nodules among all probable nodules. Using example cases from the LUNA16 dataset (https://luna16.grand-challenge.org/, accessed on 19 October 2022), the expanded SegNet and CNN classifiers were trained. The nodule classification was assessed using sensitivity, and the segmented model’s efficiency was quantified by the Dice coefficient. The statistical method confirmed the detection accuracy of the features obtained by CNN classification [32].
Another study provided a combined lung lesion detection and recognition network that can simultaneously identify, segment, and classify pulmonary nodules while accounting for inherent labeling uncertainties in the training dataset. It allows the final identification of nodules and the segmentation of the discovered nodules. For the improved analysis of 3D data, both the nodule identification and categorization subnets of the proposed combined system use a 3D encoder–decoder design. Furthermore, the categorization subnet boosts the categorization model by implementing characteristics extracted from the detection subnetwork with multiresolution nodule-specific data. When compared to the current backpropagation method, the proposed method provides useful prior knowledge to better optimize the much more difficult 3D classifier correlational design and better identify worrisome clusters from several other tissues [33].
Integrating two feedforward attentiveness processes, one at the nodule level and the other at the item level, a hierarchy attention-based multiple instance learning (HA-MIL) paradigm for clients diagnosing the disease was developed. The suggested HA-MIL platform was built by combining important parts’ representations with nodule information and then combining the nodule representation with lung recognition. The HA-MIL framework outperformed existing techniques, such as higher transaction learning in the case of MIL and embedding-space MIL, on the general populace Lung Image Database Consortium and Set Of images Resource Initiative (LIDC-IDRI) dataset (https://paperswithcode.com/dataset/lidc-idri, accessed on 19 October 2022), demonstrating the efficacy of hierarchy ensemble learning. The HA-MIL model also detected the important nodules and qualities by using greater attentiveness weights, according to the analysis of the findings [34].
Traditional Chinese Medicine (TCM) has been proven to be an excellent therapy for treating lung disease, and precise syndrome classification is critical to treatment. The creation of smart TCM syndrome discrimination algorithms is enabled by concrete evidence of TCM treatment instances and the advancement of intelligent systems. This is intended to broaden the scope of TCM’s advantages for lung cancers. This study aimed to develop end-to-end TCM diagnostic systems that might mimic lung disease classification. The developed models employed unorganized patient files as inputs to take advantage of data gathered by lung professionals for actual TCM therapy instances. The resulting algorithms outperformed those that used unstructured TCM data [35,36].Even though the articles in the related work have their significance, they can still increase their accuracy and performance. Through the understanding gained by the analysis of the above articles, this research has identified that ML-CNN with PSO and hyperparameters provides higher accuracy and greater performance and precision:
  • SVM with wolf optimization and the genetic algorithm has 93.54% accuracy;
  • R-CNN with the backpropagation algorithm and Parzen’s probability density has 98.34% accuracy;
  • The Gray-Level Co-Occurrence Matrix and chaotic crow search algorithm with a probabilistic neural network provide 90% accuracy;
  • The CNN with Association Rules and conventional Decision Tree (DT) provides 96.81% accuracy;
  • Gaussian filtering with the sunflower optimization-based wavelet neural network has less than 90% accuracy;
  • Enhanced Thermal Exchange Optimization and Kapur entropy maximization and classification techniques have less than 90% accuracy;
  • The convolutional neural network, in nine-fold cross-validation, provides 98.32% accuracy.
However, our approach, ML-CNN with PSO and hyperparameters, provides 98.85% accuracy, which is the highest among all compared methodologies.

4. System Model

In this section, the proposed system model is given in detail. Through IoMT technologies, sensors, and devices, the classification of lung cancer is implemented. Figure 2 illustrates the proposed model in detail.

4.1. Preprocessing

Firstly, lung images are denoised to improve the quality of the images. In this work, the Bayesian threshold-based Taylor series approach is proposed for denoising noisy pixels into better-quality pixels. This proposed approach is based on the Taylor series since it provides high-resolution wavelet subbands. The Taylor series defines an infinite sum of terms that are estimated from derivative functions in which the values are pixels and are varied. As a result of this, the Taylor series for a 3D frame is expressed as
T   ( X ) = ( a ) + ( X a ) T   D f ( a ) + 1 2 !   ( X a ) T     { D 2   f ( a ) } ( X a ) +
Figure 2 clearly illustrates the system architecture and functions for core comprehension. Figure 2 explains the ML-CNN and feature extraction models together with all preprocessing concepts.
In general, an image comprises different pixel values since the intensities of an image are not similar in all places of an image. Therefore, the changes in pixel values need more attention for denoising. From the defined Taylor series, the noise in subbands is computed and eliminated if it is greater than the estimated Bayesian threshold. The threshold for all subbands is formulated using the following expression:
T B = σ N 2 σ s
where σN represents the noise estimation, and TB is the threshold obtained for all bands.
When σ s 0 , it is formulated as
σ s = m a x ( ( σ y 2 σ y 2 ) , 0 )
where σy is defined as
σ y = 1 N   ( S B i )
The subbands SBi are S B i = { { L L L ,   L L H ,   L H L ,   L H H ,   H L L ,   H L H ,   H H L ,   H H H } , which is considered the total number of subbands, and σN is determined by
σ N = [ m e d i a n   ( S B ) 0.6745 ]
Upon estimating the TB value, the obtained values of the Bayesian threshold are arranged, and then using the curve fitting method, an expression is written as follows:
ϒ = p 1     β 2 + p 2     β + p 3 β + q
In Equation (6), β defines the standard noise deviation, and the values of other terms are given as p1 = 0.9592, p2 = 3.648, p3 = −0.138, and q = 0.1245. Next, a BM3D filtering algorithm is applied, and its mathematical formulation is expressed as follows:
Y ^ S S B = T 3 D 1   ( ( T 3 D ( Z S B ) ,   T B ϒ 2   l o g ( N 2 )   ) )
In Equation (7), T3D is the unitary 3D transform, Y ^ S S B is denoted as stacked subbands, Z is the subband size, and ∝ defines the threshold operator. From this final formulation, each subband is reconstructed and attains denoised frames for feature extraction.
Equation (12) D R E   ( F i , F i + 1 ) denotes the distance measure of Relative Entropy, and D R E   ( F i , F i + 1 ) gives the distance measure and Square Root of Relative Entropy between the frames. The term P i = { P i   ( 1 ) ,   P i   ( 2 )   P i ( n ) } represents the probability distribution function of frames present in a video, i.e., given as Fi and Fi+1, obtained from a normalized intensity histogram of n bins. The value of the bin is n = 256, and k represents the total number of frames present in each shot. Upon estimating the distance measures based on Relative Entropy and the Square Root of Relative Entropy, a comparison is made between them based on the time direction. From this comparison, the alterations in frames are identified, and redundantly occurring frames are eliminated. In this step, image intensity values are normalized into equal ranges for correct processing at the next level. The next step is executed as an intensity normalization process, in which the range is between 0 and 1. The Z-score normalization function is used for intensity normalization, the performance of which is better than the decimal and min–max normalization techniques. This normalization function is computed as follows:
Y i = x i μ δ
where Yi represents the normalized intensity values for the input image, μ is the mean intensity value, δ represents the variance, and xi represents the intensity position. The Z-score normalization function was used for intensity normalization, which was carried out in the processing of the image for upcoming processes, in the rational range (0–1). This step is mandatory for all images for both training and testing operations. This minimizes the overhead or complexity for similar medical image retrieval processes with the same properties (size, intensity values, and so on).

4.2. Feature Extraction

Table 1 shows the list of extracted features using the above-mentioned methods. The following features were re-extracted from the lung images:
  • Intensity-based features: These types of features describe the information on the intensity and histogram values located in the image (visibility is considered in the grayscale or color intensity histogram).
  • Shape-based features: These give image information about the shape and size.
  • Texture-based features: These describe the image intensity variations on the surface and estimate the properties of coarseness, smoothness, and regularity.
After feature extraction, the optimal set of features is extracted since the classification method cannot process a huge set of features. For this reason, the best sets of features are selected from the extracted feature set by spatial transformer networks.
The CNN is one of the types of neural network schemes with a major problem, i.e., high training complexity. Training depends upon loss values; training is frequently required. To address this issue, in this paper, the PSO algorithm is proposed, which finds the optimal values of the hyperparameters according to the input types and ranges. This type of integration helps reduce the hardware cost for CNN training and further produces fewer epochs for training and testing with the use of epochs. Further, this combination reduces the local minimum issue with regular backpropagation methodology training. The structure of the CNN is represented as follows:
  • Convolution Layers: This is the input layer that acquires the inputs, i.e., signals. This layer normalizes the range of all inputs into a single value for fast processing.
  • Pooling Layers: This layer’s purpose is to minimize the input size, which automatically influences the higher performance.
  • Activation Layer: This layer uses the ReLU activation function, which maximizes the outputs for non-linear cases.
  • Fully Connected Layer: This layer uses an activation function called softmax to control the output range. The proposed CNN computes the weight values for a given problem between neighboring neurons.
In Figure 3, deep neural networks called convolutional neural networks (CNNs) are shown classifying and segmenting pictures. Both supervised and unsupervised machine learning techniques can be used to train the CNN. With no loss of information, its built-in convolutional layer reduces the high dimensionality of pictures. In Table 2, low-level extracted features are listed, which are used for fundamental characteristics automatically retrieved from a picture without needing shape information or focusing on proximity. Table 3 shows lung nodules that are very typical and typically not a reason for alarm. Even so, finding out that one has a lesion on the lung might be scary.
Figure 4 describes the flowchart for the proposed optimized CNN with the PSO algorithm. The fitness function is computed as follows:
f i t n e s s f u n c t i o n = 1 1 N i = 1 N ( Y i Y ^ i ) 2
Based on the fitness function, the PSO algorithm computes the position of the neighboring particle that is influenced by the current position, and hence, each particle is updated using
V i ( t + 1 ) = ( c 1 × R A N D   ( ) × ( p i b e s t p i ( t ) ) ) + ( c 2 × R A N D   ( ) × ( p g b e s t p i ( t ) ) ) + V i ( t )
p i ( t + 1 ) = p i ( t ) + v i ( t )
If particle p i ( t ) lies close to the local best p i b e s t , the second term in the velocity-updating rule approaches zero, and for particles p i ( t ) lying close to their global best p g b e s t , the third term approaches zero. So, the particles close to their local best p i ( t ) or global best p g b e s t can evolve more easily than those far away from their best ones. The velocity-updating rule is as follows.
For the overall CNN layers, the performance of parameters used in training and testing is tuned using the PSO algorithm. With the use of global and local search issues, the performance is optimized. The hybrid CNN with the PSO is described below.
V i ( t + 1 ) represents the ith particle’s velocity, which is updated for every iteration. Other variables are the particle’s best position and coefficient variables. In the softmax layer of the proposed ML-CNN model, entropy functions are used for disease diagnosis and the threshold update to the given set of inputs. For the classification into the appropriate classes, the features are extracted, the similarity between probability distributions is calculated using Relative Entropy, and an additional entity is measured for determining the threshold values for the classifications, which is estimated using the Square Root of Relative Entropy. In this work, the features are extracted in the left direction toward the right for the accurate classification of the class. For this approach, the values of Relative Entropy and the Square Root of Relative Entropy are determined from the following formulations. In Algorithm 1 machine learning CNN classification pseudocode is presented for clear understanding purpose.
Algorithm 1. ML CNN classification pseudocode
Input: Sensed and image patterns Ffn
Let C    be the size of the total sequence to be classified
1. Ni = 3// no. of layers
2. Nr = 1  // no. of runs
3. Let Ns // no. of samples
4. Let Nf // no. of features
Initiallize traning set size (t фc)
Testing set size (t ψc)
// 80% training and 20% testing
Extract features from the sequence and create a list of features
Let ffn be the feature set extracted
Initialize CNN parameters
5. Let batch size = 1
6. No. of epochs = 1
Let Lbs be the labels corresponding to the selected features
Let Nc be the number of classes to be identified
7. Load ffn  // load optimized data
8. For i=1 to Nl
   Split Ffs (feature set into T (feature subset))
9.   For j = 1 to T
To find backpropagation CNN
10.     For i = 1 to size (t фc)
11.       N1. Ffn = t фc (i,:)
12.     End
13.     For i = 2:Nl
14.      Layer = i
15.      if Nl(i) = t фc (i,:)
16.      Val = Nl(i) ∗ t фc (i,:)
17.         For j = 1:length(Val)
18.            z = 0;
19.            For k = e:length(Val)
20.              Kk = Kk + 1
21.              Val = Nl(i-1) ∗ t ψc (k)
22.              Val1 = Nl(i) ∗ t ψc (k(;,:,1)
23.             End
24.            End
25.          End
    Let ffn be feature in Ffs(i,j)
    Trained TCLF estimates ffn
26. Let Rsort= sort(Tout)    //rank level
27. Let accuracy = mean(Tclf,1)
28. Tclf = ∑i(Tout/Rsort)
29.      cnti=∑ (Ffn) in belonging to samples Ns
30.     End
31. Compute Total count as CntT = i n     Cnti
   Compute probabilistic components for each class as
32.       For i = 1 to Nc
33.       Pcomp(i) = cnti/cntT
34.      End
35.End (Ending of firt i for loop)
Output: Classified output
( F i , F i + 1 ) = k = 1 n P i   ( k ) l o g P i   ( k ) P i + 1   ( k )  
D S R R E   ( F i , F i + 1 ) = k = 1 n P i   ( k ) l o g P i   ( k ) P i + 1   ( k )

5. Experimental Results and Discussion

5.1. Implementation

This study involved creating a fast and smart lung cancer detection system for use in the healthcare context. It classifies individuals with lung concerns using a deep-learning-based approach on a lung database. The information on the characteristics of the lung was determined using the ML-CNN approach. These data are used to guide the search area for the ML-CNN technique to determine the suitable class again for provided inputs using the PSO algorithm. The entire study was carried out using the Python programming language. The results are reported and organized into tables and graphs for comparison to other assessment parameters. The necessary description is provided in Table 4 and the CNN is learning using hyperparameters and PSO algorithms, which minimize iteration and time.
The above settings are applied to prevent overfitting. The hyperparameters are calculated using shallow tree and deep tree models. Once we do not have enough data, we can use the four-fold validation method; as such, it is not a heuristic method.
The above values are called hyperparameters and define the training required to build models. Changing the values of the hyperparameters changes the generated model and is generally used to determine how the algorithm learns relationships in the data.
A set of examples that are not part of the training process and do not participate in model tuning are used to evaluate the test set’s performance. This toolkit does not leak data into models, so it can be safely used to get an idea of how the model will perform in production.
During implementation, various hierarchical models of the PSO algorithm were adjusted. For such systematic methods, a variety of performance measures were employed, all of which were generated from the CNN model. The performance indicators that we utilized in our research are listed below.

5.2. Dataset Description

(i) LISS Dataset (A Public Database of Common Imaging Signs of Lung Diseases) (date of access: 19.10.2022)
We abbreviate this dataset as LISS CISL‖, a public database for researchers and academic purposes. It contains 511 2D CISLs for 252 patients, 166 3D images for 19 patients, and nearly 9 lung nodule signs captured at the Cancer Institute and Hospital at the Chinese Academy of Medical Sciences. These images were captured by GELightSpeed VCT-64 and Toshiba Aquilion 64 Slice CT scanners, Holt, MI, USA. The image format is DICOM 3.0. In the LISS database, 2D instances are covered by the nine categories of lung nodule sign images, and 3D instances are covered by only one sign (ground-glass opacity (GGO)).
In LISS, all CT images are assigned their corresponding label, and each image size is in pixels. The main reason why we considered this database is that it comprises a large number of lung images with different nodule signs. The lung image slice size in each class is 0.418–1 mm, and the mean value of the slice is 0.664 mm. The LISS database description is depicted in Table 5, which is categorized into two classes: CT imaging scans and annotated ROIs.
In the LISS database, the CT imaging scans are stored in plain text, and all of the annotated ROI CT images are stored as an individual file per nodule sign. The description of each nodule sign is given below:
  • GGO: This is a sign showing hazy maximized attenuation in the lung and vascular and bronchial margins. It is suggestive of lung adenocarcinoma and bronchioloalveolar carcinoma.
  • Lobulation: This is an indication that the septae of connective tissue, which contain fibroblasts recognised as coming from the perithymic mesenchyme, are growing (malignant lesion).
  • Calcification: This is a sign of the deposition of insoluble salts (magnesium and calcium). Its characteristics, such as distribution and morphology, are significant in distinguishing whether a lung nodule is benign or malignant. Popcorn and dense calcifications indicate that the given lung image is benign, and the central regions of lesions, spotted lesions, and irregular appearance indicate malignancy. It is determined from lung nodule areas with high-density pixels, and CT scan ranges are above 100 HU (Hounsfield Units).
  • Cavity and Vacuoles: These are signs of hollow spaces, represented by tissues. The vacuole is a type of cavity. It is suggestive of bronchioloalveolar carcinoma and adenocarcinoma, and the cavity is connected to tumors greater than 3 mm.
  • Speculation: This is a sign caused by stellate distortion in tissue lesions. The cancerous tumor’s intrusion causes this, which is one of the common lung nodule signs correlated with a desmoplastic response. This result indicates the fibrotic strand radiates with lung parenchyma. It is a sign of a malignant lesion.
  • Pleural Indentation: This is a sign of a tumor-affected area, which is diagnosed in the tissues. The sign is connected with peripheral adenocarcinomas that comprise central or subpleural anthracitic and fibrotic foci.
  • Bronchial Mucus Plugs (BMPs): These are also referred to as focal opacities. They vary in density and may have a liquefied density, which is above 100 HU. They are due to allergic bronchopulmonary aspergillosis and represent Intrabronchial Air, which the mucus transforms.
  • Air Bronchogram: This is a sign of the formation of low-attenuation bronchi over the background of the high-attenuation airless lung. It is due to (1) proximal airway patency, (b) the removal of alveolar air, which produces absorption (atelectasis) or replacement (pneumonia), or (3) the presence of both symptoms. In very rare cases, the air displacement is the result of the marked Interstitial Expansion (Lymphoma).
  • Obstructive Pneumonia: This rectangle mark available in Figure 5 is a sign of a small lung nodule volume, which is due to distal collapse. This is mainly due to proximal bronchial blocking. It is combined with adenocarcinoma and squamous cell carcinoma. For a visual representation, it represents the cuneate or flabellate area with increased density. Among the descriptions of lung nodule signs, radiologists have concluded that the GGO, Cavity and Vacuoles, Lobulation, Bronchial Mucus Plugs, and Pleural Indentation represent the malignant lung nodule lesions. Air Bronchogram, Obstructive Pneumonia, and Calcification are classified as benign or malignant lesions. Figure 5 presents the sites of nine lung nodule signs in images in the database.
(ii) LIDC-IDRI Dataset
The LIDC-IDRI dataset is the world’s biggest publicly available database for researchers and consists of thoracic CT images (1018 CT scan images) from 1010 patients (Figure 6). These images show the presence of lung cancers and nodule annotations, which means outlines. The nodule annotations are available for all patients. Still, the diagnostic information is presented for 157 patients, which consists of nodule information ratings, where 0 is an unknown class, 1 is the benign class, 2 is the primary malignant class, and 3 is the metastatic (malignant) class.
The ratings assigned for each image are based on a biopsy, surgical resection, progression, and the revision of radiological images. The diagnosis class is classified into two classes: (i) patient level and (ii) nodule level. To make this dataset for radiologists, images from 1000 patients were obtained over a long time with different CT scanners. We present the diameter size distribution for lung nodules in the LIDC-IDRI database in Table 6. Multiple lung nodules accompany the different slices. Each lung nodule slice comprises annotated data, and its presence is based on a 3 mm size.
In the LIDC database, diagnostic data are the means of ascertaining malignancy. We chose the ratings from diagnostic data; the ground truth was processed for training in the triplet CBMIR system, and it was evaluated by comparing the results with the radiologist-given ratings in the database. This database provides annotations for nodules with sizes from 3 mm to 30 mm shown in the yellow box.

5.3. Performance Measures

Equation (14) shows the accuracy rate, which is the number of accurate illness estimates relative to the total number of diagnoses. The capacity of a predictive result to reliably recognize persons without any signs of illness risks in the lung is provided in Equation (15). and represents specificity. The capacity of a decision outcome to properly identify persons with the illness is characterized as sensitivity, which is estimated in Figure 8.
The number of actual positives (i.e., risk controls and indicators for people who belong to the true-positive group) relative to the total number of people identified by the classification model (i.e., the total number classified as positive outcomes, including components incorrectly labeled as belonging to the positive class) equals the precision for a category.
The F-Score is the harmonized average mean of the specificity and sensitivity scores, as shown in Figure 11. The processing time delay is the cumulative time lag in learning the computational model and then evaluating it. It can be seen in Equation (15). In the following, the performance of the lung cancer diagnosis model is discussed and compared to existing methods. Table 7 illustrates the formulas for the performance metrics.
Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11 illustrate the performance of the model based on the accuracy, sensitivity, specificity, precision, and execution time concerning the number of instances. Based on the results shown in the figures, the performance of the proposed ML-CNN method produces the best results for lung cancer diagnosis. As a result of effective data collection, the preprocessing of images, extracted features, selected features, and classification by tuning input parameters, the performance of the proposed ML-CNN method has reached the maximum level. By addressing the limitations of poor-quality images, insufficient data, and features, the performance is improved. Furthermore, the filtering accuracy is computed for the proposed method, and the existing methods support the classification accuracy when the image-based lung cancer diagnosis is the focus.
After the selection of features (visual and semantic), they are subjected to a denoising process to remove noise, if any, present in them. To evaluate the efficiency of the method in this paper, filtering accuracy was analyzed. This filtering accuracy metric defines the range of filtering, i.e., whether the frames with noise are completely removed. The filtering accuracy is formulated in terms of the inverse of the mean square error, which is given as
Filtering   A c c u r a c y = ( 1 M S E )   100   %
where the mean square error is
MSE = 1 n   i = 1 n   ( Y i ^ Y i ) 2
In Equation (15),   ( Y i ^ Y i ) 2 is the square of errors, and 1/n ∑_(i = i)^n is the mean. According to the defined formulation, the filtering accuracy was determined and shows good results. The proposed Bayesian threshold method in this research work was compared with the previous fast bilateral filter involved in removing noises from images. The comparative results of the filtering accuracy are demonstrated in Figure 12 and Figure 13, from which the better efficiency of the proposed method is confirmed. This result is due to the involvement of a block-matching filter with a Bayesian threshold that removes errors from all parts of the image.
In contrast, a fast bilateral filter focuses on removing noise that is present at the diagonals only. This is the major reason for the reduced filtering accuracy of fast bilateral filters. The graphical comparison shows a gradual increase in filtering accuracy with the increase in time. An increase in this metric also causes an increase in classification accuracy due to the improvement in the clarity of images. Then, an optimization-based comparison was implemented for the proposed and previous approaches. In this work, hyperparameters were tuned using the PSO algorithm, which was compared to the traditional algorithms. Table 8 compares the other optimization methods to the PSO with their merits and demerits.
We analyzed the suggested approach using a variety of PSO-selected attributes to evaluate its overall performance parameters. F1 = 0.6, F2 = 0.5, and h = 1.0 have the maximum degree of precision. Table 9 represents the test results for various parameters. We ran several preliminary trials to find the optimal empirical combination of particle and iteration numbers. We discovered that 2500 particles and 29 iterations produced the final performance results given in Table 11 and Figure 7.
The PSO data, PSO iterations, and characteristics are described in Table 10, Table 11 and Table 12, which aids the model’s ability to forecast results accurately.

6. Conclusions

Virtual access to health imaging, including lung cancer computed tomography, is one of the benefits of IoT-based healthcare institutions. IoT nodes have made it easier to leverage data collected by various servers to investigate illness trends. As a result, a cancer diagnosis can be made by utilizing this information for CNN training. We developed a new classifier based on a thorough fully convolutional neural network in this study. ML-CNN is a broad sense classifier that may be used to detect and categorize biological imagery. However, this work uses ML-CNN to detect and categorize pulmonary nodules in CT scan images. The proposed ML-CNN classifier is trained on two categories: nodule (diseased—malignant or normal) and non-nodule (non-diseased—malignant or benign) (normal). Firstly, the sensor information is preprocessed using the normalization technique and then forwarded to the ML-CNN with PSO for feature extraction and classification. Secondly, the lung images are preprocessed using a Taylor series-based Bayesian threshold for noise reduction, and then Z-score normalization is used to normalize the intensity values. Then, the lung-image-based information is extracted from the image, such as color, shape, and intensity, and the performance is improved by tuning the hyperparameters using the PSO approach.
ML-CNN with PSO provides accuracy, precision, sensitivity, specificity, and F-measure values of 98.45, 98.89, 98.45, 98.62, and 98.85, respectively. The hybrid method gives high precision with greater convergence results when compared to other methods.
In deep learning, hyperparameters are parameters whose values are used to control the learning process. If we want to find some hidden aspects of the CNN, such as color, shape, and intensity, better performance is achieved by adjusting the hyperparameters. Model hyperparameters are manually set and used to help estimate model parameters. Model hyperparameters are often called parameters because they are a part of machine learning that must be manually fine-tuned.
PSO improves the accuracy, precision, sensitivity, specificity, and F-measure for both PSO and hyperparameters, which helps the CNN in feature extraction and classification.
We plan to employ more lung datasets in future research and apply hypotheses via fine-tuning and transfer learning to categorize different types of CT images, including melanoma, mammogram, and heart images. The model’s generalization for diverse types of medical imaging will also be evaluated in the continued study.

Author Contributions

Y.H.A.; software and validation, V.S.C.; formal analysis and validation, K.B.; formal analysis and investigation. R.R.M.; data curation, S.K.R.; methodology and original draft, A.T.S.; supervision, and A.K.F.; data curation and resources. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

On request we can produce the data to the concern person.

Conflicts of Interest

No authors raised any conflict of interest.

References

  1. Karthick, K.; Rajkumar, S.; Selvanathan, N.; Saravanan, U.K.; Murali, M.; Dhiyanesh, B. Analysis of Lung Cancer Detection Based on the Machine Learning Algorithm and IOT. In Proceedings of the 2021 6th International Conference on Communication and Electronics Systems (ICCES), Coimbatre, India, 8–10 July 2021; pp. 1–8. [Google Scholar]
  2. Villegas, D.; Martínez, A.; Quesada-López, C.; Jenkins, M. IoT for Cancer Treatment: A Mapping Study. In Proceedings of the 2020 15th Iberian Conference on Information Systems and Technologies (CISTI), Seville, Spain, 24–27 June 2020; pp. 1–6. [Google Scholar]
  3. Ng, T.L.; Morgan, R.L.; Patil, T.; Barón, A.E.; Smith, D.E.; Camidge, D.R. Detection of oligoprogressive disease in oncogene-addicted non-small cell lung cancer using PET/CT versus CT in patients receiving a tyrosine kinase inhibitor. Lung Cancer 2018, 126, 112–118. [Google Scholar] [CrossRef] [PubMed]
  4. Jony, M.H.; Johora, F.T.; Khatun, P.; Rana, H.K. Detection of Lung Cancer from CT Scan Images using GLCM and SVM. In Proceedings of the 2019 1st International Conference on Advances in Science, Engineering and Robotics Technology (ICASERT), Dhaka, Bangladesh, 3–5 May 2019; pp. 1–6. [Google Scholar] [CrossRef]
  5. Zanon, M.; Pacini, G.S.; de Souza, V.V.S.; Marchiori, E.; Meirelles, G.S.P.; Szarf, G.; Torres, F.S.; Hochhegger, B. Early detection of lung cancer using ultra-low-dose computed tomography in coronary CT angiography scans among patients with suspected coronary heart disease. Lung Cancer 2017, 114, 1–5. [Google Scholar] [CrossRef] [PubMed]
  6. Mishra, S.; Thakkar, H.K.; Mallick, P.K.; Tiwari, P.; Alamri, A. A sustainable IoHT based computationally intelligent healthcare monitoring system for lung cancer risk detection. Sustain. Cities Soc. 2021, 72, 103079. [Google Scholar] [CrossRef]
  7. Alakwaa, W.; Mohammad, N.; Amr, B. Lung cancer detection and classification with 3D convolutional neural network (3D-CNN). Int. J. Adv. Comput. Sci. Appl. 2017, 8, 8. [Google Scholar] [CrossRef] [Green Version]
  8. Hatuwal, B.K.; Himal, C.T. Lung cancer detection using convolutional neural network on histopathological images. Int. J. Comput. Trends Technol. 2020, 68, 10, 21–24. [Google Scholar] [CrossRef]
  9. Sasikumar, S.; Renjith, P.N.; Ramesh, K.; Sankaran, K.S. Attention Based Recurrent Neural Network for Lung Cancer Detection. In Proceedings of the 2020 Fourth International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud) (I-SMAC), Palladam, India, 7–9 October 2020; pp. 720–724. [Google Scholar]
  10. Al-Yasriy, H.F.; AL-Husieny, M.S.; Mohsen, F.Y.; Khalil, E.A.; Hassan, Z.S. Diagnosis of lung cancer based on CT scans using CNN. IOP Conf. Ser. Mater. Sci. Eng. 2020, 928, 022035. [Google Scholar] [CrossRef]
  11. Kirubakaran, J.; Venkatesan, G.K.D.P.; Kumar, K.S.; Kumaresan, M.; Annamalai, S. Echo state learned compositional pattern neural networks for the early diagnosis of cancer on the internet of medical things platform. J. Ambient. Intell. Humaniz. Comput. 2020, 12, 3303–3316. [Google Scholar] [CrossRef]
  12. Bahat, B.; Görgel, P. Lung Cancer Diagnosis via Gabor Filters and Convolutional Neural Networks. In Proceedings of the 2021 Innovations in Intelligent Systems and Applications Conference (ASYU), Elazig, Turkey, 6–8 October 2021; pp. 1–6. [Google Scholar]
  13. Hosseini, H.M.; Monsefi, R.; Shadroo, S. Deep Learning Applications for Lung Cancer Diagnosis: A systematic review. arXiv 2020, arXiv:2201.00227. [Google Scholar]
  14. Ozdemir, O.; Russell, R.L.; Berlin, A.A. A 3D Probabilistic Deep Learning System for Detection and Diagnosis of Lung Cancer Using Low-Dose CT Scans. IEEE Trans. Med. Imaging 2019, 39, 1419–1429. [Google Scholar] [CrossRef] [Green Version]
  15. Capuano, R.; Catini, A.; Paolesse, R.; Di Natale, C. Sensors for Lung Cancer Diagnosis. J. Clin. Med. 2019, 8, 235. [Google Scholar] [CrossRef] [Green Version]
  16. Shimizu, R.; Yanagawa, S.; Monde, Y.; Yamagishi, H.; Hamada, M.; Shimizu, T.; Kuroda, T. Deep learning appli-cation trial to lung cancer diagnosis for medical sensor systems. In Proceedings of the 2016 International SoC Design Conference (ISOCC), Jeju, Korea, 23–26 October 2016; pp. 191–192. [Google Scholar]
  17. Mishra, S.; Chaudhary, N.K.; Asthana, P.; Kumar, A. Deep 3d convolutional neural network for automated lung cancer diagnosis. In Computing and Network Sustainability: Proceedings of IRSCNS 2018; Springer: Singapore, 2019; pp. 157–165. [Google Scholar]
  18. Zhang, Y.; Simoff, M.J.; Ost, D.; Wagner, O.J.; Lavin, J.; Nauman, B.; Hsieh, M.-C.; Wu, X.-C.; Pettiford, B.; Shi, L. Understanding the patient journey to diagnosis of lung cancer. BMC Cancer 2021, 21, 402. [Google Scholar] [CrossRef] [PubMed]
  19. Alsammed, S.M.Z.A. Implementation of Lung Cancer Diagnosis based on DNN in Healthcare System. Webology 2021, 18, 798–812. [Google Scholar] [CrossRef]
  20. Pradhan, K.; Chawla, P. Medical Internet of things using machine learning algorithms for lung cancer detection. J. Manag. Anal. 2020, 7, 591–623. [Google Scholar] [CrossRef]
  21. Valluru, D.; Jeya, I.J.S. IoT with cloud based lung cancer diagnosis model using optimal support vector machine. Heal. Care Manag. Sci. 2019, 23, 670–679. [Google Scholar] [CrossRef] [PubMed]
  22. Souza, L.F.D.F.; Silva, I.C.L.; Marques, A.G.; Silva, F.H.D.S.; Nunes, V.X.; Hassan, M.M.; Filho, P.P.R. Internet of Medical Things: An Effective and Fully Automatic IoT Approach Using Deep Learning and Fine-Tuning to Lung CT Segmentation. Sensors 2020, 20, 6711. [Google Scholar] [CrossRef]
  23. Chakravarthy, S.; Rajaguru, H. Lung Cancer Detection using Probabilistic Neural Network with modified Crow-Search Algorithm. Asian Pac. J. Cancer Prev. 2019, 20, 2159–2166. [Google Scholar] [CrossRef] [Green Version]
  24. Palani, D.; Venkatalakshmi, K. An IoT Based Predictive Modelling for Predicting Lung Cancer Using Fuzzy Cluster Based Segmentation and Classification. J. Med. Syst. 2018, 43, 21. [Google Scholar] [CrossRef]
  25. Faruqui, N.; Abu Yousuf, M.; Whaiduzzaman; Azad, A.; Barrosean, A.; Moni, M.A. LungNet: A hybrid deep-CNN model for lung cancer diagnosis using CT and wearable sensor-based medical IoT data. Comput. Biol. Med. 2021, 139, 104961. [Google Scholar] [CrossRef]
  26. Ozsandikcioglu, U.; Atasoy, A.; Yapici, S. Hybrid Sensor Based E-Nose For Lung Cancer Diagnosis. In Proceedings of the 2018 IEEE International Symposium on Medical Measurements and Applications (MeMeA), Rome, Italy, 11–13 2018; pp. 1–5. [Google Scholar]
  27. Kiran, S.V.; Kaur, I.; Thangaraj, K.; Saveetha, V.; Grace, R.K.; Arulkumar, N. Machine Learning with Data Science-Enabled Lung Cancer Diagnosis and Classification Using Computed Tomography Images. Int. J. Image Graph. 2021, 1, 2240002. [Google Scholar] [CrossRef]
  28. Shan, R.; Rezaei, T. Lung Cancer Diagnosis Based on an ANN Optimized by Improved TEO Algorithm. Comput. Intell. Neurosci. 2021, 2021, 6078524. [Google Scholar] [CrossRef]
  29. Jeffree, A.I.; Karman, S.; Ibrahim, S.; Ab Karim, M.S.; Rozali, S. Biosensors Approach for Lung Cancer Diagnosis—A Review. In Lecture Notes in Mechanical Engineering, Proceedings of the 6th International Conference on Robot Intelligence Technology and Applications, Kuala Lumpur, Malaysia, 16–18 December 2018; Springer: Singapore, 2020; pp. 425–435. [Google Scholar]
  30. Rustam, Z.; Hartini, S.; Pratama, R.Y.; Yunus, R.E.; Hidayat, R. Analysis of Architecture Combining Convolutional Neural Network (CNN) and Kernel K-Means Clustering for Lung Cancer Diagnosis. Int. J. Adv. Sci. Eng. Inf. Technol. 2020, 10, 1200–1206. [Google Scholar] [CrossRef]
  31. Anitha, J.; Agnes, S.A. Appraisal of deep-learning techniques on computer-aided lung cancer diagnosis with computed tomography screening. J. Med. Phys. 2020, 45, 98–106. [Google Scholar] [CrossRef] [PubMed]
  32. Chenyang, L.; Chan, S.-C. A Joint Detection and Recognition Approach to Lung Cancer Diagnosis From CT Images With Label Uncertainty. IEEE Access 2020, 8, 228905–228921. [Google Scholar] [CrossRef]
  33. Wang, Q.; Zhou, Y.; Huang, J.; Liu, Z.; Li, L.; Xu, W.; Cheng, J.-Z. Hierarchical Attention-Based Multiple Instance Learning Network for Patient-Level Lung Cancer Diagnosis. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Korea, 16–19 December 2020; pp. 1156–1160. [Google Scholar]
  34. Liu, Z.; He, H.; Yan, S.; Wang, Y.; Yang, T.; Li, G. End-to-End Models to Imitate Traditional Chinese Medicine Syn-drome Differentiation in Lung Cancer Diagnosis: Model Development and Validation. JMIR Med. Inform. 2020, 8, e17821. [Google Scholar] [CrossRef] [PubMed]
  35. Dodia, S.; Annappa, B.; Mahesh, P.A. Recent advancements in deep learning based lung cancer detection: A sys-tematic review. Eng. Appl. Artif. Intell. 2022, 116, 105490. [Google Scholar] [CrossRef]
  36. Civit-Masot, J.; Bañuls-Beaterio, A.; Domínguez-Morales, M.; Rivas-Pérez, M.; Muñoz-Saavedra, L.; Corral, J.M.R. Non-small cell lung cancer diagnosis aid with histopathological images using Explainable Deep Learning techniques. Comput. Methods Programs Biomed. 2022, 226, 107108. [Google Scholar] [CrossRef]
Figure 1. Pipeline for IoMT-based lung disease diagnosis.
Figure 1. Pipeline for IoMT-based lung disease diagnosis.
Bioengineering 10 00320 g001
Figure 2. System architecture.
Figure 2. System architecture.
Bioengineering 10 00320 g002
Figure 3. Machine Learning—CNN for classification.
Figure 3. Machine Learning—CNN for classification.
Bioengineering 10 00320 g003
Figure 4. Flowchart for CNN with PSO.
Figure 4. Flowchart for CNN with PSO.
Bioengineering 10 00320 g004
Figure 5. Abnormal signs in LISS dataset.
Figure 5. Abnormal signs in LISS dataset.
Bioengineering 10 00320 g005
Figure 6. Lung classification results.
Figure 6. Lung classification results.
Bioengineering 10 00320 g006
Figure 7. Accuracy vs. number of instances.
Figure 7. Accuracy vs. number of instances.
Bioengineering 10 00320 g007
Figure 8. Sensitivity vs. number of instances.
Figure 8. Sensitivity vs. number of instances.
Bioengineering 10 00320 g008
Figure 9. Specificity vs. number of instances.
Figure 9. Specificity vs. number of instances.
Bioengineering 10 00320 g009
Figure 10. Precision vs. number of instances.
Figure 10. Precision vs. number of instances.
Bioengineering 10 00320 g010
Figure 11. F-score vs. number of instances.
Figure 11. F-score vs. number of instances.
Bioengineering 10 00320 g011
Figure 12. Execution time vs. number of instances.
Figure 12. Execution time vs. number of instances.
Bioengineering 10 00320 g012
Figure 13. Filtering accuracy vs. number of instances.
Figure 13. Filtering accuracy vs. number of instances.
Bioengineering 10 00320 g013
Table 1. List of extracted features.
Table 1. List of extracted features.
Feature Type Feature DescriptionFormula
Texture featuresCorrelationIt is defined by the similar direction of rows and columns i , j = 0 n 1 p i , j ( i j ) 2
ContrastIt indicates the brightness or contrast for each nodule. When the contrast level is high, texture features are not clear. i , j = 0 n 1 p i , j ( ( i μ i ) ( j μ j ) σ i σ j )
HomogeneityIt is defined as the distribution closeness of entities in the matrix. It is computed in all directions of the image. i , j p ( i , j ) 1 + | i j |
EntropyIt is defined by the complexity of textures or unevenness E = g = 1 G g = 1 G p ( g , g ) log p ( g , g )
p ( i , j ) is the normalized image matrix for one pixel coordinate, n is the total number of distinct gray levels in the image
Shape featuresAreaThe number of pixels in the largest axial slice computes it l a r a S multiplied by the resolution of the pixels I ( x , y ) × Δ A
Aspect RatioIt is computed by the major axis and minor axis length M A ( x , y ) [ L ] M I ( x , y ) [ L ]
RoundnessIt is computed by the similarity of the lung nodule region to the circular shape 4 π A L
PerimeterIt is defined as the structural property for the list of coordinates and also the sum of the distance from each coordinate ( X i X i 1 ) 2 + ( Y i Y i 1 ) 2
CircularityIt is computed by the 〖lar〗_aS of each lung nodule image 4 π A Q 2
Δ A is the area of one pixel in the shape of I ( x , y ) , X i and Y i is the ith pixel coordinates, A is the nodule area, and L is the nodule region boundary length, M A ( x , y ) [ L ] is the major axis length and M I ( x , y ) [ L ] is the minor axis length.
Intensity featuresUniformityIt is defined by the intensity uniformity of the Histograms i = 0 l 1 H 2 ( R i )
MeanThe average intensity measure represents it i = 0 l 1 R i × H ( R i )
Standard VarianceIt is represented by the 2nd moment of the average values i = 0 l 1 ( R i M ) 3 × H ( R i )
KurtosisIt is defined by the 4th moment of mean i = 0 l 1 ( R i M ) 4 × H ( R i )
SkewnessIt is defined by the 3rd moment of the mean values i = 0 l 1 ( R i M ) 3 × H ( R i )
SmoothnessThe relative intensity differences in the given region define it 1 1 ( 1 + σ 2 )
R i Is a random attribute that indicates the intensity, H(R_i) is the Histogram of the Intensity in the nodule region, l denotes the number of potential intensity levels, and σ represents the standard deviation
Table 2. List of extracted low-level features.
Table 2. List of extracted low-level features.
Feature Name
Texture featuresCorrelation
Contrast
Homogeneity
Sum of square variance
Spectral, Spatial, and Entropy
Shape featuresArea
Irregularity
Roundness
Perimeter
Circularity
Intensity featuresIntensity
Mean
Standard Variance
Kurtosis
Skewness
Median
Table 3. High-level (semantic) features for lung nodule classification.
Table 3. High-level (semantic) features for lung nodule classification.
Nodule Size<4 mm, 4–7 mm, 8–20 mm, and >20 mm
SpeculationThe range between 0 and 1
DiameterNodule diameter
LocationUpper lobe or not
MorphologySmooth, Lobulated, Spiculated, and Irregular
Table 4. Hyperparameters of CNN and PSO.
Table 4. Hyperparameters of CNN and PSO.
Methods Used ParametersDescription
CNNNumber of Epochs5
Non-Linear Activation FunctionReLU
Activation Function Softmax
Learning FunctionAdam Optimizer
PSONumber of Particles10
Iterations 10
Inertial Weight (W)0.85
Social Constant (W2)2
Cognitive Constant (W1)2
Table 5. Dataset description.
Table 5. Dataset description.
CT Imaging ScansAnnotated ROIs
GGO25–2D
19–3D
45–2D
166–3D
Lobulation2141
Calcification2047
Cavity and Vacuoles75147
Speculation1829
Pleural Dragging2680
Air Bronchogram2223
Bronchial Mucus Plugs1981
Obstructive
Pneumonia
1618
Sum271677
Table 6. LIDC-IDRI database description.
Table 6. LIDC-IDRI database description.
Nodule DescriptionNumber of Nodules
No. of nodules or lesions marked by at least 1 radiologist7371
No. of <3 mm nodules marked by at least 1 radiologist2669
No. of <3 mm nodules marked by all 4 radiologists928
Table 7. Parameters for lung cancer diagnosis.
Table 7. Parameters for lung cancer diagnosis.
Formula
Accuracy A C C   =   T P   +   T N T P   +   T N   +   F P   +   F N   ×   100 %  
Sensitivity S e n s i t i v i t y   =   T P T P   +   T N   ×   100 %  
Specificity S p e c i f i c i t y   =   T N T N   +   F P   ×   100 %  
Precision P r e c i s i o n   =   T P T P   +   F P   ×   100 %  
F-Score F = 2     P r e c i s i o n     S e n s i t i v i t y P r e c i s i o n   +   S e n s i t i v i t y   ×   100 %
where TP = true positive, TN = true negative, FP = false positive, FN = false negative
Table 8. Comparison of optimization approaches.
Table 8. Comparison of optimization approaches.
Approach DescriptionMerits and Demerits
Genetic algorithmA genetic algorithm is an evolutionary algorithm that deals with chromosomes. Operations such as crossover and mutation are performed. Merits:
1. Parallel processing
2. Flexible
Demerits:
1. Gives close to the optimal solution
2. Lot of computations are required
Ant Colony OptimizationThis algorithm is based on the behavior of ants, which deposit chemical components called pheromones.Merits:
1. Faster discovery with appropriate solutions
2. Solves travel salesman problem
Demerits:
1. Inappropriate for continuous problems
2. Weakened local search
Bee Colony OptimizationIndividual proactive abilities and self-organizing capacity are developed with honey bee behaviorMerits:
1. Faster convergence
2. Uses fewer control parameters
Demerits:
1. Allows only a limited number of search space
Particle swarm optimizationThe population-based algorithm initializes the input and evaluates the fitness value.Merits:–
1. Robust
2. Less computation time
Table 9. Comparison of numerical results for proposed and existing methods.
Table 9. Comparison of numerical results for proposed and existing methods.
AlgorithmAttacksAccuracy (%)Precision (%)Sensitivity (%)Specificity (%)F-Measure (%)
GA200084.5286.7585.4586.7585.7
400086.5487.4586.1287.4585.9
600087.4588.1586.7588.7885.6
800088.0288.1686.7888.9585.9
10,00088.1588.1986.7988.2386
ACO200088.4588.7888.9188.9689
400088.688.7588.9088.9789.15
600088.6988.8988.9288.9889.45
800088.7288.9088.9488.9889.55
10,00089.7588.9188.9588.9989.6
BCO200090.191.2391.2591.4491.63
400090.2591.4591.3591.5691.65
600090.2691.7591.3991.5891.67
800090.3291.8891.4091.6091.69
10,00090.4591.8991.4591.6291.85
PSO200098.198.2398.2598.4498.63
400098.2598.4598.3598.5698.65
600098.2698.7598.3998.5898.67
800098.3298.8898.4098.6098.69
10,00098.4598.8998.4598.6298.85
Table 10. Algorithm parameters for PSO using empirical data.
Table 10. Algorithm parameters for PSO using empirical data.
F2hAccuracy
0.80.61.098.45
0.80.60.997.73
0.80.61.098.12
0.70.61.098.09
0.60.51.099.46
Table 11. PSO method results when utilizing a constant number of particles and an increasing number of iterations.
Table 11. PSO method results when utilizing a constant number of particles and an increasing number of iterations.
IterationsAccuracyPrecisionF-Measure
25002597.9097.8997.12
25002698.0697.0397.56
25002798.4596.4396.49
25002898.2397.6398.62
25002999.5699.5499.32
25003097.9697.8797.51
Table 12. Observations of the PSO algorithm with different feature sizes.
Table 12. Observations of the PSO algorithm with different feature sizes.
FeaturesAccuracyPrecisionF-Measure
1099.4599.0399.89
1298.0997.4697.43
1598.8398.0398.69
1898.2398.6797.52
2097.1297.2398.86
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hussain Ali, Y.; Sabu Chooralil, V.; Balasubramanian, K.; Manyam, R.R.; Kidambi Raju, S.; T. Sadiq, A.; Farhan, A.K. Optimization System Based on Convolutional Neural Network and Internet of Medical Things for Early Diagnosis of Lung Cancer. Bioengineering 2023, 10, 320. https://doi.org/10.3390/bioengineering10030320

AMA Style

Hussain Ali Y, Sabu Chooralil V, Balasubramanian K, Manyam RR, Kidambi Raju S, T. Sadiq A, Farhan AK. Optimization System Based on Convolutional Neural Network and Internet of Medical Things for Early Diagnosis of Lung Cancer. Bioengineering. 2023; 10(3):320. https://doi.org/10.3390/bioengineering10030320

Chicago/Turabian Style

Hussain Ali, Yossra, Varghese Sabu Chooralil, Karthikeyan Balasubramanian, Rajasekhar Reddy Manyam, Sekar Kidambi Raju, Ahmed T. Sadiq, and Alaa K. Farhan. 2023. "Optimization System Based on Convolutional Neural Network and Internet of Medical Things for Early Diagnosis of Lung Cancer" Bioengineering 10, no. 3: 320. https://doi.org/10.3390/bioengineering10030320

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop