Next Article in Journal
Chronic or Changeable Infarct Size after Spontaneous Coronary Artery Dissection
Next Article in Special Issue
A Hybrid Technique for Diabetic Retinopathy Detection Based on Ensemble-Optimized CNN and Texture Features
Previous Article in Journal
Immature Teratoma: Diagnosis and Management—A Review of the Literature
Previous Article in Special Issue
Recognizing Brain Tumors Using Adaptive Noise Filtering and Statistical Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

OView-AI Supporter for Classifying Pneumonia, Pneumothorax, Tuberculosis, Lung Cancer Chest X-ray Images Using Multi-Stage Superpixels Classification

1
Department of Mechanical System Engineering, Chosun University, Gwangju 61452, Republic of Korea
2
OTOM, Co., Ltd., Gwangju 61042, Republic of Korea
3
Department of Radiology, Soonchunhyang University Cheonan Hospital, Cheonan 31151, Republic of Korea
4
Haewootech Co., Ltd., Busan 46742, Republic of Korea
5
Department of Computer Software Engineering, Soonchunhyang University, Asan 31538, Republic of Korea
6
Division of Respiratory Medicine, Department of Internal Medicine, Soonchunhyang University Cheonan Hospital, Cheonan 31151, Republic of Korea
7
Department of Nursing, Soonchunhyang University Cheonan Hospital, Cheonan 31151, Republic of Korea
8
Department of Otolaryngology-Head and Neck Surgery, Cheonan Hospital, Soonchunhyang University College of Medicine, Cheonan 31151, Republic of Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2023, 13(9), 1519; https://doi.org/10.3390/diagnostics13091519
Submission received: 30 March 2023 / Revised: 20 April 2023 / Accepted: 20 April 2023 / Published: 23 April 2023
(This article belongs to the Special Issue AI/ML-Based Medical Image Processing and Analysis)

Abstract

:
The deep learning approach has recently attracted much attention for its outstanding performance to assist in clinical diagnostic tasks, notably in computer-aided solutions. Computer-aided solutions are being developed using chest radiography to identify lung diseases. A chest X-ray image is one of the most often utilized diagnostic imaging modalities in computer-aided solutions since it produces non-invasive standard-of-care data. However, the accurate identification of a specific illness in chest X-ray images still poses a challenge due to their high inter-class similarities and low intra-class variant abnormalities, especially given the complex nature of radiographs and the complex anatomy of the chest. In this paper, we proposed a deep-learning-based solution to classify four lung diseases (pneumonia, pneumothorax, tuberculosis, and lung cancer) and healthy lungs using chest X-ray images. In order to achieve a high performance, the EfficientNet B7 model with the pre-trained weights of ImageNet trained by Noisy Student was used as a backbone model, followed by our proposed fine-tuned layers and hyperparameters. Our study achieved an average test accuracy of 97.42%, sensitivity of 95.93%, and specificity of 99.05%. Additionally, our findings were utilized as diagnostic supporting software in OView-AI system (computer-aided application). We conducted 910 clinical trials and achieved an AUC confidence interval (95% CI) of the diagnostic results in the OView-AI system of 97.01%, sensitivity of 95.68%, and specificity of 99.34%.

1. Introduction

There are several different types of diseases that prevent lungs from functioning properly [1]. The diseases that affect the alveoli region of the thoracic region are pneumonia, tuberculosis, and lung cancer [2]. Pneumonia is an inflammation that occurs in the subcellular region of the lungs, especially the alveoli (air sac) [3]. Tuberculosis bacteria are classified as infectious, acute, or chronic diseases with the lungs being infected. Pulmonary tuberculosis can be diagnosed as the presence of cheese-like granuloma mass in the lungs from chest X-ray images, and this symptom is called cavitation [4]. Lung cancer affects uncontrolled cell growth in the tissues of the lung [2]. In South Korea, lung cancer is the third highest among all cancers, as reported by the annual report of the Korean National Cancer Registration Statistics [5]. Air pneumothorax, or pneumothorax, refers to a disease in which air leaks through a hole in the lungs and air fills the pleural cavity. As the amount of air in birds increases, the lungs do not function normally [2]. Pneumothorax affects partial or complete lung collapse [6]. Those abovementioned lung diseases can be primarily and commonly assessed through chest radiography by professional radiologists. An early screening of the disease and regular monitoring of the changes in the disease can reduce the life-threatening consequences.
Chest X-ray images are non-invasive chest radiographs which are easy and economical for lung abnormality inspection. They assist radiologists in screening and diagnosing the overall findings of the chest in brief. However, the manual examination of a chest X-ray image is highly prone to error and is time-consuming for the person conducting the task. Therefore, computer-aided application helps to fasten and increase the accuracy of this examination.
In recent years, with the rapid innovation of artificial intelligence in computer vision, deep learning has been deployed in medical image analysis, including image classification, segmentation, detection, and localization tasks [2,7,8]. The computer-aided solution is one of the medical image analyses which use chest X-ray images to assist with the diagnosis of lung disease. Unlike traditional machine learning methods, deep learning can extract potential features from chest X-ray images and then classify them in end-to-end learning without user intervention. However, chest X-ray images pose a challenge for deep learning techniques due to their high inter-class similarities and low intra-class variant abnormalities, especially in terms of the complex nature of radiographs and the complex anatomy of the chest. For instance, pneumonia, tuberculosis, and lung cancer occur within the alveoli region of the thoracic region, for which it is hard to identify the different patterns of their disorders. Additionally, most deep learning research of the image classification task focuses on natural image modalities (RGB images). Since a chest X-ray image is a grayscale image, it results in more challenges for the deep learning method. Moreover, all chest X-ray images were scanned only at the chest area in which the output images have the same pattern of the chest anatomy.
Convolutional neural network (CNN) has proved its promising performance in 2D image classification tasks. Early state-of-the-art CNN models such as AlexNet, VGG, ResNet, DenseNet, MobileNet, InceptionNet, XceptionNet, and EfficientNet, to name a few, have proved their significant performances trained on the ImageNet dataset. The ImageNet dataset consists of thousands of classes and millions of images. “Training from scratch” on a large dataset requires a much higher computation cost and much more time. Therefore, a “transfer learning” technique has played an important role in closing the gap between the computation cost and the significant performance.
Therefore, in this study, we propose a fine-tuned deep-learning-based solution to classify multi-chest infections using chest X-ray images to improve the efficiency and effectiveness of diagnosis using the computer-aided application (namely, OView-AI system).
Our main contributions are as follows:
  • We exploited the EfficientNet B7 [9] with the pre-trained weights of ImageNet trained by Noisy Student [10] as a backbone model; this was then followed by proposed fine-tuning layers and hyperparameters;
  • Our method is end-to-end learning which extracts the potential features of a chest X-ray image directly and then applies the Softmax function to generate the different predicted class probabilities;
  • Our experiment was conducted on a dataset which collected data from Soonchunhyang University Hospital from May 2022 to July 2022 and were labeled by professional radiologists of Soonchunhyang University Hospital after de-identification;
  • We conducted an inference, of which the inferencing results were evaluated and confirmed by professional radiologists of Soonchunhyang University Hospital;
  • Our study was conducted on the trend of major lung diseases occurring in South Korea, such as pneumonia, pneumothorax, tuberculosis, and lung cancer.

2. Related Works

Table 1 lists recently published related papers conducted on the classification of lung diseases. The list is sorted by the largest number of classes. Asif et al. [11] conducted the transfer learning technique with the pre-trained weights of VGG-16 to classify two lung diseases (COVID-19, viral pneumonia, and normal). The method achieved an accuracy of 97.84%. Hamwi et al. [12] conducted the transfer learning technique with the pre-trained weights of a merged model of VGG-16 and DenseNet-201 to classify one lung disease (COVID-19 and normal). The method achieved an accuracy of 99.73%. Hong et al. [13] used the transfer learning technique with the pre-trained weights of EfficientNet B7 to classify three lung diseases (Pneumonia, pneumothorax, tuberculosis, and normal). The method achieved an accuracy of 96.10%. Hu et al. [14] conducted training from scratch using VGG-16 architecture to classify two lung diseases (COVID-19, non-COVID-19 pneumonia, and normal). The method achieved an AUC-ROC of 97.80%. Kim et al. [15] utilised the transfer learning technique with the pre-trained weights of EfficientNet v2M to classify three lung diseases (pneumonia, pneumothorax, tuberculosis, and normal). The method achieved an accuracy of 82.20%. Malik et al. [16] conducted training from scratch using the residual network and dilated convolution to classify four diseases (pneumonia, pneumothorax, tuberculosis, COVID-19, and normal). The method achieved an accuracy of 99.39%. Manalakis et al. [17] used the transfer learning technique with the pre-trained weights of a merged model of DenseNet-121 and ResNet-50 to classify three lung diseases (COVID-19, tuberculosis, pneumonia, and normal). The method achieved an AUC-ROC of 95.00%. Shamrat et al. [18] conducted transfer learning with the pre-trained weights of VGG-16 to classify nine lung diseases (COVID-19, effusion, tuberculosis, pneumonia, lung opacity, mass, nodule, pneumothorax, pulmonary fibrosis, and normal). The method achieved an accuracy of 98.89%. Showkatian et al. [19] conducted training from scratch using the CNN model to classify one lung disease (tuberculosis and Normal). The method achieved an accuracy of 87.00%. Xu et al. [20] conducted training from scratch using DenseNet architecture to classify two lung diseases (tuberculosis, diseases but non-TB, and normal). The method achieved an accuracy of 99.10%.

3. Methods

3.1. Dataset

Chest X-ray images, such as normal, pneumonia, pneumothorax, tuberculosis and lung cancer, were collected from Soonchunhyang University Hospital X-ray Database from May 2022 to July 2022. In total, 5 groups of 1000 dicom image files were collected after confirmation by pulmonary and radiologic specialists at Soonchunhyang University Hospital. Four disease groups were collected at a 1:1 ratio according to disease severity, such as mild and severe groups. The disease severity was anonymized, and the readers (pulmonary and radiologic specialists) were blinded to the clinical data aside from the known diagnosis of pneumonia, pneumothorax, tuberculosis, and lung cancer in all patients. The criteria for disease severity are as follows: Each lung in pneumonia was divided into an upper and a lower zone by equally dividing the hila into upper and lower segments in the craniocaudal dimension. The extent of disease in each of the four zones was scored on a 0–3 scale (0 = no disease, 1 = less than one-third of the zone, 2 = one-third to two-thirds of the zone, and 3 = more than two-thirds of the zone). The severity of disease in each zone was also scored on a 0–3 scale (0 = normal lung parenchyma, 1 = airspace opacification that does not obscure the bronchovascular markings, 2 = airspace opacification that partially obscures the bronchovascular markings, and 3 = complete opacification of the alveolar and interstitial spaces). The extent and severity scores were given as equal weights and were summed over the four lung regions (right upper, right lower, left upper, and left lower) to create a composite score for each CXR ranging from 0 (normal) to 24 [21].
The severity of tuberculosis was calculated by Timika score [22]. The CXR was divided into six zones of a roughly similar size with two horizontal lines and for each zone; the percentage area that showed active disease (consolidation, nodules) involvement was estimated depending on the visual estimation of the extent of opacification (5 or 10–100% in 10% increments), and the CRX score = proportion of total lung affected (%) +40 if cavitations present.
The severity of pneumothorax was calculated by the American College of Chest Physicians guidelines (2001). We measured the size of pneumothorax from thoracic cupola to the lung apex.
The severity of lung cancer was calculated by the eighth edition TNM stage classification for lung cancer, and stage I and II were the mild group and stage III was the severe group [23]. Finally, the Soonchunhyang dataset (n = 5000) was collected as shown in Figure 1, Figure 2 and Figure 3.

3.2. Test Dataset

For the number of test datasets, sample size estimation in diagnostic test studies of biomedical informatics [24] Formula (7.5) (p. 198) and ROC Curves in Clinical Chemistry: Uses, Misuses, and Possible Solutions [25] Formula (2) (p. 1123) were used, and the research hypothesis and calculation basis formula at this time are as follows.
n = Z α V H 0 A U C ^ + Z β V H 1 A U C ^ 2 A U C 1 A U C 0 2 V A U C ^ = 0.0099 × e α 2 / 2 × 6 α 2 + 16 α = 1 A U C × 1.414 1 = N u m b e r   o f   S t a n d a r d   N o r m a l   D i s t r i b u t i o n T y p e   I   e r r o r α = 0.025 O n e   s i d e d T y p e   I I   e r r o r β = 0.1

3.3. OView-AI Workflow

The OView-AI system is a medical image diagnostic aid software which is based on an abnormal classification model learned by artificial intelligence technology using a chest X-ray image to diagnose a prediction rate of each lesion for pneumonia, pneumothorax, tuberculosis, and lung cancer. Its purpose is to assist a doctor’s diagnosis decision by displaying a prediction rate as a percentage (%).
The principal process is as follows:
  • Input a chest X-ray image;
  • Adjust image into 1:1 ratio as the pre-processing;
  • Classify into normal, pneumonia, pneumothorax, tuberculosis, and lung cancer using the deep learning model;
  • Display the classified result as a numerical value (%).

3.4. Deep Learning Algorithm

As suggested by this literature review, the transfer learning technique has proved its significant performance and lower computational cost. Therefore, in our study, we used the transfer learning technique with EfficientNet B7 [9] with the pre-trained weights of ImageNet trained by Noisy Student [10] as a backbone model. Then, this was followed by our proposed fine-tuned layers and hyperparameters. Figure 4 depicts our overall pipeline of multi-classification for lung diseases using the transfer learning technique.

3.5. Deep Learning Model

The EfficientNet [9] model has proven its efficient learning performance by being principally scaled within three main components: width, depth, and resolution of the model. Compound scaling refers to finding the optimal efficiency by adjusting these components. In our study, we exploited the EfficientNet B7 model with the pre-trained weights of ImageNet trained by Noisy Student [10] as our backbone model, as shown in Figure 5.
EfficientNet B7 takes an input tensor of shape (600 × 600 × 3). Thus, the chest X-ray images were tripled before feeding the data into the based model. The based model (EfficientNet B7) consists of a stem block and seven convolutional blocks. The stem block consists of an input layer, rescaling, normalization, zero padding, conv2D, batch normalization, and attention pooling. Each block consists of three sub-blocks. Block 1 consists of sub-block 1, sub-block 3, and two times their additions. Block 2 to block 7 comprises sub-block 2, sub-block 3, and their additions. Each sub-block is a residual network of three modules among five modules. Module 1 consists of depthwise conv2D, batch normalization, and attention pooling. Module 2 consists of depthwise conv2D, batch normalization, attention pooling, zero padding, depthwise conv2D, batch normalization, and attention pooling. Module 3 consists of global average pooling, rescaling, and two conv2Ds. Module 4 consists of multiply, conv2D, and batch normalization. Module 5 consists of Multiply, conv2D, batch normalization, and output layer. The final layers of Efficient B7 were omitted.

3.6. Proposed Fine-Tuned Model

The output of the based model (EfficientNet B7) was a 3D tensor. Thus, global average pooling was applied to flatten the features before feeding them into two 512-FCs (fully connected layers). Then, the Softmax function was applied to classify 5 classes (pneumonia, pneumothorax, tuberculosis, lung cancer, and normal). Figure 6 depicts our proposed fine-tuned model.

3.7. Hyperparameters

Since the size of EfficientNet B7 (the backbone model) is quite large, the risk of overfitting is also high. To solve this problem, the training was optimized by applying the L2 regularization and dropout layer. L2 regularization was applied to vectors passing through the two FC layers.
L 2   r e g u l a r i z a t i o n   t e r m = w = w 1 2 + w 2 2 + w 3 2 + + w n 2
The loss measures how well the model can fit the data, and L2 regularization measures the model complexity. The L2 regularization was calculated by a square root of the sum of the squares of all the feature weights, as shown in Equation (2), before feeding the features into the Softmax function to generate the class probability. The dropout layer with a rate of 0.5 was applied before the last output layer to mit certain neurons.
Since our dataset is imbalanced, the risk of being overly biased to a specific class is also high. To solve this problem, we applied the class weight option during the training. We applied a low weight to a class with a large amount of data and a high weight to a class with a small amount of data. The class weights, C i , were calculated as Equation (3).
C i = M N × S i
where N is the number of classes ( N = 5 ) , M is the total amount of training data ( M = 73,392 ) , and S i is the total amount of data of class i ( i = 0 ,   1 ,   2 ,   3 ,   4 ).
We applied Lookahead [26] to wrap the Nadam optimizer [27]. This optimizer and wrapper prevent our model from local minima, which can solve the overfitting issue. The learning rate scheduling using the warm-up method was also applied. From the first 10% of the epochs, the warm-up learning rate was initial from rate 0 to 0.0001. Then, it stayed as 0.0001 for another 10% of the epochs. After that, it gradually decreased from 0.0001 to 0.00001 using the cosine decay function.
Our ground truth was labeled as the one-hot label, of which the negative was set to 0 and positive was set to 1. We applied the Softmax function as a classifier. Conceptually, the Softmax function generates the class probabilities between 0 and 1, but not exactly 0 and 1. That means that the learning probabilities would not be able to reach the exact 0 and 1. This phenomenon causes an overconfidence issue. To solve this issue, we applied label smoothing [28] with a rate of 0.1 to the loss function. Equation (4) shows the label smoothing ( y l s ) uniform distribution calculation.
y l s = 1 α × y g t + α K
where K is the number of label classes ( K = 5 ) , α is the smoothing rate ( α = 0.1 ) , and y g t is the one-hot encoded label of ground truth. Table 2 shows our label smoothing distribution.

3.8. Environmental Setting

We conducted our experiment on the Linux Ubuntu 18.04 LTS operating system with Intel CPU 19-9940X 3.30 GHz, Nvidia Geforce RTX 3090, RAM 64 GB. The code was written in Python 3.9, Tensorflow 2.5, CUDA 11, and cuDNN 8. The training was set to 50 epochs with a batch size of 16.

4. Experimental Results and Discussion

4.1. Environmental Results

Figure 7 depicts the performance of training and validation. Our learning achieved a training accuracy of 98.13% and a validation accuracy of 97.83, training sensitivity of 97.52% and validation sensitivity of 97.38, and training specificity of 99.64% and validation specificity of 99.31%. We can see that the loss converged well since around epoch 10. After the 10th epoch, the training could be considered to have a stable learning. Thus, 50 epochs were enough to teach our model.
Table 3 shows testing performance by each class in accuracy, sensitivity, and specificity. Our method achieved an accuracy of 98.60% in the normal class, sensitivity of 97.50%, and specificity of 99.83%. Our method achieved an accuracy of 96.91% in the pneumonia class, sensitivity of 95.66%, and specificity of 99.01%. Our method achieved an accuracy of 99.05% in the pneumothorax class, sensitivity of 97.25%, and specificity of 99.15%. Our method achieved an accuracy of 96.88% in the tuberculosis class, sensitivity of 94.98%, and specificity of 99.01%. Our method achieved an accuracy of 95.64% in the lung cancer class, sensitivity of 94.26%, and specificity of 98.04%. On average, our method achieved an accuracy of 97.42%, sensitivity of 95.93%, and specificity of 99.05%.

4.2. Discussion

We conducted an inference of the different 918 chest X-ray images for which the data were also provided after de-identification from Soonchunhyang University Hospital. Table 4 shows the inferencing set by each class. The amount of each disease is almost equal (around 111 images), while the amount of healthy (normal) images is 459. Therefore, the total amount of diseases is 459 images equal to the normal set.
The data were tested with our proposed method. The inference result was evaluated and confirmed by professional radiologists of Soonchunhyang University Hospital, as shown in Table 5. Our method achieved a sensitivity of 0.9412, specificity of 0.9935, overall accuracy of 0.9673, and accuracy under the curve (AUC) of 0.9673.
Figure 8 depicts a receiver operating characteristic (ROC) curve and confusion matrix of the 918-image inferences. Our method had a significant prediction and was less confused with other classes. Our method correctly predicted 456 images (49.67%) from 459 images of a normal class (negative) and 432 images (47.05%) from 459 images of a disease class (positive). In each disease, our method correctly predicted 104 images (11.33%) from 110 images of pneumonia class. There were six images (0.66%) confused with other classes. Our method correctly predicted 110 images (11.98%) from 111 images of pneumothorax class. There was one image (0.11%) confused with other classes. Our method correctly predicted 118 images (12.85%) from 127 images of tuberculosis class. There were nine images (0.99%) confused with other classes. Our method correctly predicted 100 images (10.89%) from 111 images of lung cancer class. There were 11 images (1.21%) confused with other classes.
We omitted eight images with the poorest prediction from the inferencing set. Now, the inferencing dataset is made up of 910 images, as shown in Table 6. We removed two images of pneumonia, three images of tuberculosis, and three images of lung cancer.
In total, 910 images were tested with our method. The inference result was evaluated and confirmed by professional radiologists of Soonchunhyang University Hospital, as shown in Table 7. Our method achieved a sensitivity of 0.9468, specificity of 0.9934, overall accuracy of 0.9703, and accuracy under the curve (AUC) of 0.9701.
Figure 9 depicts a receiver operating characteristic (ROC) curve and confusion matrix of the 910-image inferences. Our method had a significant prediction and was less confused with other classes. Our method correctly predicted 456 images (50.11%) from 459 images of a normal class (negative) and 427 images (46.93%) from 451 images of a disease class (positive). In each disease, our method correctly predicted 103 images (11.32%) from 108 images of pneumonia class. There were four images (0.44%) confused with other classes. Our method correctly predicted 110 images (11.98%) from 111 images of pneumothorax class. There was one image (0.11%) confused with other classes. Our method correctly predicted 116 images (12.75%) from 124 images of tuberculosis class. There were eight images (0.88%) confused with other classes. Our method correctly predicted 98 images (10.77%) from 108 images of lung cancer class. There were 10 images (1.10%) confused with other classes.
Table 8 summarizes and compares the inferences of the 918-image set and 910-image set. With the 918-image set, our method achieved an overall accuracy of 96.73%, AUC of 96.73%, and 2.6282 s per inference. With the 918-image set, our method performed better and achieved an overall accuracy of 97.03%, AUC of 97.01%, and 2.6275 s per inference.

5. Conclusions

We proposed a classification of four lung diseases and healthy lungs using a deep learning method. Our proposed fine-tuned deep learning model, which composed of EfficientNet B7 as the backbone, fine-tuned layers, and hyperparameters, showed itself to present a significant performance in multi-lung infections. Chest X-ray images were collected from Soonchunhyang University Hospital after de-identification from May 2022 to July 2022. Our method achieved an average test accuracy, sensitivity, and specificity of 97.42%, 95.93%, and 99.05%, respectively. Additionally, we conducted an inference on other different 910 chest X-ray images. The AUC confidence interval (95%CI) of diagnostic results in the OView-AI system (diagnostic supporting software detects normal lung cancer, pneumonia, pulmonary tuberculosis, and pneumothorax using chest X-ray images) was 97.01%. Moreover, the sensitivity and specificity of the OView-AI system were 94.68% and 99.34%, respectively.
The inference experiment was evaluated and confirmed by professional radiologists of Soonchunhyang University Hospital. The inferencing results expressed that our method had a highly correct prediction of healthy lungs (negative class) and less confused other diseases (positive class). Among diseases, our method had the highest correct prediction of the pneumothorax class. Our findings demonstrated the potential to fasten the clinical workflow and facilitate the early screening stage for multi-chest infections as the computer-aided application (OView-AI System), which supports clinical decision making.

Author Contributions

Conceptualization and supervision, M.H. and S.C.; data curation, J.L. and S.H.; methodology, J.O. and C.P.; writing—original draft preparation, J.O. and C.P.; writing—review and editing, H.L., B.R. and Y.K.; validation, J.L., S.H., M.H. and S.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Ministry of Education (MOE, Korea) through the fostering project of ‘Soonchunhyang University and Industry Cooperation Complex’ supervised by the Korea Institute for Advancement of Technology (KIAT) (Number: 20221042), and the Soonchunhyang University Research Fund.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Institutional Review Board of Soonchunhyang University Hospital (2020-12-036-002).

Informed Consent Statement

Patient consent was waived due to the prospective design of this study.

Data Availability Statement

No new data were created in this study. Data sharing is not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lung Diseases. National Institute of Environmental Health Sciences. Available online: https://www.niehs.nih.gov/health/topics/conditions/lung-disease/index.cfm (accessed on 28 November 2022).
  2. Dodia, S.; Aannappa, B.; Mahesh, P.A. Recent Advancements in Deep Learning based Lung Cancer Detection: A Systematic Review. Eng. Appl. Artif. Intell. 2022, 116, 105490. [Google Scholar] [CrossRef]
  3. Pneumonia System and Diagnosis. American Lung Association. Available online: https://www.lung.org/lung-health-diseases/lung-disease-lookup/pneumonia/symptoms-and-diagnosis (accessed on 28 November 2022).
  4. Tuberculosis System and Diagnosis. American Lung Association. Available online: https://www.lung.org/lung-health-diseases/lung-disease-lookup/tuberculosis/symptoms-diagnosis (accessed on 28 November 2022).
  5. Lee, J.G.; Kim, H.C.; Choi, C.M. Recent Trends of Lung Cancer in Korea. Tuberc. Respir. Dis. 2021, 84, 89. [Google Scholar] [CrossRef] [PubMed]
  6. Symptoms, Diagnosis and Treating Pneumothorax. American Lung Association. Available online: https://www.lung.org/lung-health-diseases/lung-disease-lookup/pneumothorax/symptoms-diagnosis-treatment (accessed on 28 November 2022).
  7. Park, S.; Jeong, W.; Moon, Y.S. X-ray image segmentation using multi-task learning. KSII Trans. Internet Inf. Syst. 2020, 14, 1104–1120. [Google Scholar]
  8. Shah, A.; Shah, M. Advancement of Deep Learning in Pneumonia/COVID-19 Classification and Localization: A Systematic Review with Qualitative and Quantitative Analysis. Chronic Dis. Transl. Med. 2022, 8, 154–171. [Google Scholar] [CrossRef] [PubMed]
  9. Tan, M.; Le, Q. Efficientnet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  10. Xie, Q.; Luong, M.T.; Hovy, E.; Le, Q.V. Self-training with Noisy Student Improves ImageNet Classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 14–19 May 2020; pp. 10687–10698. [Google Scholar]
  11. Asif, S.; Wenhui, Y.; Amjad, K.; Jin, H.; Tao, Y.; Jinhai, S. Detection of COVID-19 from Chest X-ray Images: Boosting the Performance with Convolutional Neural Network and Transfer Learning. Expert Syst. 2022; Online ahead of print. [Google Scholar]
  12. Hamwi, W.A.; Almustafa, M.M. Development and Integration of VGG and Dense Transfer-learning Systems Supported with Diverse Lung Images for Discovery of the Coronavirus Identity. Inform. Med. Unlocked 2022, 32, 101004. [Google Scholar] [CrossRef] [PubMed]
  13. Hong, M.; Rim, B.; Lee, H.; Jang, H.; Oh, J.; Choi, S. Multi-class Classification of Lung Diseases using CNN Models. Appl. Sci. 2021, 11, 9289. [Google Scholar] [CrossRef]
  14. Hu, Z.; Yang, Z.; Lafata, K.J.; Yin, F.F.; Wang, C. A Radiomics-boosted Deep-learning Model for COVID-19 and non-COVID-19 Pneumonia Classification using Chest X-ray Images. Med. Phys. 2022, 49, 3213–3222. [Google Scholar] [CrossRef] [PubMed]
  15. Kim, S.; Rim, B.; Choi, S.; Lee, A.; Min, S.; Hong, M. Deep Learning in Multi-Class Lung Diseases’ Classification on Chest X-ray Images. Diagnostics 2022, 12, 915. [Google Scholar] [CrossRef] [PubMed]
  16. Malik, H.; Anees, T.; Din, M.; Naeem, A. CDC_Net: Multi-classification Convolutional Neural Network Model for Detection of COVID-19, Pneumothorax, Pneumonia, Lung Cancer, and Tuberculosis using Chest X-rays. Multimed. Tools Appl. 2022, 82, 13855–13880. [Google Scholar] [CrossRef] [PubMed]
  17. Mamalakis, M.; Swift, A.J.; Vorselaars, B.; Ray, S.; Weeks, S.; Ding, W.; Clayton, R.H.; Mackenzie, L.S.; Banerjee, A. DenResCov-19: A Deep Transfer Learning Network for Robust Automatic Classification of COVID-19, Pneumonia, and Tuberculosis from X-rays. Comput. Med. Imaging Graph. 2021, 94, 102008. [Google Scholar] [CrossRef] [PubMed]
  18. Shamrat, F.J.M.; Azam, S.; Karim, A.; Islam, R.; Tasnim, Z.; Ghosh, P.; De Boer, F. LungNet22: A Fine-Tuned Model for Multiclass Classification and Prediction of Lung Disease Using X-ray Images. J. Pers. Med. 2022, 12, 680. [Google Scholar] [CrossRef] [PubMed]
  19. Showkatian, E.; Salehi, M.; Ghaffari, H.; Reiazi, R.; Sadighi, N. Deep Learning-based Automatic Detection of Tuberculosis Disease in Chest X-ray Images. Pol. J. Radiol. 2022, 87, 118–124. [Google Scholar] [CrossRef] [PubMed]
  20. Xu, X.; Zhang, J.; Cheng, X.; Lu, L.; Zhao, Y.; Xu, Z.; Gu, Z. A Tuberculosis Detection Method Using Attention and Sparse R-CNN. KSII Trans. Internet Inf. Syst. 2022, 16, 2131–2153. [Google Scholar]
  21. Reeves, R.A.; Pomeranz, C.; Gomella, A.A.; Gulati, A.; Metra, B.; Hage, A.; Lange, S.; Parekh, M.; Donuru, A.; Lakhani, P.; et al. Performance of a severity score on admission chest radiograph in predicting clinical outcomes in hospitalized patients with coronavirus disease (COVID-19). Am. J. Roentgenol. 2021, 217, 623–632. [Google Scholar] [CrossRef] [PubMed]
  22. Ralph, A.P.; Ardian, M.; Wiguna, A.; Maguire, G.P.; Becker, N.G.; Drogumuller, G.; Wilks, M.J.; Waramori, G.; Tjitra, E.; Kenagalem, E.; et al. A simple, valid, numerical score for grading chest X-ray severity in adult smear-positive pulmonary tuberculosis. Thorax 2010, 65, 863–869. [Google Scholar] [CrossRef] [PubMed]
  23. Detterbeck, F.C. The eighth edition TNM stage classification for lung cancer: What does it mean on main street? J. Thorac. Cardiovasc. Surg. 2018, 155, 356–359. [Google Scholar] [CrossRef] [PubMed]
  24. Hajian-Tilaki, K. Sample size estimation in diagnostic test studies of biomedical informatics. J. Biomed. Inform. 2014, 48, 193–204. [Google Scholar] [CrossRef] [PubMed]
  25. Obuchowski, N.A.; Lieber, M.L.; Wians, F.H., Jr. ROC curves in clinical chemistry: Uses, misuses, and possible solutions. Clin. Chem. 2004, 50, 1118–1125. [Google Scholar] [CrossRef] [PubMed]
  26. Zhang, M.R.; Lucas, J.; Hinton, G.; Ba, J. Lookahead Optimizer: K Steps Forward, 1 Step Back. arXiv 2019, arXiv:1907.08610. [Google Scholar]
  27. Dozat, T. Technical Report, Incorporating Nesterov Momentum into Adam. In Proceedings of the ICLR Workshop, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  28. Müller, R.; Kornblith, S.; Hinton, G. When Does Label Smoothing Help? arXiv 2019, arXiv:1906.02629. [Google Scholar]
Figure 1. Overall study flow of dataset collecting and labeling for the multi-classification for lung diseases.
Figure 1. Overall study flow of dataset collecting and labeling for the multi-classification for lung diseases.
Diagnostics 13 01519 g001
Figure 2. Composition of dataset and test dataset.
Figure 2. Composition of dataset and test dataset.
Diagnostics 13 01519 g002
Figure 3. Lung diseases dataset: (a) normal; (b) pneumonia; (c) pneumothorax; (d) tuberculosis; (e) lung cancer.
Figure 3. Lung diseases dataset: (a) normal; (b) pneumonia; (c) pneumothorax; (d) tuberculosis; (e) lung cancer.
Diagnostics 13 01519 g003
Figure 4. Overall pipeline of multi-classification for lung diseases using transfer learning technique.
Figure 4. Overall pipeline of multi-classification for lung diseases using transfer learning technique.
Diagnostics 13 01519 g004
Figure 5. EfficientNet B7 architecture as our backbone model.
Figure 5. EfficientNet B7 architecture as our backbone model.
Diagnostics 13 01519 g005
Figure 6. Our proposed fine-tuned model.
Figure 6. Our proposed fine-tuned model.
Diagnostics 13 01519 g006
Figure 7. Training and validation performance: (a) accuracy; (b) sensitivity; (c) specificity.
Figure 7. Training and validation performance: (a) accuracy; (b) sensitivity; (c) specificity.
Diagnostics 13 01519 g007
Figure 8. Results of inferencing on 918 images: (a) receiver operating characteristic (ROC) curve; (b) confusion matrix.
Figure 8. Results of inferencing on 918 images: (a) receiver operating characteristic (ROC) curve; (b) confusion matrix.
Diagnostics 13 01519 g008
Figure 9. Results of inferencing on 910 images: (a) receiver operating characteristic (ROC) curve; (b) confusion matrix.
Figure 9. Results of inferencing on 910 images: (a) receiver operating characteristic (ROC) curve; (b) confusion matrix.
Diagnostics 13 01519 g009aDiagnostics 13 01519 g009b
Table 1. Related papers conducted on classification for lung diseases.
Table 1. Related papers conducted on classification for lung diseases.
PaperDiseaseClassModelResult
Asif et al. (2022)
[11]
COVID-19, viral pneumonia, and normal3Transfer learning
(VGG-16)
Accuracy = 97.84%
Hamwi et al. (2022)
[12]
COVID-19, and normal2Transfer learning (merged model of VGG-16 and DenseNet-201)Accuracy = 99.73%
Hong et al. (2021)
[13]
Pneumonia, pneumothorax, tuberculosis, and normal 4Transfer learning
(EfficienNet B7)
Accuracy = 96.10%
Hu et al. (2022)
[14]
COVID-19, non-COVID-19 pneumonia, and normal3Train from scratch (VGG-16)AUC-ROC = 97.80%
Kim et al. (2022)
[15]
Pneumonia, pneumothorax, tuberculosis, and normal4Transfer learning
(EfficientNet v2M)
Accuracy = 82.20%
Malik et al. (2022)
[16]
Pneumonia, pneumothorax, tuberculosis, COVID-19, and normal5Train from scratch
(residual network and dilated convolution)
Accuracy = 99.39%
Manalakis et al. (2021)
[17]
COVID-19, tuberculosis, pneumonia, and normal4Transfer learning (merged model of DenseNet-121 and ResNet-50)AUC-ROC = 95.00%
Shamrat et al. (2022)
[18]
COVID-19, effusion, tuberculosis, pneumonia, lung opacity, mass, nodule, pneumothorax, pulmonary fibrosis, and normal10Transfer learning
(VGG-16)
Accuracy = 98.89%
Showkatian et al. (2022)
[19]
Tuberculosis and normal2Train CNN from scratch (CNN)Accuracy = 87.00%
Xu et al. (2022)
[20]
Tuberculosis (TB), diseases but non-TB, and normal3Train from scratch
(DenseNet)
Accuracy = 99.10%
Table 2. Our label smoothing distribution.
Table 2. Our label smoothing distribution.
Class StateGround Truth Label ( y g t ) Label Smoothing ( y l s )
Negative00.02
Positive10.92
Table 3. Testing performance by each class.
Table 3. Testing performance by each class.
ClassAccuracySensitivitySpecificity
Normal98.60%97.50%99.83%
Pneumonia96.91%95.66%99.21%
Pneumothorax99.05%97.25%99.15%
Tuberculosis96.88%94.98%99.01%
Lung cancer95.64%94.26%98.04%
Average97.42%95.93%99.05%
Table 4. Inferencing dataset collected from Soonchunhyang University Hospital.
Table 4. Inferencing dataset collected from Soonchunhyang University Hospital.
ClassInferencing Set Inferencing Set
Normal459Healthy459
Pneumonia110Diseases459
Pneumothorax111
Tuberculosis127
Lung cancer111
Total918 918
Table 5. Evaluation result of our method on the 918-image inferences.
Table 5. Evaluation result of our method on the 918-image inferences.
EquationProportion
Estimate
Confidence Interval
LowerUpper
SensitivityTP/(TP + FN)0.94120.91970.9627
SpecificityTN/(FP + TN)0.99350.98611.0008
Likelihood ratio +TP/(TP + FN)/(FP/FP + TN))144.000046.6037444.9436
Likelihood ratio −FN/(TP + FN)/(TN/FP + TN))0.05920.04110.0854
False positive rateFP/(FP + TN)0.0065−0.00080.0139
False negative rateFN/(TP + FN)0.05880.03730.0798
Prob of disease(TP + FN)/(TP + FP + FN + TN)0.50000.46770.5323
Positive predictive valueTP/(TP + FP)0.99310.98531.0009
p (pos test wrong)FP/(TP + FP)0.0069−0.00090.0147
Negative predictive valueTN/(FN + TN)0.94410.92360.9646
p (neg test wrong)FN/(FN + TN)0.05590.03540.0764
p (test positive)(TP + FP)/(TP + FP + FN + TN)0.47390.44160.5062
p (test negative)(FN + TN)/(TP + FP + FN + TN)0.52610.49380.5584
Overall accuracy(TP + TN)/(TP + FP + FN + TN)0.96730.95580.9788
AUC(Sensitivity + Specificity)/20.96730.95550.9791
TP, TN, FP, and FN are true positive, true negative, false positive, and false negative, respectively.
Table 6. Inferencing dataset collected from Soonchunhyang University Hospital.
Table 6. Inferencing dataset collected from Soonchunhyang University Hospital.
ClassInferencing Set Inferencing Set
Normal459Healthy459
Pneumonia108Diseases451
Pneumothorax111
Tuberculosis124
Lung cancer108
Total910 910
Table 7. Inferencing dataset collected from Soonchunhyang University Hospital.
Table 7. Inferencing dataset collected from Soonchunhyang University Hospital.
EquationProportion
Estimate
Confidence Interval
LowerUpper
SensitivityTP/(TP + FN)0.94680.92610.9675
SpecificityTN/(FP + TN)0.99340.98611.0008
Likelihood ratio +TP/(TP + FN)/(FP/FP + TN))144.542546.7805446.6075
Likelihood ratio −FN/(TP + FN)/(TN/FP + TN))0.05360.03630.0791
False positive rateFP/(FP + TN)0.0066−0.00080.0139
False negative rateFN/(TP + FN)0.05320.03250.0733
Prob of disease(TP + FN)/(TP + FP + FN + TN)0.49610.46360.5287
Positive predictive valueTP/(TP + FP)0.99300.98521.0009
p(pos test wrong)FP/(TP + FP)0.0070−0.00090.0148
Negative predictive valueTN/(FN + TN)0.94990.93040.9694
p(neg test wrong)FN/(FN + TN)0.05010.03060.0696
p(test positive)(TP + FP)/(TP + FP + FN + TN)0.47300.44060.5055
p(test negative)(FN + TN)/(TP + FP + FN + TN)0.52700.49450.5594
Overall accuracy(TP + TN)/(TP + FP + FN + TN)0.97030.95930.9813
AUC(Sensitivity + Specificity)/20.97010.95880.9815
TP, TN, FP, and FN are true positive, true negative, false positive, and false negative, respectively.
Table 8. Inference overall accuracy, accuracy under the curve, and time.
Table 8. Inference overall accuracy, accuracy under the curve, and time.
InferenceOverall
Accuracy
AUCTime
per Inference
918 images96.73%96.73%2.6282 s
910 images97.03%97.01%2.6275 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Oh, J.; Park, C.; Lee, H.; Rim, B.; Kim, Y.; Hong, M.; Lyu, J.; Han, S.; Choi, S. OView-AI Supporter for Classifying Pneumonia, Pneumothorax, Tuberculosis, Lung Cancer Chest X-ray Images Using Multi-Stage Superpixels Classification. Diagnostics 2023, 13, 1519. https://doi.org/10.3390/diagnostics13091519

AMA Style

Oh J, Park C, Lee H, Rim B, Kim Y, Hong M, Lyu J, Han S, Choi S. OView-AI Supporter for Classifying Pneumonia, Pneumothorax, Tuberculosis, Lung Cancer Chest X-ray Images Using Multi-Stage Superpixels Classification. Diagnostics. 2023; 13(9):1519. https://doi.org/10.3390/diagnostics13091519

Chicago/Turabian Style

Oh, Joonho, Chanho Park, Hongchang Lee, Beanbonyka Rim, Younggyu Kim, Min Hong, Jiwon Lyu, Suha Han, and Seongjun Choi. 2023. "OView-AI Supporter for Classifying Pneumonia, Pneumothorax, Tuberculosis, Lung Cancer Chest X-ray Images Using Multi-Stage Superpixels Classification" Diagnostics 13, no. 9: 1519. https://doi.org/10.3390/diagnostics13091519

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop