Next Article in Journal
Molecularly Confirmed Pseudomyogenic Hemangioendothelioma with Unusual EGFL7::FOSB Fusion in the Head and Neck Region of an Older Patient
Previous Article in Journal
The Role of New Morphological Parameters Provided by the BC 6800 Plus Analyzer in the Early Diagnosis of Sepsis
Previous Article in Special Issue
Retinal Disease Diagnosis Using Deep Learning on Ultra-Wide-Field Fundus Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Severe Lung Infection on Chest Radiographs of COVID-19 Patients: Robustness of AI Models across Multi-Institutional Data

1
Department of Radiology, University of Michigan, Ann Arbor, MI 48109, USA
2
Office of Science and Engineering Laboratories, Center for Devices and Radiological Health, U.S. Food and Drug Administration, Silver Spring, MD 20993, USA
3
Department of Radiology, New York University, New York, NY 10016, USA
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(3), 341; https://doi.org/10.3390/diagnostics14030341
Submission received: 27 November 2023 / Revised: 24 January 2024 / Accepted: 30 January 2024 / Published: 5 February 2024
(This article belongs to the Special Issue Artificial Intelligence in Biomedical Image Analysis)

Abstract

:
The diagnosis of severe COVID-19 lung infection is important because it carries a higher risk for the patient and requires prompt treatment with oxygen therapy and hospitalization while those with less severe lung infection often stay on observation. Also, severe infections are more likely to have long-standing residual changes in their lungs and may need follow-up imaging. We have developed deep learning neural network models for classifying severe vs. non-severe lung infections in COVID-19 patients on chest radiographs (CXR). A deep learning U-Net model was developed to segment the lungs. Inception-v1 and Inception-v4 models were trained for the classification of severe vs. non-severe COVID-19 infection. Four CXR datasets from multi-country and multi-institutional sources were used to develop and evaluate the models. The combined dataset consisted of 5748 cases and 6193 CXR images with physicians’ severity ratings as reference standard. The area under the receiver operating characteristic curve (AUC) was used to evaluate model performance. We studied the reproducibility of classification performance using the different combinations of training and validation data sets. We also evaluated the generalizability of the trained deep learning models using both independent internal and external test sets. The Inception-v1 based models achieved AUC ranging between 0.81 ± 0.02 and 0.84 ± 0.0, while the Inception-v4 models achieved AUC in the range of 0.85 ± 0.06 and 0.89 ± 0.01, on the independent test sets, respectively. These results demonstrate the promise of using deep learning models in differentiating COVID-19 patients with severe from non-severe lung infection on chest radiographs.

1. Introduction

Deep-learning-based models, commonly referred to as artificial intelligence (AI) models, have been developed to assist physicians in analyzing medical images and in making diagnostic decisions in the past decade. AI models have been developed to analyze chest radiographs (CXR) for classifying various lung diseases. For COVID-19, AI has been shown to quantify the extent of lung involvement that correlates with ICU admission, intubation, and death [1]. Reliable quantification of disease characteristics using AI models has the potential to help monitor disease progression or regression, reducing some of the healthcare burden.
Characterization of the level of lung infection on chest radiograph of COVID-19 patients is generally a challenging task due to the wide variation of its appearance. Nair et al. [2] conducted a study that quantified reader agreement in their diagnosis of COVID-19. In their study, there were four groups of readers: chest consultant, general consultant, specialist registrar radiologists, and infectious disease clinicians. They concluded that the reader agreement was low. AI-aided reading could potentially mitigate this issue as has been shown in other types of clinical tasks [3,4,5]. For instance, Figure 1 illustrates the different degrees of severity from COVID-19 infection: (1) normal, (2) mild, (3) moderate, and (4) severe [6,7,8]. Visually, in terms of pixel density, shape, and texture, while normal and mild may be distinguishable from those of severe cases, moderate and severe COVID-19 infections are much more similar.
Previous studies have proposed AI methods for identifying severe COVID-19 infection. Most of the methods use CT scans, because CT scans provide more information than two-dimensional X-ray images. On the other hand, CT requires higher cost equipment and relatively higher radiation dose than CXR and may not be readily available in resource-limited communities. CXR is a widely used low-cost and efficient detection method that may be used in the evaluation of the degree of lung involvement at the initial screening of suspected cases of COVID-19 infection and subsequent monitoring over the course of treatment. Most importantly, CXR can be acquired with portable bedside equipment without moving the patient from isolated infectious disease wards to the CT suites.
Not all previous studies used imaging information. Feyaerts et al. [9] for instance proposed an integrated computational approach to analyze the combined plasma and single-cell proteomic data and the severity was classified as mild, moderate, or severe. The results provided a set of early determinants of COVID-19 severity that may point to therapeutic targets for the prevention of COVID19 progression, with the area under the receiver operating characteristic curve (AUC) ranging between 0.773 and 0.799. In a more recent work, Danilov et al. [10] proposed an approach for automatic scoring of COVID-19 severity by X-ray imaging based on a deep learning workflow. Their approach first segmented the lung and then scored the severity. Frid-Adar et al. [11] estimated the severity of pneumonia in COVID-19 patients. They segmented the lung and the region of pneumonia and then performed severity classification. Table 1 shows an overview of the related work. There is a limited number of published articles for detection of severe COVID-19 infection versus other severity classes as moderate, mild, and normal lungs. However, a larger amount of work was devoted to classifying COVID-19 infection versus normal lungs and/or pneumonia without considering its severity. The following works considered multi-class deep learning classification approaches for stratifying X-ray lung images into different number of categories. For instance, Namburu et al. [12] considered seven categories: COVID-19, virus, bacteria, ARDS, SARS, Streptococcus, and normal. Ren et al. [13] used four categories: COVID-19 infection, bacterial infection, virus infection, and normal lungs. Hadi et al. [14] considered only three categories: COVID-19, normal, and viral pneumonia, so did Constantinou et al. [15]: COVID-19, non-COVID-19, and normal, and Ullah et al. [16]: COVID-19, pneumonia, and normal. Some studies performed multi-region or multi-class classification of COVID-19 infection. Samala et al. [17] demonstrated that deep-learning-based quantitative severity descriptors on CXR images had significant correlation with radiologist’s severity ratings and had the potential to classify normal lungs from those with COVID-19 infection. Park et al. [18] divided the CXR images into six regions and the AI model analyzed each region to quantify the COVID-19 severity: 0 for normal and 1 for severe. In the end, COVID-19 severity was scored in a range between 0 and 6. Sahoo et al. [19] proposed a multi-stage system for COVID-19 severity assessment, classifying as mild, moderate, severe, and critical, on CXR images.
Although CXRs of COVID-19 are taken all around the world every day, available datasets for AI model training are still limited due to the lack of expert labels and the difficulties in sharing patient data outside the hospital for privacy issues [18]. There is a considerable number of datasets containing labels as negative or positive for COVID-19 [24,25,26]. For our work, we utilized the datasets with COVID-19 severity labels. To our knowledge, publicly available datasets of CXRs containing severity labels for COVID-19 are those shown in Table 2.
In this study, we classified the patient’s lung infection into severe vs. non-severe because severe COVID infection carries a higher risk for the patient and requires prompt treatment with oxygen therapy and hospitalization while less severe levels of infection often stay on observation. Also, severe infections are more likely to have long standing residual changes in their lungs and may need follow up imaging [33]. As summarized in Table 1, most of the previous studies that used CXRs classified COVID-19 versus non-COVID-19 rather than severe versus non-severe infection.
The specific contributions of our work include:
(1)
Tackling a challenging task of classifying severe from non-severe lung infection in COVID patients using a dataset of CXR images containing various degrees of severity (normal, mild, moderate, and severe) infection.
(2)
Studying the reproducibility of the performances of the deep learning models trained with data sets of limited sizes and from multi-institutional sources.
(3)
Evaluating the generalizability of the trained deep learning models by using both independent internal and external test sets.

2. Materials and Methods

2.1. Datasets

We selected four databases in our study where images and/or cases contained severity labels as shown in Table 3. The selected databases were COVIDGR [23,31], BrixIA [28,29,30], MIDRC [32], and our institutional database from the University of Michigan (UMICH). Institutional Review Board (IRB) approval was obtained for retrospective collection of the UMICH database with waiver of informed consent. We did not include the data set by Cohen et al. [27] because their definitions of the levels of severity were very different from those of the other data sets. The combined dataset consisted of 5748 cases and 6193 CXR images with severity ratings as described in Table 2.
For the segmentation step, we used an ImageNet pre-trained U-Net model and further pre-trained it with a subset of the NIH CXR images (N = 277). The U-Net model was then fine-tuned and evaluated with a small subset of the MIDRC data set, including 30 images for training, 7 images for validation, and 39 images for testing. The reference segmentation of the lungs in these development data sets was manually outlined by trained researchers.
For our severity assessment study, we considered the images that had been labeled as one of the following: negative, mild, moderate, or severe defined by TCIA [7,8]:
  • Negative: negative for pneumonia, no lung opacities.
  • Mild: Required if not negative for pneumonia. Opacities in 1–2 lung zones.
  • Moderate: Required if not negative for pneumonia. Opacities in 3–4 lung zones.
  • Severe: Required if not negative for pneumonia. Opacities in >4 lung zones.
The MIDRC set included images in DICOM format. Images were acquired with flat-panel digital detectors (DX) and computed radiography (CR). Each image was read, and the COVID-19 severity was rated independently by three physicians. Most of the images had consistent ratings from the three readers. For the images that obtained different severity ratings, we used the majority (in case of 2 equal ratings) or average (in case of 3 different ratings) of the severity ratings as the label. The BrixIA set was contributed by an Italian group and included CXR images of COVID-19 subjects acquired with both CR and DX modalities. All data were anonymized DICOM files and annotation files in CSV with BrixIA score and relevant metadata. The severity ratings had consensus from five radiologists [30]. The COVIDGR set included anonymized X-ray images in Joint Photographic Experts Group (JPG) format. It was collected under a collaboration with an expert radiologist team of the Hospital Universitario San Cecilio, Spain [31]. The UMICH set included de-identified DX images in DICOM format. One experienced chest radiologist provided the severity rating according to the TCIA definition. A DICOM reader was used to read the CXR images in DICOM format that provides all information about the image in the header including the image size and gray level depth. The header information was used to read out the image as it was originally stored. The images were preprocessed as described in “Section 2.2.2.1. Data Harmonization”.
We proposed to develop a deep learning model that would classify severe lung infection versus the rest (normal, mild, and moderate) in COVID-19 patients. Table 3 shows the partitioning of the data sets by case into the training, validation, and test subsets, i.e., all images from the same patient were always grouped into the same subset to maintain the independence among the three subsets. It also shows the distribution of severe and non-severe images for each set. For training deep learning models, it is preferable to have a balanced dataset. However, severe and non-severe cases often are not proportionally distributed. Because of the limited sizes of the available datasets, we used all eligible images for this study.

2.2. Deep-Learning-Based Processing Pipeline

In this work, we implemented a deep-learning-based pipeline for automatically identifying severe lung infection on CXR images of COVID-19 patients. Figure 2 illustrates our proposed processing pipeline that consisted of two main stages: (1) lung segmentation and (2) severity classification. It combined deep learning models and conventional methods such as edge detection, hole filling, and basic statistical operations. Each stage is described in detail in the following subsections.

2.2.1. Lung Segmentation

Figure 3 shows the process of our segmentation method. The segmentation was intended to focus the AI model on the regions within the lungs and avoid shortcut learning. We used U-Net [34] based architecture for lung segmentation.
The U-Net was trained using transfer learning with two pre-training stages: first by ImageNet, then by 277 CXR images of the NIH database [35], and finally fine-tuned by a small subset of the MIDRC data set. The pre-training with the NIH dataset adapted the weights to CXR images in general while the fine tuning with the MIDRC data set further adapted the U-Net to CXRs of COVID patients, thus improving the segmentation accuracy. The U-Net was trained to output the segmented right and left lungs (Figure 3). The U-Net segmented output image was thresholded and dilated to obtain the final binary lung mask. However, sometimes the U-Net output noise or disconnected parts of the lungs. For these situations, an automated post-processing technique was used to select and keep only the two biggest objects which were usually the lungs. Finally, a hole-filling technique was used to remove the holes, if any, in the segmented lung masks.

2.2.2. Severity Classification

We trained three models by using different training datasets:
  • Model M: Trained only with the MIDRC training set.
  • Model MB: Trained with the combined MIDRC + BrixIA training set.
  • Model MBC: Trained with the combined MIDRC + BrixIA + COVIDGR training set.
For severity assessment, we used the Inception-v1 [36] and Inception-v4 [37] deep neural network architectures. The selection of the Inception architectures was based on our prior studies related to breast cancer diagnosis on mammograms and breast tomosynthesis images [38,39,40], where these architectures showed robust and accurate performance. Inception-v1 includes 9 inception blocks and 5 million parameters. Inception-v4 is a deeper architecture with 14 inception blocks and 43 million parameters and potentially has better learning capacity. Because of the limited training set sizes and training efficiency, the Inception-v1 structure was used in most of the experiments. Both models were pre-trained with ImageNet dataset. We fine-tuned the last 3 blocks of Inception-v1 and the last 2 blocks of Inception-v4. We also compared the performances of the Inception-v1 model with the Inception-v4 model.
The segmented lung regions from the U-Net were cropped with a square bounding box. The cropped region was resampled to a matrix size of 480 × 480 pixels. The segmented lung images focused the attention of the Inception network on only relevant information within the lungs by excluding the anatomical background structures.

2.2.2.1. Data Harmonization

To harmonize the wide variations of pixel intensity distributions of the CXR images from different equipment and clinical sites, we experimentally designed the following image pre-processing steps and selected the parameters using small subsets of the training and validation sets that were not part of the independent test set. We pre-processed the masked lung images by reducing the gray levels of all images to 8 bits, shifting the mean pixel intensity in the lung regions to 128, and scaling the standard deviation of all images to one single value, 22 for DX images and 26 for CR images. These values (22 and 26) were the average of the standard deviations of the pixel intensity distributions over images of the respective modality from the training set. The pixels outside the lung regions were set to a constant value of zero. Figure 4 presents six images where the images on the top row do not have the mean and standard deviation shifted and the images on the bottom row are processed by adjusting the mean and standard deviation. It can be observed that the images on the bottom row have more uniform appearance compared to the images on the top row.

2.2.2.2. Data Augmentation

We applied data augmentation to the training data set where each image was flipped horizontally. The augmented data set was used as input to train the deep learning models.
Our experiments were performed on GPUs—Nvidia GeForce GTX 1080Ti with 11 GB of memory (Santa Clara, CA, USA).

3. Results

3.1. Lung Segmentation

The lung segmentation performance was assessed by comparing the U-Net model’s segmentation results to the manual outlines on the MIDRC test subset of 39 images. To assess the segmentation accuracy, we calculated the quantitative measures including the Jaccard index, the Dice Coefficient, the Hausdorff distance, and the average Euclidean distance relative to the reference. Table 4 shows the four mean segmentation performance measures obtained by averaging over the images in the test subset.
Although the lungs in general are correctly segmented, the lung segmentation on CR images is less accurate than on DX images. This is because the U-Net was trained with only DX images due to DX is the predominant modality and the manual outlines were obtained only for DX images.
Figure 5 illustrates six segmentation results on test images.
The three images on the left side demonstrate the limitation of the segmentation method where the lung region was almost invisible due to infection. The other three cases represent images for which the segmentation agrees well with the manual outlines.

3.2. Severity Assessment

To study the effects of different training sets on the performance of the deep learning model for severity classification, we trained the Inception-v1 model with three different training sets, described above. Figure 6 shows the AUC of three different models, M, MB, and MBC on three different validation sets as the number of iterations increased during training. There was no notable improvement with the addition of more data by combining the different data sets.
We evaluated the consistency of the models when we repeated the training with different random initialization. For each repeated experiment, the weights of the last fully connected layer were randomly initialized and the training dataset batches were sampled with a different random seed. We studied the impact of batch size on models trained on the MIDRC training dataset. Figure 7 shows the validation AUC-vs-epoch curves for batch sizes of 16, 32, and 64 for the Inception-v1 model trained with the MIDRC training set. We plotted the horizontal axis in terms of epoch because the number of epochs was kept the same for the training of the different models while the number of iterations scaled with the batch size (i.e., number of iterations = total number of training images/batch size). For each batch size, the model was trained with five different random initializations and deployed on the three different validation sets (MIDRC, BrixIA, and COVIDGR). The results show that the batch size of 16 obtained a more stable performance on the validation sets. The same trend can also be seen in Figure 8 where the models were trained with the combined MIDRC + BrixIA training set. The batch size of 16 was therefore used in most of our model training unless specified.
Figure 9 compares the performance of two different deep learning architectures, Inception-v1 and Inception-v4, on three different validation sets. The Inception-v1 model achieved better and more stable performance than the Inception-v4 model on the three different validation sets when they were both trained with the MIDRC training set. With the combined MIDRC + BrixIA training set, the Inception-v4 model could also converge to stable performance after about 150 epochs, indicating that the MIDRC training set may be too small to train the Inception-v4 model.

3.3. Severity Classification

After the models were trained, we selected a checkpoint as a frozen model in the relatively stable region of the validation curves and deployed the models to the independent held-out test sets. From the validation curves, we observed that 200 epochs could reach stability for the different conditions that we studied so that it was selected as the end point for all models. For each training set and batch size, we trained the model five times with different random initialization to perform sensitivity analysis and estimated the mean and standard deviation of the AUC on our deep learning pipeline for severity assessment. Table 5 shows the mean and the standard deviation of AUC deployed on different test sets for the Inception-v1 models trained with different training sets and batch sizes.
Similar results are shown in Table 6 for the Inception-v4 model except that it was trained only at batch sizes of 16 and 32. The Inception-v4 model with batch size of 64 could not be trained because of the limited memory size (11 GB) of our graphic processing units.
In addition to AUC, we also estimated the corresponding accuracy, sensitivity, and specificity at a cut-off point on the ROC curve determined by the maximum Youden Index for all the conditions reported in Table 5 and Table 6. These results are included in the Appendix A in Table A1, Table A2, Table A3 and Table A4.

4. Discussion

Deep-learning-based models have been proposed for detection of COVID-19 [41]. However, few studies were conducted for classification of severe versus non-severe lung infection on chest radiographs of COVID-19 patients using deep learning. Chest radiography is low cost and easily accessible even in resource-limited regions, especially that sequential imaging, including a reference baseline exam, is often needed for surveillance of infection progression or regression in many COVID patients. A deep learning model that can assist physicians in consistently identifying severe lung infection in COVID-19 patients would be useful for treatment management, especially in pandemic situations.
We demonstrated that it is possible to train the Inception-v1 and Inception-v4 models using the limited data sets with proper severity labels. For Inception-v1, the small MIDRC training set appeared to be sufficient and the addition of the other training sets did not substantially improve its performance. On the other hand, for Inception-v4, the model trained with the small MIDRC set was unstable likely because Inception-v4 had a much larger number of weights to be trained than Inception-v1. The larger BrixIA training set was needed to stabilize the training. The Inception-v4 model could achieve better performance than Inception-v1 when it was trained with the larger training set. This is consistent with the expectation that the larger Inception-v4 model has larger learning capacity but requires larger training sample size to learn properly.
The training sets for the deep learning models in this study were obtained from three public data sets, MIDRC, BrixIA, and COVIDGR, which were collected from different populations and different imaging equipment. The BrixIA and the COVIDGR sets were used in combination with the MIDRC set, in comparison to the MIDRC set alone. The independent test results in Table 5, Table 6, Table A1, Table A2, Table A3 and Table A4 showed a similar trend that the test performance of the trained model increased slightly when the training set size increased. For a given test set, the AUC increased when the training samples from the same population as the test set were included in the combined training set. The COVIDGR set appeared to be somewhat distinct from the other two training sets such that adding the COVIDGR set would change the test performance for the MIDRC and BrixIA test sets in both directions. One possible cause is that the COVIDGR set was provided in JPG format, which might change the image quality compared to the images in DICOM format.
Because of the limited size of each data set, we allocated only a small portion of the data set for validation. As can be seen from the validation curves, the AUC performance was quite different from the final test results. However, the small validation sets served adequately for the purpose of monitoring the convergence of the training procedures, indicating whether the training reached stability and eventually converged to a plateau region under the various training conditions. The test results were much more consistent among the different training conditions and test sets.
The UMICH test set was collected from a different institution than the training or validation sets. It served as an “external validation set” in this study. The results indicated that the trained deep learning models can be generalized to an unseen dataset from a different population.
It is difficult to make a direct comparison between the performance of our models and the previous studies (Table 1). The performance of a model depends strongly on the characteristics of the data set and the reference standard. The previous studies used different data sets than those in our study. It is also unclear if any of the studies that used CXRs as input targeted the same task, i.e., classifying severe vs. non-severe lung infection in COVID-19 patients since the definition of the severity levels in the different studies appeared very different. Ref. [23] might be the closest but it did not report AUC; the accuracy, sensitivity, and specificity were comparable to our values. Regardless of the details of the studies, most of the reported AUCs were lower than those achieved by our models.
There are limitations in this study. The sizes of all four data sets were small. One of the challenges is that most of the publicly available data sets did not use the same severity ratings defined by TCIA, probably because they require experienced chest radiologists’ reading. The MIDRC data set has been enlarged substantially since the early set that we used for this study. However, different methods are used to assess the severity of COVID-19 infection for the newly added chest radiographs so that the labels are inconsistent with the early set and cannot be combined together for our purpose. This therefore reveals an important issue in the collection of large database such as MIDRC; it is preferable to plan from the beginning the methods and standards for collecting the labels and annotations required for the various applications of the data so that they can be more uniform for the entire data set. The changes in labeling or associated information from time to time would fragment the database into smaller subsets that contain consistent information for a given study, partly defeating the purpose of collecting large public database.
A second limitation is that the image characteristics of the public data sets were heterogeneous. The MIDRC set included images acquired with CR, for which the image quality and the processing methods were very different from images acquired with flat-panel digital detectors. The images in the COVIDGR set were stored in JPG format, different from the DICOM standard format for medical images. The heterogeneity of the image characteristics might have reduced the effectiveness of combining the training sets together to increase the training set sizes. As seen in Table 5 and Table 6, there were only relatively small improvements in the test AUCs when the BrixIA training set, which was several times larger, was added to the MIDRC training set. On the other hand, there may be an advantage that the model trained with such heterogeneous data may be more robust when it is deployed to local patient images of different quality, also apparent from the tables showing relatively consistent performances across the different independent test sets from multi-institutions and different countries.
A third limitation is that we did not compare many different deep learning structures, including the more recent visual transformer approach. However, the purpose of this study is to evaluate the feasibility of training a robust model despite the limited sample size and heterogeneous quality of the available data, which would likely be the more important factors that determined the model performance than the deep learning architectures.

5. Conclusions

In this study we developed a deep learning pipeline to differentiate severe from non-severe lung infection in COVID-19 patients using chest radiographs. A U-Net model was developed to segment the lungs. Inception-v1 and Inception-v4 models were trained for the classification of severe vs. non-severe lung infection. Multi-institutional datasets were used to develop and evaluate the models. Different combinations of training and validation data sets were used to study the reproducibility of classification performance. The generalizability of the trained deep learning models was validated with multi-institutional test sets and an external test set. The Inception-v1 models achieved AUCs ranging from 0.81 ± 0.02 to 0.84 ± 0.0, while the Inception-v4 models achieved AUCs in the range between 0.85 ± 0.06 and 0.89 ± 0.01 on the independent test sets. Our study shows promise in differentiating COVID-19 patients with severe from non-severe lung infection on chest radiographs. Future work should improve the performance of the deep learning models by increasing the training sample sizes, if available. COVID-19 is evolving over time and vaccination also changes the course of the viral infection manifestation and symptoms. It is hoped that these early works can serve as a foundation for continued development of updated image analysis tools that can assist in early and efficient detection of the disease and thus facilitate proper treatment decision.

Author Contributions

Conceptualization, H.-P.C., L.M.H., R.K.S., J.S., P.P.A. and A.S.; methodology, L.M.H., A.S., H.-P.C. and R.K.S.; software, A.S., R.K.S. and L.M.H.; validation, A.S., L.M.H. and H.-P.C.; formal analysis, A.S., L.M.H. and H.-P.C.; investigation, A.S., L.M.H., H.-P.C. and R.K.S.; resources, H.-P.C. and L.M.H.; data curation, A.S., H.-P.C. and L.M.H.; writing—original draft preparation, A.S.; writing—review and editing, A.S., L.M.H., H.-P.C., R.K.S., P.P.A., C.Z. and J.S.; visualization, A.S.; supervision, L.M.H. and H.-P.C.; project administration, L.M.H. All authors have read and agreed to the published version of the manuscript.

Funding

A.S., H.-P.C., R.K.S., P.P.A., C.Z. and J.S. received no external funding. L.M.H. was partially funded through The Medical Imaging Data Resource Center (MIDRC), which is funded by the National Institute of Biomedical Imaging and Bioengineering (NIBIB) of the National Institutes of Health under Subcontract No. AWD101462-T (Contract No. 75N92020D00021).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of University of Michigan (HUM00186640, 9 September 2020).

Informed Consent Statement

Patient consent was waived due to the designation of this study as a Health Insurance Portability and Accountability Act (HIPAA)-compliant retrospective cohort study.

Data Availability Statement

Data available upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

We also have estimated the accuracy, sensitivity, and specificity for all the experiments in Table 5 and Table 6. The results are presented in Table A1, Table A2, Table A3 and Table A4 below. The accuracy, sensitivity, and specificity were estimated at a cut-off point on the ROC curve determined by the maximum Youden Index.
Table A1. Accuracy values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v1 models trained with different training sets and batch sizes of 16, 32, and 64. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Table A1. Accuracy values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v1 models trained with different training sets and batch sizes of 16, 32, and 64. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Batch SizeModelsMIDRC BrixIA COVIDGR UMICH
Accuracy
16M0.76 ± 0.010.74 ± 0.010.74 ± 0.040.73 ± 0.01
MB0.77 ± 0.010.77 ± 0.020.72 ± 0.030.75 ± 0.02
MBC0.77 ± 0.020.77 ± 0.010.76 ± 0.060.73 ± 0.02
32M0.75 ± 0.020.75 ± 0.020.71 ± 0.050.73 ± 0.01
MB0.75 ± 0.030.77 ± 0.020.76 ± 0.060.75 ± 0.02
MBC0.78 ± 0.010.78 ± 0.010.76 ± 0.040.75 ± 0.02
64M0.74 ± 0.040.73 ± 0.030.66 ± 0.070.68 ± 0.04
MB0.77 ± 0.030.78 ± 0.020.76 ± 0.100.74 ± 0.02
MBC0.75 ± 0.030.75 ± 0.010.74 ± 0.030.73 ± 0.02
Table A2. Accuracy values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v4 models trained with different training sets and batch sizes of 16 and 32. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Table A2. Accuracy values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v4 models trained with different training sets and batch sizes of 16 and 32. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Batch SizeModelsMIDRCBrixIACOVIDGRUMICH
Accuracy
16M0.77 ± 0.050.75 ± 0.080.82 ± 0.050.77 ± 0.06
MB0.80 ± 0.010.81 ± 0.020.74 ± 0.050.84 ± 0.01
MBC0.80 ± 0.020.79 ± 0.020.81 ± 0.050.83 ± 0.02
32M0.77 ± 0.030.78 ± 0.020.71 ± 0.070.80 ± 0.02
MB0.81 ± 0.020.79 ± 0.020.71 ± 0.020.83 ± 0.01
MBC0.80 ± 0.010.79 ± 0.010.80 ± 0.070.83 ± 0.03
Table A3. Sensitivity and specificity values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v1 models trained with different training sets and batch sizes of 16, 32, and 64. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Table A3. Sensitivity and specificity values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v1 models trained with different training sets and batch sizes of 16, 32, and 64. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Batch SizeModelsPerformance MetricMIDRC BrixIA COVIDGR UMICH
16MSensitivity0.75 ± 0.06 0.77 ± 0.02 0.77 ± 0.04 0.77 ± 0.05
Specificity0.77 ± 0.050.73 ± 0.020.74 ± 0.040.72 ± 0.03
MBSensitivity0.77 ± 0.040.76 ± 0.040.75 ± 0.11 0.71 ± 0.02
Specificity0.78 ± 0.040.77 ± 0.030.71 ± 0.060.77 ± 0.02
MBCSensitivity0.78 ± 0.030.79 ± 0.01 0.76 ± 0.09 0.77 ± 0.03
Specificity0.76 ± 0.050.77 ± 0.010.76 ± 0.080.71 ± 0.03
32MSensitivity0.78 ± 0.030.75 ± 0.050.68 ± 0.06 0.71 ± 0.05
Specificity0.73 ± 0.030.75 ± 0.040.72 ± 0.070.73 ± 0.02
MBSensitivity0.76 ± 0.060.76 ± 0.040.69 ± 0.12 0.71 ± 0.02
Specificity0.74 ± 0.080.78 ± 0.040.78 ± 0.090.77 ± 0.02
MBCSensitivity 0.79 ± 0.04 0.76 ± 0.020.84 ± 0.04 0.73 ± 0.05
Specificity0.77 ± 0.030.78 ± 0.020.74 ± 0.060.76 ± 0.05
64MSensitivity 0.75 ± 0.04 0.66 ± 0.050.67 ± 0.11 0.66 ± 0.08
Specificity0.73 ± 0.040.75 ± 0.050.66 ± 0.070.69 ± 0.06
MBSensitivity 0.78 ± 0.03 0.75 ± 0.04 0.67 ± 0.11 0.69 ± 0.05
Specificity0.77 ± 0.040.79 ± 0.030.79 ± 0.150.77 ± 0.03
MBCSensitivity 0.77 ± 0.030.78 ± 0.02 0.69 ± 0.08 0.69 ± 0.02
Specificity0.74 ± 0.060.74 ± 0.020.76 ± 0.040.75 ± 0.02
Table A4. Sensitivity and specificity values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v4 models trained with different training sets and batch sizes of 16 and 32. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Table A4. Sensitivity and specificity values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v4 models trained with different training sets and batch sizes of 16 and 32. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Batch SizeModelsPerformance MetricMIDRC BrixIA COVIDGR UMICH
16MSensitivity 0.75 ± 0.11 0.71 ± 0.15 0.80 ± 0.16 0.77 ± 0.12
Specificity0.79 ± 0.040.76 ± 0.060.82 ± 0.040.77 ± 0.04
MBSensitivity 0.82 ± 0.04 0.78 ± 0.03 0.72 ± 0.09 0.82 ± 0.04
Specificity0.78 ± 0.040.81 ± 0.030.75 ± 0.070.85 ± 0.03
MBCSensitivity 0.82 ± 0.05 0.78 ± 0.01 0.79 ± 0.03 0.81 ± 0.05
Specificity0.79 ± 0.040.79 ± 0.020.82 ± 0.060.84 ± 0.03
32MSensitivity 0.77 ± 0.04 0.78 ± 0.04 0.69 ± 0.08 0.84 ± 0.02
Specificity0.77 ± 0.050.78 ± 0.040.71 ± 0.080.78 ± 0.02
MBSensitivity 0.81 ± 0.02 0.79 ± 0.02 0.75 ± 0.06 0.84 ± 0.05
Specificity0.80 ± 0.030.79 ± 0.030.70 ± 0.030.83 ± 0.03
MBCSensitivity 0.78 ± 0.04 0.81 ± 0.03 0.85 ± 0.09 0.83 ± 0.03
Specificity0.81 ± 0.030.79 ± 0.020.79 ± 0.100.83 ± 0.05

References

  1. Laino, M.E.; Ammirabile, A.; Posa, A.; Cancian, P.; Shalaby, S.; Savevski, V.; Neri, E. The Applications of Artificial Intelligence in Chest Imaging of COVID-19 Patients: A Literature Review. Diagnostics 2021, 11, 1317. [Google Scholar] [CrossRef] [PubMed]
  2. Nair, A.; Procter, A.; Halligan, S.; Parry, T.; Ahmed, A.; Duncan, M.; Taylor, M.; Chouhan, M.; Gaunt, T.; Roberts, J.; et al. Chest radiograph classification and severity of suspected COVID-19 by different radiologist groups and attending clinicians: Multi-reader, multi-case study. Eur. Radiol. 2023, 33, 2096–2104. [Google Scholar] [CrossRef] [PubMed]
  3. Way, T.; Chan, H.-P.; Hadjiiski, L.; Sahiner, B.; Chughtai, A.; Song, T.K.; Poopat, C.; Stojanovska, J.; Frank, L.; Attili, A.; et al. Computer-aided diagnosis of lung nodules on CT scans: ROC study of its effect on radiologists’ performance. Acad. Radiol. 2010, 17, 323–332. [Google Scholar] [CrossRef] [PubMed]
  4. Sun, D.; Hadjiiski, L.; Alva, A.; Zakharia, Y.; Joshi, M.; Chan, H.P.; Garje, R.; Pomerantz, L.; Elhag, D.; Cohan, R.H.; et al. Computerized Decision Support for Bladder Cancer Treatment Response Assessment in CT Urography: Effect on Diagnostic Accuracy in Multi-Institution Multi-Specialty Study. Tomography 2022, 8, 644–656. [Google Scholar] [CrossRef]
  5. Hadjiiski, L.M.; Chan, H.-P.; Sahiner, B.; Helvie, M.A.; Roubidoux, M.; Blane, C.; Paramagul, C.; Petrick, N.; Bailey, J.; Klein, K.; et al. Improvement of Radiologists’ Characterization of Malignant and Benign Breast Masses in Serial Mammograms by Computer-Aided Diagnosis: An ROC Study. Radiology 2004, 233, 255–265. [Google Scholar] [CrossRef]
  6. Litmanovich, D.E.; Chung, M.; Kirkbride, R.R.; Kicska, G.; Kanne, J.P. Review of Chest Radiograph Findings of COVID-19 Pneumonia and Suggested Reporting Language. J. Thorac. Imaging 2020, 35, 354–360. [Google Scholar] [CrossRef]
  7. Tsai, E.; Simpson, S.; Lungren, M.P.; Hershman, M.; Roshkovan, L.; Colak, E.; Erickson, B.J.; Shih, G.; Stein, A.; Kalpathy-Cramer, J.; et al. Data from Medical Imaging Data Resource Center (MIDRC)—RSNA International COVID Radiology Database (RICORD) Release 1c—Chest X-ray, COVID+ (MIDRC-RICORD-1C). The Cancer Imaging Archive. 2021. Available online: https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=70230281 (accessed on 1 October 2023).
  8. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and Operating a Public Information Repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef]
  9. Feyaerts, D.; Hedou, J.; Gillard, J.; Chen, H.; Tsai, E.S.; Peterson, L.S.; Ando, K.; Manohar, M.; Do, E.; Dhondalay, G.K.R.; et al. Integrated plasma proteomic and single-cell immune signaling network signatures demarcate mild, moderate, and severe COVID-19. Cell Rep. Med. 2022, 3, 100680. [Google Scholar] [CrossRef]
  10. Danilov, V.V.; Litmanovich, D.; Proutski, A.; Kirpich, A.; Nefaridze, D.; Karpovsky, A.; Gankin, Y. Automatic scoring of COVID-19 severity in X-ray imaging based on a novel deep learning workflow. Sci. Rep. 2022, 12, 12791. [Google Scholar] [CrossRef]
  11. Frid-Adar, M.; Amer, R.; Gozes, O.; Nassar, J.; Greenspan, H. COVID-19 in CXR: From Detection and Severity Scoring to Patient Disease Monitoring. IEEE J. Biomed. Health Inform. 2021, 25, 1892–1903. [Google Scholar] [CrossRef] [PubMed]
  12. Namburu, A.; Sumathi, D.; Raut, R.; Jhaveri, R.H.; Dhanaraj, R.K.; Subbulakshmi, N.; Balusamy, B. FPGA-Based Deep Learning Models for Analysing Corona Using Chest X-ray Images. Mob. Inf. Syst. 2022, 2022, 2110785. [Google Scholar] [CrossRef]
  13. Ren, K.; Hong, G.; Chen, X.; Wang, Z. A COVID-19 medical image classification algorithm based on Transformer. Sci. Rep. 2023, 13, 5359. [Google Scholar] [CrossRef] [PubMed]
  14. Hadi, M.U.; Qureshi, R.; Ahmed, A.; Iftikhar, N. A lightweight CORONA-NET for COVID-19 detection in X-ray images. Expert Syst. Appl. 2023, 225, 120023. [Google Scholar] [CrossRef] [PubMed]
  15. Constantinou, M.; Exarchos, T.; Vrahatis, A.G.; Vlamos, P. COVID-19 Classification on Chest X-ray Images Using Deep Learning Methods. Int. J. Environ. Res. Public Health 2023, 20, 2035. [Google Scholar] [CrossRef] [PubMed]
  16. Ullah, Z.; Usman, M.; Latif, S.; Gwak, J. Densely attention mechanism based network for COVID-19 detection in chest X-rays. Sci. Rep. 2023, 13, 261. [Google Scholar] [CrossRef]
  17. Samala, R.K.; Hadjiiski, L.; Chan, H.P.; Zhou, C.; Stojanovska, J.; Agarwal, P.; Fung, C. Severity assessment of COVID-19 using imaging descriptors: A deep-learning transfer learning approach from non-COVID-19 pneumonia. In Medical Imaging 2021: Computer-Aided Diagnosis; SPIE: Bellingham, WA, USA, 2021; Volume 11597. [Google Scholar] [CrossRef]
  18. Park, S.; Kim, G.; Oh, Y.; Seo, J.B.; Lee, S.M.; Kim, J.H.; Moon, S.; Lim, J.-K.; Ye, J.C. Multi-task vision transformer using low-level chest X-ray feature corpus for COVID-19 diagnosis and severity quantification. Med. Image Anal. 2022, 75, 102299. [Google Scholar] [CrossRef]
  19. Sahoo, P.; Saha, S.; Sharma, S.K.; Mondal, S.; Gowda, S. A Multi-stage framework for COVID-19 detection and severity assessment from chest radiography images using advanced fuzzy ensemble technique (R). Expert Syst. Appl. 2024, 238, 121724. [Google Scholar] [CrossRef]
  20. Bergantini, L.; d’Alessandro, M.; Cameli, P.; Otranto, A.; Luzzi, S.; Bianchi, F.; Bargagli, E. Cytokine profiles in the detection of severe lung involvement in hospitalized patients with COVID-19: The IL-8/IL-32 axis. Cytokine 2022, 151, 155804. [Google Scholar] [CrossRef]
  21. Huyut, M.T. Automatic Detection of Severely and Mildly Infected COVID-19 Patients with Supervised Machine Learning Models. IRBM 2023, 44, 100725. [Google Scholar] [CrossRef]
  22. Al Rahhal, M.M.; Bazi, Y.; Jomaa, R.M.; AlShibli, A.; Alajlan, N.; Mekhalfi, M.L.; Melgani, F. COVID-19 Detection in CT/X-ray Imagery Using Vision Transformers. J. Pers. Med. 2022, 12, 310. [Google Scholar] [CrossRef]
  23. Tabik, S.; Gomez-Rios, A.; Martin-Rodriguez, J.L.; Sevillano-Garcia, I.; Rey-Area, M.; Charte, D.; Guirado, E.; Suarez, J.L.; Luengo, J.; Valero-Gonzalez, M.A.; et al. COVIDGR Dataset and COVID-SDNet Methodology for Predicting COVID-19 Based on Chest X-ray Images. IEEE J. Biomed. Health Inform. 2020, 24, 3595–3605. [Google Scholar] [CrossRef]
  24. PadChest. Available online: https://bimcv.cipf.es/bimcv-projects/padchest/ (accessed on 2 October 2023).
  25. Stony Brook University COVID-19 Positive Cases (COVID-19-NY-SBU). Available online: https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=89096912 (accessed on 2 October 2023).
  26. Chest Imaging with Clinical and Genomic Correlates Representing a Rural COVID-19 Positive Population (COVID-19-AR). Available online: https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=70226443 (accessed on 2 October 2023).
  27. Cohen, J.P.; Morrison, P.; Dao, L.; Roth, K.; Duong, T.Q.; Ghassemi, M. COVID-19 Image Data Collection: Prospective Predictions Are the Future. arXiv 2020, arXiv:2006.11988. Available online: https://github.com/ieee8023/covid-chestxray-dataset (accessed on 1 June 2023). [CrossRef]
  28. Borghesi, A.; Maroldi, R. COVID-19 outbreak in Italy: Experimental chest X-ray scoring system for quantifying and monitoring disease progression. Radiol. Medica 2020, 125, 509–513. [Google Scholar] [CrossRef]
  29. Signoroni, A.; Savardi, M.; Benini, S.; Adami, N.; Leonardi, R.; Gibellini, P.; Vaccher, F.; Ravanelli, M.; Borghesi, A.; Maroldi, R.; et al. BS-Net: Learning COVID-19 pneumonia severity on a large chest X-ray dataset. Med. Image Anal. 2021, 71, 102046. [Google Scholar] [CrossRef]
  30. BrixIA Home Page, #BrixIA COVID19 Severity Score Assessment Project and Database. Available online: https://brixia.github.io/ (accessed on 1 October 2023).
  31. COVIDGR Home Page, GitHub—Ari-Dasci/OD-Covidgr: Datasets of X-ray Imaging for Detection of COVID-19. Available online: https://github.com/ari-dasci/OD-covidgr/ (accessed on 1 October 2023).
  32. MIDRC—Medical Imaging and Data Resource Center. Available online: https://www.midrc.org/ (accessed on 2 October 2023).
  33. Murphy, M.C.; Little, B.P. Chronic Pulmonary Manifestations of COVID-19 Infection: Imaging Evaluation. Radiology 2023, 307, e222379. [Google Scholar] [CrossRef] [PubMed]
  34. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the 18th International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  35. Gielczyk, A.; Marciniak, A.; Tarczewska, M.; Lutowski, Z. Pre-processing methods in chest X-ray image classification. PLoS ONE 2022, 17, e0265949. [Google Scholar] [CrossRef] [PubMed]
  36. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  37. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the 31st AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 4278–4284. [Google Scholar]
  38. Samala, R.K.; Chan, H.P.; Hadjiiski, L.; Helvie, M.A.; Richter, C.D.; Cha, K.H. Breast Cancer Diagnosis in Digital Breast Tomosynthesis: Effects of Training Sample Size on Multi-Stage Transfer Learning Using Deep Neural Nets. IEEE Trans. Med. Imaging 2019, 38, 686–696. [Google Scholar] [CrossRef] [PubMed]
  39. Samala, R.K.; Chan, H.P.; Hadjiiski, L.M.; Helvie, M.A.; Richter, C.D. Generalization error analysis for deep convolutional neural network with transfer learning in breast cancer diagnosis. Phys. Med. Biol. 2020, 65, 105002. [Google Scholar] [CrossRef] [PubMed]
  40. Samala, R.K.; Chan, H.P.; Hadjiiski, L.; Helvie, M.A. Risks of feature leakage and sample size dependencies in deep feature extraction for breast mass classification. Med. Phys. 2021, 48, 2827–2837. [Google Scholar] [CrossRef] [PubMed]
  41. Apostolopoulos, I.D.; Mpesiana, T.A. Covid-19: Automatic detection from X-ray images utilizing transfer learning with convolutional neural networks. Phys. Eng. Sci. Med. 2020, 43, 635–640. [Google Scholar] [CrossRef]
Figure 1. Different degrees of COVID severity.
Figure 1. Different degrees of COVID severity.
Diagnostics 14 00341 g001
Figure 2. Processing pipeline for classification of severe versus non-severe COVID-19 patients using chest radiographs.
Figure 2. Processing pipeline for classification of severe versus non-severe COVID-19 patients using chest radiographs.
Diagnostics 14 00341 g002
Figure 3. Lung Segmentation by U-Net: (a) Input image, (b) manually outlined lungs—the target output, (c) U-Net prediction and (d) final lung regions.
Figure 3. Lung Segmentation by U-Net: (a) Input image, (b) manually outlined lungs—the target output, (c) U-Net prediction and (d) final lung regions.
Diagnostics 14 00341 g003
Figure 4. Original images are presented on the top row. Bottom images are harmonized by shifting the mean and the standard deviation of the pixel intensities within the lung regions. The background pixel intensity outside the lung region was set to a constant value of zero for all images.
Figure 4. Original images are presented on the top row. Bottom images are harmonized by shifting the mean and the standard deviation of the pixel intensities within the lung regions. The background pixel intensity outside the lung region was set to a constant value of zero for all images.
Diagnostics 14 00341 g004
Figure 5. Images in the top row are input and images in the bottom row are segmentation results. Color labels: Black—background, white—labeled lung, red—false negative, and yellow—false positive.
Figure 5. Images in the top row are input and images in the bottom row are segmentation results. Color labels: Black—background, white—labeled lung, red—false negative, and yellow—false positive.
Diagnostics 14 00341 g005
Figure 6. AUC-versus-iteration performance of Inception-v1 models trained with different training sets: Model M (blue line), Model MB (orange line), and Model MBC (purple line). The trained models were deployed on different validation sets: (a) MIDRC validation, (b) BrixIA validation, and (c) COVIDGR validation. The horizontal axis was plotted as the number of iterations because the number of images in each training set was very different.
Figure 6. AUC-versus-iteration performance of Inception-v1 models trained with different training sets: Model M (blue line), Model MB (orange line), and Model MBC (purple line). The trained models were deployed on different validation sets: (a) MIDRC validation, (b) BrixIA validation, and (c) COVIDGR validation. The horizontal axis was plotted as the number of iterations because the number of images in each training set was very different.
Diagnostics 14 00341 g006
Figure 7. AUC-versus-epoch curves of Inception-v1 models trained on the MIDRC training set using different batch sizes (16, 32, and 64) and deployed on three different validation sets. Each model was trained with five different random initializations.
Figure 7. AUC-versus-epoch curves of Inception-v1 models trained on the MIDRC training set using different batch sizes (16, 32, and 64) and deployed on three different validation sets. Each model was trained with five different random initializations.
Diagnostics 14 00341 g007
Figure 8. AUC-versus-epoch curves of Inception-v1 models trained on the combined MIDRC + BrixIA training set using different batch sizes (16, 32, and 64) and deployed on three different validation sets. Each model was trained with five different random initializations.
Figure 8. AUC-versus-epoch curves of Inception-v1 models trained on the combined MIDRC + BrixIA training set using different batch sizes (16, 32, and 64) and deployed on three different validation sets. Each model was trained with five different random initializations.
Diagnostics 14 00341 g008
Figure 9. Comparison of the AUC-versus-epoch curves of Inception-v1 and Inception-v4 models trained on the MIDRC training set and the combined MIDRC + BrixIA training set and deployed on three different validation sets. Each model was trained with five different random initializations. Batch size = 16.
Figure 9. Comparison of the AUC-versus-epoch curves of Inception-v1 and Inception-v4 models trained on the MIDRC training set and the combined MIDRC + BrixIA training set and deployed on three different validation sets. Each model was trained with five different random initializations. Batch size = 16.
Diagnostics 14 00341 g009
Table 1. Related work on classification of COVID-19 severity or from other lung infections.
Table 1. Related work on classification of COVID-19 severity or from other lung infections.
MethodYearDataClass GroupsResults
Statistical approach [9]2021Plasma and single-cell proteomicMild, moderate, severeAUCtraining = 0.799
AUCvalidation = 0.773
Deep Learning [10]2022CXRCOVID-19 and normalMean absolute error of 0.30
Deep Learning [11]2021CXRSeverity level (scores from 0 to 8)confusion matrix:
Sensitivity = 0.94
Specificity = 0.98
Deep learning [12]
ResNet
2022CXRCOVID-19, virus, bacteria, ARDS, SARS, Streptococcus, normalAccuracy = 98%
Deep learning [13]2023CXRCOVID-19, virus, bacteria, and normalAccuracy = 97.65%
Deep learning [14] CORONA-NET2023CXRCOVID-19, viral pneumonia, normal Accuracy = 99.57%
Deep learning [15]
ResNet
2023CXRCOVID-19, non-COVID-19, and normalAccuracy = 96%
Deep learning [16] DAM-Net2023CXRCOVID-19, pneumonia, and normalAccuracy = 97.22%
Sensitivity = 96.87%
Specificity = 99.12%
Deep learning [17]2021CXRSeverity descriptors vs. radiologist’s severity ratings,
normal vs. abnormal
Correlation = 0.68 (p < 0.0001)
AUC = 0.78
Cytokine profiles with statistical approaches [20]2022Cytokine concentrationSevere, moderate, and mildAUC = 0.83
Supervised machine learning models [21]2023Feature-dataset consisting of the routine blood values and demographic data that affect the prognosis of COVID-19Severely and mildlyAUC range: 0.75 to 0.95
Accuracy range: 94.05% to 97.86%
Vision Transformers [22]2022CXR and CTCOVID-19, pneumonia, and normalAccuracy = 94.62%
COVID_SDNet [23]2020CXRSevere, moderate, and mildAccuracy = 81.0% ± 2.9%
Sensitivity = 76.8% ± 6.3%
Specificity = 85.2% ± 5.4%
Multi-task vision [18]2022CXRSeverity degree from 0 to 6Accuracy range: 78.7% to 97.7%
Multi-stage framework [19]2023CXRMild, moderate, severe, and criticalAccuracy = 97.63%
Table 2. Available databases for COVID-19 (public data sets with severity rating labels available for the current study).
Table 2. Available databases for COVID-19 (public data sets with severity rating labels available for the current study).
Name of DatabaseYear of Data CollectionType of LabelsPublic AvailableSeverity COVID19 RateNumber of Images
Cohen et al. [27]2020Defined by other paper [9]:
  • if “survival” is false, it is called “high severity”;
  • if “survival” and “went-ICU” are true, it is termed “moderate severity”;
  • if “survival” is true and “went-ICU” is false, it is named “low severity”
yesyes679 X-ray images from 412 patients
BrixIA [28,29,30]2020The lungs were divided into six regions and each region was scored ranging from 0 to 3. Consensus from 5 radiologistsyesyes4695
COVIDGR [23,31]2020Data were rated by 1 radiologist as negative, mild, moderate, or severeyesyes852
MIDRC [32]2022Data were rated by 3 radiologists as negative, mild, moderate, or severeyesyes823 with severity labels for COVID-19
Table 3. Data collection. Distribution of severe and non-severe images and cases per training, validation, and testing dataset.
Table 3. Data collection. Distribution of severe and non-severe images and cases per training, validation, and testing dataset.
TRAININGVALIDATIONTESTING
TotalSevereNon-SevereTotalSevereNon-SevereTotalSevereNon-Severe
MIDRCimages582129 (22%)453 (78%)6818 (26%)50 (74%)17370 (40%)103 (60%)
cases425 50 26
BrixIAimages3285563 (17%)2722 (83%)47058 (12%)412 (88%)940184 (20%)756 (80%)
cases3206 464 917
COVIDGRimages30056 (19%)244 (81%)427 (17%)35 (83%)8315 (18%)68 (82%)
cases300 42 83
UMICHimages------------25086 (34%)164 (66%)
cases-- -- 235
Table 4. The performance measures for lung segmentation in the MIDRC test set using the U-Net deep learning approach.
Table 4. The performance measures for lung segmentation in the MIDRC test set using the U-Net deep learning approach.
Mean of JaccardMean of Dice CoefficientMean of Hausdorff DistanceMean of Average Euclidean Distance
Mean0.820.9037.7 mm5.4 mm
Standard deviation0.070.0424.6 mm2.7 mm
Table 5. AUC values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v1 models trained with different training sets and batch sizes of 16, 32, and 64. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Table 5. AUC values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v1 models trained with different training sets and batch sizes of 16, 32, and 64. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Batch SizeModelsMIDRC BrixIA COVIDGR UMICH
AUC
16M0.82 ± 0.010.82 ± 0.010.80 ± 0.040.80 ± 0.01
MB0.84 ± 0.010.84 ± 0.010.78 ± 0.020.80 ± 0.02
MBC0.84 ± 0.020.85 ± 0.000.82 ± 0.020.80 ± 0.01
32M0.82 ± 0.020.81 ± 0.010.75 ± 0.040.77 ± 0.01
MB0.82 ± 0.020.84 ± 0.010.76 ± 0.030.80 ± 0.01
MBC0.83 ± 0.020.84 ± 0.000.84 ± 0.010.81 ± 0.02
64M0.79 ± 0.030.76 ± 0.020.62 ± 0.070.71 ± 0.05
MB0.83 ± 0.020.83 ± 0.010.78 ± 0.060.78 ± 0.02
MBC0.82 ± 0.030.83 ± 0.010.77 ± 0.030.78 ± 0.02
Table 6. AUC values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v4 models trained with different training sets and batch sizes of 16 and 32. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Table 6. AUC values for classification of severe and non-severe COVID-19 based on chest radiographs in the held-out independent test sets by the Inception-v4 models trained with different training sets and batch sizes of 16 and 32. The mean and standard deviation were estimated from the models trained with five random initializations for a given training set and batch size.
Batch SizeModelsMIDRC BrixIA COVIDGR UMICH
AUC
16M0.82 ± 0.090.79 ± 0.140.82 ± 0.150.82 ± 0.10
MB0.86 ± 0.010.88 ± 0.010.79 ± 0.040.89 ± 0.01
MBC0.88 ± 0.010.87 ± 0.010.84 ± 0.020.89 ± 0.01
32M0.84 ± 0.030.85 ± 0.010.74 ± 0.050.87 ± 0.02
MB0.88 ± 0.020.88 ± 0.010.79 ± 0.030.89 ± 0.02
MBC0.87 ± 0.010.88 ± 0.010.85 ± 0.060.89 ± 0.01
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sobiecki, A.; Hadjiiski, L.M.; Chan, H.-P.; Samala, R.K.; Zhou, C.; Stojanovska, J.; Agarwal, P.P. Detection of Severe Lung Infection on Chest Radiographs of COVID-19 Patients: Robustness of AI Models across Multi-Institutional Data. Diagnostics 2024, 14, 341. https://doi.org/10.3390/diagnostics14030341

AMA Style

Sobiecki A, Hadjiiski LM, Chan H-P, Samala RK, Zhou C, Stojanovska J, Agarwal PP. Detection of Severe Lung Infection on Chest Radiographs of COVID-19 Patients: Robustness of AI Models across Multi-Institutional Data. Diagnostics. 2024; 14(3):341. https://doi.org/10.3390/diagnostics14030341

Chicago/Turabian Style

Sobiecki, André, Lubomir M. Hadjiiski, Heang-Ping Chan, Ravi K. Samala, Chuan Zhou, Jadranka Stojanovska, and Prachi P. Agarwal. 2024. "Detection of Severe Lung Infection on Chest Radiographs of COVID-19 Patients: Robustness of AI Models across Multi-Institutional Data" Diagnostics 14, no. 3: 341. https://doi.org/10.3390/diagnostics14030341

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop