Next Article in Journal
Investigating the Impact of Two Major Programming Environments on the Accuracy of Deep Learning-Based Glioma Detection from MRI Images
Next Article in Special Issue
Distinctions between Choroidal Neovascularization and Age Macular Degeneration in Ocular Disease Predictions via Multi-Size Kernels ξcho-Weighted Median Patterns
Previous Article in Journal
The Role of Multimodality Imaging Approach in Acute Aortic Syndromes: Diagnosis, Complications, and Clinical Management
Previous Article in Special Issue
Brain Stroke Classification via Machine Learning Algorithms Trained with a Linearized Scattering Operator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intracranial Hemorrhage Detection Using Parallel Deep Convolutional Models and Boosting Mechanism

1
Department of Computer Science, COMSATS University Islamabad, Islamabad 44000, Pakistan
2
School of Electrical Engineering and Computer Science (SEECS), National University of Sciences and Technology (NUST), Islamabad 44500, Pakistan
3
Institute of Computing, Kohat University of Science & Technology, Kohat 26000, Pakistan
4
Computer Science Department, National University of Computer and Emerging Sciences (NUCES-FAST), Islamabad 44000, Pakistan
5
Industrial Engineering Department, College of Engineering, King Saud University, Riyadh 11421, Saudi Arabia
6
Centre for Smart Systems, AI and Cybersecurity, Staffordshire University, Stoke-on-Trent ST4 2DE, UK
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(4), 652; https://doi.org/10.3390/diagnostics13040652
Submission received: 21 December 2022 / Revised: 31 January 2023 / Accepted: 2 February 2023 / Published: 9 February 2023

Abstract

:
Intracranial hemorrhage (ICH) can lead to death or disability, which requires immediate action from radiologists. Due to the heavy workload, less experienced staff, and the complexity of subtle hemorrhages, a more intelligent and automated system is necessary to detect ICH. In literature, many artificial-intelligence-based methods are proposed. However, they are less accurate for ICH detection and subtype classification. Therefore, in this paper, we present a new methodology to improve the detection and subtype classification of ICH based on two parallel paths and a boosting technique. The first path employs the architecture of ResNet101-V2 to extract potential features from windowed slices, whereas Inception-V4 captures significant spatial information in the second path. Afterwards, the detection and subtype classification of ICH is performed by the light gradient boosting machine (LGBM) using the outputs of ResNet101-V2 and Inception-V4. Thus, the combined solution, known as ResNet101-V2, Inception-V4, and LGBM (Res-Inc-LGBM), is trained and tested over the brain computed tomography (CT) scans of CQ500 and Radiological Society of North America (RSNA) datasets. The experimental results state that the proposed solution efficiently obtains 97.7% accuracy, 96.5% sensitivity, and 97.4% F1 score using the RSNA dataset. Moreover, the proposed Res-Inc-LGBM outperforms the standard benchmarks for the detection and subtype classification of ICH regarding the accuracy, sensitivity, and F1 score. The results prove the significance of the proposed solution for its real-time application.

1. Introduction

Intracranial hemorrhage (ICH) occurs within the cranium due to a traumatic brain injury, tumor, stress, vascular abnormality, arteriovenous malformations, and smoking [1,2,3]. One of the major concerns of ICH is the high death rate of about 35% to 52% in the first 30 days [4,5]. Other concerns such as disability, epilepsy, vascular issues, blood clotting, loss of memory, and vision are also faced by the survivors [6,7]. Therefore, a rapid and accurate mechanism is required to give medical treatment at the initial level to reduce the high mortality rate of ICH. Primarily, computed tomography (CT) scans are used by radiologists to locate the ICH region and type manually. These scans are based on adjacent slices examined by the radiologist to identify the region and pattern of hemorrhage [8]. The ICH region can be intra-axial or extra-axial. The intra-axial represents the bleeding inside brain tissues, whereas the bleeding inside of cranial but outside of the brain tissues is known as extra-axial. The first region covers two types: intraparenchymal hemorrhage (IPH) and intraventricular hemorrhage (IVH), whereas subarachnoid hemorrhage (SAH), subdural hemorrhage (SDH) and epidural hemorrhage (EDH) are covered by the second region [9,10].
The primary way of ICH detection and subtype classification is to physically analyze the CT images of the human head by radiologists. However, due to subtle complexities, overloaded burden, and the time-consuming task of ICH detection, the expertise of radiologists is required to rapidly and accurately detect the ICH region and its cause [9,11]. Another issue that leads to a high death rate is the unavailability of expert radiologists at an emergency time in many developing countries [11]. Therefore, the junior radiologists analyze the CT scans and detect the cause and region of ICH. Although there is a difference in the detection of ICH by expert radiologists than the juniors, which affects on the treatment and medication [9,12]. Therefore, to deal with the large dataset, streamline workflow, timely availability of the detection system, and improve the ICH detection accuracy, an artificial-intelligence-based automated mechanism is required [7,12,13]. Recently, many studies have been proposed to automate ICH detection and subtype characterization, classified into machine learning and deep-learning-based solutions.
Machine-learning-based solutions for identifying ICH have risen significantly in the past few years. ICH has different types in which subarachnoid hemorrhage (SAH) is one of them. The authors in [14,15] propose a decision tree (DT) and random forest 50 (RF)-based model using the clinical data of patients to detect the SAH.
Likewise, Ramos et al. [16] employ four efficient machine learning models, known as DT, RF, support vector machine (SVM), and multilayer perceptron (MLP), to improve the detection of SAH using the laboratory dataset of patients. Another study [17] proposes a hybrid of deep learning and machine learning model where a deep convolutional neural network (CNN) is used to extract features and linear SVM to detect ICH using head CT scans.
Raima et al. [11] detect the presence of ICH using SVM. Afterwards, U-Net segmentation is performed to locate and classify the ICH region. The researchers in [18] show the performance comparison of six machine learning models for the spontaneous detection of ICH. The machine above learning algorithms improve ICH detection and subtype classification results. However, the presented solutions require significant manual data preprocessing, parameters optimization, and feature engineering. Furthermore, these solutions become less efficient and lead to overfitting as the size of the dataset increases [8,9,13]. Thus, with the increasing number of cases globally, data size is increasing, which requires a more effective approach, known as deep learning, to overcome the shortcomings.
Deep learning approaches are efficient in handling large datasets and automating the feature extraction from brain CT slices, attracting the research community’s attention. Sage et al. [19] employ two parallel paths based on ResNet50 to capture the potential features from the head CT scans. These features are forwarded as input to the random forest (RF) and SVM for the detection of ICH. The researchers in [20] propose a deep CNN model based on five convolutional layers to extract features and two dense layers for the identification of ICH. Grewal et al. [21] perform the ICH detection using the 3D CT images where a baseline architecture of DenseNet is used along with the recurrent neural network (RNN). Wu et al. [8] employ the deep CNN model, known as EfficientNet-B0, in two parallel paths to obtain both brain tissue and spatial-based features. Afterwards, the ensemble mechanism is applied to the outputs to perform the ICH detection. In [12], Lee et al. execute four deep learning models, VGG16, Inception-ResNetV2, Inception-V3, and ResNet50, where CNN is used as the baseline architecture. The models are combined to form an ensemble mechanism for improving the detection of ICH. Mansour et al. [22] perform the segmentation to identify the affected regions using a segmentation technique known as Graph-Cut.
Furthermore, the Capsule network is used for feature extraction and fuzzy deep neural network to detect and characterize ICH. Another work in [23] demonstrates the importance of brain windows setting to improve the detection of ICH. The authors attach a window estimator module with the deep CNN model to automate the estimation of brain windows setting.
Anupama et al. [24] locate the ICH regions using the Grab-Cut segmentation mechanism. Moreover, a synergic deep learning concept is used for feature extraction and subtype classification of ICH using head CT images. Wang et al. [2] and Ye et al. [25] identify ICH and its subtypes through a hybrid of CNN- and RNN-based models using 3D CT images. The intention behind this is to extract important features using CNN. Then, RNN performs sequential learning and detection of ICH following a similar approach [5,26,27] where a CNN-based architecture is used to capture features from 3D brain CT images. Then, the long short-term memory (LSTM) model uses the high-level features provided by CNN and performs correlation analysis at slices to identify the ICH and its subtypes. In [28], a faster R-CNN-based architecture is proposed to localize the ICH region and its type categorization using brain CT images.
Lee et al. [29] apply a conventional artificial neural network (ANN) model to demonstrate its acceptability over the CNN model regarding detection and subtype classification of ICH. The aforementioned deep learning solutions significantly rise to ICH detection and its subtype classification performance. However, some of the answers are based on 3D CT images, which require extra time and memory for model training due to a high number of parameters [21,25,28]. Furthermore, it is important to capture spatial features from adjacent CT slices in a similar way that have a significant impact on the ICH detection performance, which is not covered by the studies [5,11,21,25,27]. Few of the studies [14,17,19] have employed machine learning techniques such as SVM, DT, and RF for the ICH detection using the CNN-based features, which are less effective for large dataset and prone to overfitting.
To address the limitations mentioned above, we propose a new deep learning and machine learning-based hybrid model for ICH detection and subtype classification. The proposed solution employs two deep learning models, ResNet101-V2 and Inception-V4, in parallel ways to extract the potential features from brain CT slices. The architecture of these solutions consists of CNN. The ResNet101-V2 captures intensity-related elements from windowed CT slices, whereas the spatial features from adjacent pieces are obtained through the Inception-V4. The prime intention behind extracting spatial and intensity features are to mimic radiologists’ real-world diagnostic processes. Afterwards, a more efficient and fast technique, a light gradient boosting machine (LGBM), is applied to the extracted features by windowed and spatial features to detect the ICH and its subtypes. Therefore, the proposed solution is a combined ResNet101-V2, Inception-V4, and LGBM (Res-Inc-LGBM) based model for ICH detection. To evaluate the performance of the proposed Res-Inc-LGBM, extensive experimentation is performed using the dataset of intracranial hemorrhage detection challenge (IHDC) provided by the Radiological Society of North America (RSNA). Moreover, the proposed solution is tested on the CQ500 dataset to analyze its generalization.

2. Materials and Methods

In this section, we present a detailed description of the proposed methodology for detection and subtypes classification of ICH, as shown in Figure 1. The proposed method consists of three major phases: preprocessing, feature extraction, and ICH detection. Thereafter, in the preprocessing phase, noncontrast brain CT images are used as input data. This phase includes techniques such as skull removal and multiwindowing to enhance ICH detection. Afterwards, the feature extraction phase utilizes two excellent deep learning models to capture intensity-based and spatial features from the given CT scans, as shown in Figure 1. Then, an LGBM model detects the ICH and its subtypes in the third phase using the extracted features from the previous phase. In this regard, the proposed method efficiently overcomes the shortcomings of the existing models for ICH detection and subtype classification. The description of each phase is given in the following subsections.

2.1. Data Preprocessing

The proposed mechanism for the identification and characterization of ICH requires the preprocessing of input to smooth the training process and reduce the training time of the model. Therefore, we apply preprocessing techniques such as skull removal, windowing to capture brain and bone tissues, and data augmentation.

2.1.1. Dataset

The proposed solution is trained and evaluated over the brain hemorrhage data of ICH patients, known as IHDC, which was made available in 2019 by the RSNA [2,30,31]. The dataset is collected from three universities, which is further evaluated by about 60 radiologists of RSNA and made available to develop automated solutions for the detection and subtype classification of ICH [2]. The dataset consists of noncontrast head CT slices in digital imaging and communications in medicine (DICOM) annotated with IPH, IVH, SAH, SDH, and EDH. The slices that contain no hemorrhage or more than one are annotated as ‘Any’. The selected dataset for the training and validation of the proposed Res-Inc-LGBM solution includes information on 13,334 patients with different types of brain hemorrhages. The total number of brain CT images containing hemorrhage is 4579, as shown in Figure 2. The general representation of subtypes of ICH is shown in Figure 3. In addition, another dataset is employed to analyze the proposed solution’s generalisation ability, known as CQ500 [32]. It consists of head CT scans collected from different radiological centres, annotated by expert radiologists. Furthermore, detailed information about both datasets is given in Table 1.

2.1.2. Multiwindowing and Adjacent Slicing

Here, we elaborate on the two types of preprocessing, windowing and adjacent slicing. Before that, we apply the skull removal process that improves ICH detection by specifically focusing on the brain region. Moreover, it improves the model’s training without focusing on the skull region. Therefore, Otsu’s morphological method removes the skulls of head CT images [33,34]. Afterwards, we applied the windowing strategy, known as Hounsfield units, to capture the intensity-based features. This strategy improves the contrast level of head CT scans through window width and level to capture the bone, brain, and subdural tissues [2,35].
Moreover, in the actual application, the radiologists also adjust the windows of brain CT slices to improve the contrast for better locating the lesions. The prime intention is to focus on the specific tissues to discover the hidden complexities of ICH [8]. In this paper, we have applied three intensity windows, termed brain window, bone window, and subdural window, to improve the visualization of brain tissues, skull lesions, and soft tissues, as shown in Figure 4. The first window focuses on the brain tissues where window width and level are defined as 80 and 40, respectively. Afterwards, the window width of 380 and window level 40 are set to focus on the bone tissues. The third window, the subdural window, is obtained by setting the window level to 80 and the width to 200, which captures the subdural hematomas. The three intensity windows are combined to form a single image with three channels.
The other type of preprocessing, the adjacent slicing mechanism, helps the proposed model to focus on the adjacent slices for locating the bleeding regions. Furthermore, this mechanism mimics the real-time diagnostic process of radiologists that use adjacent slices to find the subtle bleeding spots and improve the detection process Wu et al. [8,12,36]. Therefore, to capture the spatial features, a slice interpolation, also known as an adjacent slicing mechanism, is utilized to improve the ICH detection process. In this mechanism, two adjacent slices and the centre slice are concatenated to form a single image with three channels. After applying windowing and adjacent slices, an image augmentation procedure is used to enhance the generalization ability of the proposed solution. In addition, it resolves the overfitting issue. The augmentation method includes scaling, flipping, translating, zooming, and shearing range to get different representations. Afterwards, the slices are rescaled to 224 × 244 to match the model’s input representation because the original CT slices are given in the shape of 512 × 512. Image rescaling is important to reduce the proposed model’s training time because of fewer parameters and memory usage.

2.2. Feature Extraction Using ResNet101-V2 and Inception-V4

Feature extraction is important as it reduces the model’s parameters and computation cost, ultimately enhancing the ICH detection and subtype classification performance. Therefore, the windowing and adjacent slicing-based inputs are passed to the two feature extraction techniques, ResNet101-V2 and Inception-V4. These models extract intensity-based and spatial features in a similar way to enhance the identification of ICH and its subtype. To extract the potential windowed features, this paper proposes a new variant of the ResNet model, known as ResNet101-V2. It depends on the baseline architecture of CNN that extracts high-level features from the given input. Many studies demonstrate the importance of deep learning to capture the hidden complexities efficiently and significantly improve the ICH detection performance [37,38]. The authors Wang et al. in [2,8,39,40] employ deep learning feature extractors such as EfficientNet-B0, DenseNet121, Capsule network, and Visual geometry group (VGG), respectively. Although, with the deep architecture, the gradient vanishes during back-propagation and leads to poor performance, resolved through the skip connections in the ResNet architecture. Therefore, this paper utilizes the pre-trained deep learning architecture of ResNet101-V2 that consists of residual blocks and skip links to efficiently obtain the intensity base features from CT scans. Specifically, it is deep with 101 layers that efficiently find the subtle lesions. The deep architecture of ResNet101-V2 has achieved more excellent performance than the shallow levels architecture such as ResNet50 and ResNet30 Rahman et al. [37]. In this regard, the pre-trained architecture of ResNet101-V2 is employed as a feature extractor to enhance the ICH detection performance, given Keras et al. [41,42].
The spatial features are obtained parallel to the intensity features to follow the real-world diagnostic process of ICH detection. The spatial information in the proposed methodology is obtained through the Inception-V4 model that receives the adjacent slices as input. The prime intention behind using the inception-V4 is the identification of hidden bleeding regions at low computation cost and high performance [43]. It has achieved the highest performance in the classification of ImageNet due to efficient training speed, and fewer model parameters [44]. Likewise to the ResNet101-V2, it is based on the CNN architecture and has both residual skip connections and inception blocks to improve the ICH detection rate. Therefore, in this paper, we use an effective pre-trained feature extractor, Inception-V4, as stated Keras et al. [41,43]. The ResNet101-V2 returns the 2048 intensity-based features, whereas 1536 spatial features are obtained through the Inception-V4. Afterwards, the features provided by both feature extractors are concatenated to form a new input for ICH detection and subtype characterization.

2.3. ICH Detection and Subtype Classification

The final detection and five subtype classification of ICH are performed by the LGBM which uses the concatenated windowed and spatial features. The researchers use SVM [11], DT [14], RF [19], and DT-, RF-, SVM-, and MLP-based four machine learning models in [16] to identify and characterize the ICH. However, these machine learning models become expensive for large datasets, leading to overfitting problems. Moreover, these models have slow feature learning mechanisms and poorly detect ICH. Therefore, in this paper, we have used the boosting mechanism known as LGBM, which is light and has fast training. The boosting strategy learns from the mistakes of previous classifiers and optimizes the performance of the upcoming classifier [45]. It significantly enhances the performance of ICH detection and its subtype classification through rapid learning, less memory consumption, and minimization of previous mistakes during the training process.
Moreover, it can efficiently handle a large dataset, which makes it suitable for this study [46]. LGBM follows the leaf wise strategy to solve the hidden complexities compared to other tree-based models, such as DT or extreme gradient boosting (Xgboost) that employ a level-wise approach. Therefore, the leaf wise strategy of LGBM helps find the hidden complexities of ICH and enhance its detection score. LGBM has several important hyperparameters, such as the number of classifiers, depth, and learning rate. These parameters have a significant impact on the ICH detection performance. Thus, the hyperparameters of LGBM are tuned through the grid-search technique.

3. Experimental Results

This section describes the experimental results of the proposed methodology to assess its performance for subtype classification and detection of ICH. The proposed solution is compared with the existing benchmark techniques for ICH detection to demonstrate its effectiveness and suitability for real-time application.
The proposed Res-Inc-LGBM solution is developed over the workstation with specifications of Intel core i7-7200, 16GB RAM, and a 1TB hard drive using the Python language. In addition, the TensorFlow and Keras libraries and the graphics processing unit (GPU) of NVIDIA are employed to form and train the deep learning models. The proposed Res-Inc-LGBM is developed and assessed over the dataset of IHDC, which is mentioned in Section 2.1.1. The dataset is divided into two subsets, training and testing, with a ratio of 70 and 30, respectively. In this study, the pre-trained architecture of ResNet101-V2 and Inception-V4 are employed to capture the windowed and spatial features [45].
The proposed architectures are trained using the adaptive moment estimation (Adam) learning technique and weighted binary cross-entropy (WBC) loss function. The weighted loss function is used to deal with the effects of imbalanced data. The weight value is assigned concerning the class importance, such as 1.0, 0.29, 0.22, 0.15, 0.16, and 0.18 for ‘Any’, EDH, IVH, IPH, SDH, and SAH subtypes. The CT scans are reshaped to 224 × 244 and 299 × 299 for forwarding as an input of ResNet101-V2 and Inception-V4, respectively. The feature extractors use the batch size of 16 and learning rate of 1×10-3 to train over 50 iterations where the performance of models is evident. Afterwards, the extracted features are received as input by LGBM for ICH detection and its subtype categorization, which is mentioned in Section 2.3.

3.1. Experimental Results of Proposed Solution

To assess the performance of the proposed Res-Inc-LGBM algorithm, several classification-based performance indicators are employed in this study. These indicators are obtained through the confusion matrix that provides four outcomes: true positive (TP), false positive (FP), true negative (TN), and false negative (FN).
These four outcomes provide the basis to form several potential performance indicators that are used in this study, such as F1-measure, precision, sensitivity (also known as recall) and TP rate (TPR), specificity, also identified as TN rate (TNR), area under the precision-recall (AUPR), accuracy, and area under the curve (AUC) Sage et al. [19,46].
After the preprocessing of data, feature extraction techniques, Inception-V4, and ResNet101-V2, are applied to obtain important features. The Inception-V4 extracts the important elements from the adjacent CT slices. Specifically, it finds the complex bleeding regions through the adjacent CT slices. Figure 5 shows the performance of the Inception-V4 while obtaining the spatial features. It is seen that the deep architecture of Inception-V4 efficiently learns the subtle complexities and minimizes the loss function.
The Inception-V4 has uniform performance for both training and validation data, such as the loss values of 0.01. During the validation, the model loss sometimes fluctuates due to the complexity of ICH. Afterwards, the accuracy of the Inception-V4 is shown in Figure 6 during the model training. It achieves an accuracy level of 0.85 during the training, showing the proposed solution’s excellency.
The performance of Inception-V4 in terms of AUC and AUPR is shown in Figure 7 and Figure 8 during the model execution, respectively. As it is clear that the model excellently finds the hidden complexities of ICH, which the AUC and PRAU validate. The Inception-V4 feature extractor obtains scores of 0.97 and 0.98 for AUC and PRAU, respectively. These scores are obtained using the training and validation sets. The execution results of Inception-V4 validate that the proposed mechanism has significant potential to enhance ICH detection and its subtype classification.
In the parallel way of spatial features, the intensity-based representations are achieved through another feature extractor, ResNet101-V2. It has the potential to efficiently extract the brain, bone, and subdural tissues based on subtle complexities, as shown in Figure 9. It is depicted that the second model also achieves excellent results for feature extraction on training and validation sets. However, the fluctuations are seen for the validation data due to complexities and similarities between the subtypes of ICH, such as EDH and SDH. Except that, the model efficiently minimizes the loss value and obtains 0.10 for the validation set, as shown in Figure 9. Likewise, the accuracy of the ResNet101-V2 is depicted in Figure 10. The figure shows the accuracy of training and validation sets regarding the extraction of brain, bone, and subdural tissues.
In contrast to the loss value, the accuracy value gradually increases for both training and validation sets. It is evident that ResNet101-V2 efficiently finds insights from CT scans and improves the accuracy level. In addition, to demonstrate the efficiency of the ResNet101-V2 for the extraction of intensity features, the AUC and AUPR scores are also depicted in Figure 11 and Figure 12. It is shown in Figure 11 that ResNet101-V2 smoothly optimizes the AUC for both training and validation sets. The AUC score of 0.98 is obtained by the model for training and unseen dataset. Likewise, Figure 12 highlights the AUPR-based performance over the model training iterations. The figure depicts that the ResNet101-V2 uniformly improves the AUPR score for the training and testing sets regarding the extraction of intensity features.
The AUC- and AUPR-based performance of the ResNet101-V2 validates that the employed feature extractor has significant capability to obtain the hidden bleeding regions from intensity-based CT slices. Furthermore, the AUC-based performance of the proposed solution in terms of the receiver operating characteristics (ROC) curve is shown in Figure 13. The ROC curves depict the TPR versus FPR to demonstrate the ICH detection and its subtype classification ability of the model. The high value of the curve states the excellent detection power of the proposed mechanism. In a similar fashion to AUC, the proposed solution’s performance regarding AUPR is depicted in Figure 14 where precision against recall is demonstrated. The curves in the figure show the performance for each subtype of ICH. In general, the feature extraction results of ResNet101-V2 and Inception-V4 validate that these models efficiently the detection and subtype classification of ICH.
To assess the performance of the proposed Res-Inc-LGBM solution for ICH detection and its subtype categorization, Table 2 presents the complete analysis of the results. The table illustrates the performance results for each of the subtypes of ICH. It is seen that the proposed Res-Inc-LGBM achieves 0.985 AUC, 0.954 sensitivity, 0.972 specificities, 0.947 precision, 0.963 F1-score, 0.974 AUPR, and 0.975 accuracies for ICH detection. The proposed solution obtains efficient results for ICH detection due to the efficiency of hybrid deep learning and machine learning techniques. The deep learning features extractors capture the important intensity and spatial features that significantly improve the ICH subtype classification and detection, as stated in Table 2.
Furthermore, the performance results for SDH subtype detection are 0.974 AUC, 0.965 sensitivity, 0.983 specificities, 0.963 precision, 0.972 F1-score, 0.971 AUPR, and 0.987 accuracies. These results show significant improvement in detecting SDH due to its similarity with EDH. Likewise, 0.980, 0.972, 0.963, 0.966, 0.952, 0.964, and 0.975 are obtained for detecting EDH in terms of AUC, sensitivity, specificity, precision, F1-score, AUPR, and accuracy, respectively. The proposed solution secures excellent results for EDH, which is a significant improvement. The traditional studies have fewer detection scores for SDH and EDH due to their hidden subtleties and similarities.
Moreover, the dataset contains less number of EDH cases, which are usually falsely classified by the model. However, the proposed solution has the detection capability that efficiently learns the bleeding regions and reduces the misclassification of EDH. Afterwards, the detection of other ICH subtypes such as SAH, IPH, and IVH are also efficiently covered by the proposed solution with a minimum false detection rate. The detection results of SAH, IPH, and IVH by the Res-Inc-LGBM are also similar to the SDH and EDH, as given in Table 2. In general, the results demonstrate that the proposed solution has significantly improved ICH’s detection and subtype classification.
To validate the effectiveness of ResNet101-V2 for intensity base features and Inception-V4 for spatial, Table 2 presents the results in terms of ResNet101-V2-LGBM (Res-LGBM) and Inception-V4-LGBM (Inc-LGBM). These are the two variants of the proposed Res-Inc-LGBM mechanism to show the importance of intensity and spatial features. In the first variant Res-LGBM, the Inception-V4 is removed from the mechanism to demonstrate the importance of spatial features. Res-LGBM receives windowed CT slices as an input and outputs the intensity features, which are fed as an input to the LGBM for the detection and subtype characterization of ICH. Likewise, the other variant Inc-LGBM highlights the significance of intensity features by removing the Inception-V4 module. The Inc-LGBM model is trained with the adjacent slices as an input and outputs the spatial features.
The LGBM uses spatial features to identify and classify the ICH into subtypes. In terms of ICH detection, the Res-LGBM-based variant gains 0.938 AUC, 0.764 sensitivity, 0.926 specificities, 0.825 precision, 0.843 F1-score, 0.892 AUPR, and 0.926 accuracies. It is seen that the spatial features have a significant impact, such as about 5% to 13% in terms of AUC, sensitivity, F1-score, and AUPR. Likewise, the Inc-LGBM achieves 0.952, 0.803, 0.926, 0.863, 0.875, 0.952, and 0.936 for AUC, sensitivity, specificity, precision, F1-score, AUPR, and accuracy in terms of ICH detection, respectively. The results demonstrate that the intensity-based features also impact the detection score, such as about 3% to 15% regarding AUC, sensitivity, F1-score, and AUPR. It is concluded that the proposed feature extractors have improved the ICH detection and subtype categorization performance due to excellent feature extraction capabilities.
The ICH subtype classification results of Res-LGBM and Inc-LGBM are quite inferior to the results of the proposed Res-Inc-LGBM solution, as given in Table 2. The table states that the performance results of Res-LGBM for EDH are 0.927 AUC, 0.804 sensitivity, 0.896 specificities, 0.883 precision, 0.832 F1-score, 0.844 AUPR, and 0.936 accuracies. In comparison, the Inc-LGBM obtains 0.935 AUC, 0.822 sensitivity, 0.885 specificities, 0.895 precision, 0.875 F1-score, 0.902 AUPR, and 0.945 accuracies for EDH. It is seen that the Inc-LGBM has high AUC, sensitivity, precision, F1-score, AUPR, and accuracy results than the Res-LGBM due to the significance of spatial features and Inception-V4 architecture. Afterwards, for SDH subtype detection, Res-LGBM scores 0.951 AUC, 0.924 sensitivity, 0.986 specificities, 0.905 precision, 0.892 F1-score, 0.931 AUPR, and 0.992 accuracies, whereas Inc-LGBM gains 0.957 AUC, 0.931 sensitivity, 0.963 specificities, 0.962 precision, 0.893 F1-score, 0.945 AUPR, and 0.996 accuracies.
The Inc-LGBM has results superior to those of the Res-LGBM for AUC, sensitivity, F1-score, AUPR, and accuracy in terms of SDH. Specifically, the Inc-LGBM has more excellent results for ICH detection and subtype categorization than the Res-LGBM, as stated in Table 2, although it is clear that the spatial features have more impact on ICH detection performance than the intensity-based features. In the combination of both spatial and windowed features, the proposed Res-Inc-LGBM achieves more excellent results than its two variants, as validated by the performance results in Table 2.
The proposed solution’s generalization ability can be seen in Table 3 by testing it on the CQ500 dataset. The table presents ICH detection and its subcategorization results regarding recall (sensitivity), specificity, F1-score, and accuracy. The performance results clearly state the proposed methodology has excellent generalization ability for ICH detection. Specifically, using the CQ500 dataset, the proposed Res-Inc-LGBM achieves 95.4%, 94.2%, 94.1%, and 95.1% for sensitivity, specificity, F1-score, and accuracy, respectively. Therefore, the proposed solution efficiently handles the complex task of ICH detection using the unseen dataset.

3.2. Performance Comparison with Benchmarks

Here, we demonstrate the performance of the proposed solution against other state-of-the-art techniques, Wang et al. [2,8,17,19,26,27] for ICH detection, and subtype classification. The authors in [2] achieve 0.964 AUC, 0.944 specificities, and 0.887 sensitivity for ICH detection compared to the proposed model that scores 0.985 AUC, 0.972 specificities, and 0.954 sensitivity. The proposed solution achieves more excellent results due to its efficient strategy of LGBM with rapid training. Afterwards, the proposed Res-Inc-LGBM compares with the approach presented in [8] where the benchmark model covers 0.943 for AUC and 0.859 for sensitivity in ICH detection, whereas the proposed mechanism achieves 0.985 AUC and 0.954 sensitivity. Furthermore, the approach presented in [8] achieves 0.792 sensitivity, 0.823 F1-score, and 0.986 accuracies for EDH. In contrast, the proposed Res-Inc-LGBM model obtains 0.972 sensitivity, 0.952 F1-score, and 0.975 accuracy to detect the EDH, which is quite a complex subtype of ICH due to its similarity with other subtypes of ICH.
The researchers Keshavamurthy et al. [17] achieve an accuracy level of 0.905, a sensitivity of 0.936, 0.964 AUC, and 0.882 specificities to detect the brain injury using CNN and linear SVM model. The proposed technique finds more excellent results, such as an accuracy level of 0.975, sensitivity of 0.954, 0.985 AUC, and 0.972 specificities. Another similar mechanism is proposed in [19] that gets a score of 0.902 for AUC, 0.894 for specificity, 0.897 for accuracy, and 0.901 for sensitivity to identify the ICH. On the other hand, the Res-Inc-LGBM has more excellent results for ICH detection and subtype categorization, as mentioned in Table 2. For instance, the approach based on [19] achieves 0.727 for identifying EDH subtype in terms of F1-score, whereas the Res-Inc-LGBM covers 0.952 F1-score. Bruja et al. [26] present a CNN- and LSTM-based model that achieves a score of 0.865 for sensitivity, 0.957 for specificity, 0.972 for AUC, and 0.96 for accuracy regarding the ICH detection. In contrast, the proposed model has more excellent results for ICH detection than the [26], as given in Table 2. Another study [27] also implements the same mechanism of CNN and LSTM that achieves 0.956 for AUC versus the proposed solution that covers the AUC score of 0.985 for ICH detection.
The proposed Res-Inc-LGBM overcomes the shortcomings of standard techniques regarding subtype classification and detection of ICH. The main reason behind this is the efficiency in feature extraction and detection through the LGBM, which is faster and more efficient for the identification and subtype classification of ICH. Therefore, the proposed mechanism is an efficient and suitable choice for ICH detection and its subtype categorization.

4. Conclusions

In this paper, ICH detection and its subtype classification are improved by presenting the new solution, known as Res-Inc-LGBM, using the dataset of IHDC. The proposed Res-Inc-LGBM mechanism uses two deep learning feature extractors, ResNet101-V2 and Inception-V4, to find the significant spatial and intensity features. Afterwards, the deep learning feature extractors are combined with the more efficient machine learning boosting technique, LGBM, that uses the extracted features as input and performs the subtype categorization and detection of ICH. The complete methodology is evaluated by extensive experimentation to show its excellent performance and suitability for real-time ICH detection.
The experiment results of the proposed Res-Inc-LGBM for ICH detection show that the discriminative capability of the model was excellent in terms of AUC, sensitivity, specificity, precision, F1-score, AUPR, and accuracy, respectively. Moreover, the proposed Res-Inc-LGBM is compared against the state-of-the-art techniques regarding ICH detection and its subtype characterization. The proposed model overcomes the performance of existing techniques for ICH detection, which shows its excellent detection capability and suitability for real-world problems. The proposed solution is developed and assessed over the dataset of IHDC. Therefore, different datasets will be used further to enhance the ICH subtype classification and detection performance.

Author Contributions

Conceptualization: M.A.; Data curation: M.A.; Formal analysis: M.A. and M.A.S.; Funding acquisition: E.A.N. and H.T.R.; Investigation: M.A., H.A.K. and E.A.; Methodology: M.A.S. and H.A.K.; Resources: M.A.S. and H.A.K.; Software: S.M.; Supervision: M.A.S. and H.A.K.; Validation: M.A.S. and H.A.K.; Visualization: M.A.S. and H.A.K.; Writing—original draft: M.A., S.M., E.A., E.A.N. and H.T.R.; Writing—review & editing: M.A. and H.A.K. All authors have read and agreed to the published version of the manuscript.

Funding

The authors present their appreciation to King Saud University for funding this research through the Researchers Supporting Program number (RSP2023R164), King Saud University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

The organization’s employee’s data used to consent and confidentiality.

Data Availability Statement

Data will be available upon reasonable request to authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hssayeni, M.D.; Croock, M.S.; Salman, A.D.; Al-khafaji, H.F.; Yahya, Z.A.; Ghoraani, B. Intracranial hemorrhage segmentation using a deep convolutional model. Data 2020, 5, 14. [Google Scholar] [CrossRef]
  2. Wang, X.; Shen, T.; Yang, S.; Lan, J.; Xu, Y.; Wang, M.; Zhang, J.; Han, X. A deep learning algorithm for automatic detection and classification of acute intracranial hemorrhages in head CT scans. NeuroImage Clin. 2021, 32, 102785. [Google Scholar] [PubMed]
  3. Heit, J.J.; Iv, M.; Wintermark, M. Imaging of intracranial hemorrhage. J. Stroke 2017, 19, 11. [Google Scholar] [PubMed]
  4. Cordonnier, C.; Demchuk, A.; Ziai, W.; Anderson, C.S. Intracerebral haemorrhage: Current approaches to acute management. Lancet 2018, 392, 1257–1268. [Google Scholar] [PubMed]
  5. Ko, H.; Chung, H.; Lee, H.; Lee, J. Feasible study on intracranial hemorrhage detection and classification using a CNN-LSTM network. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 1290–1293. [Google Scholar]
  6. Øie, L.R.; Madsbu, M.A.; Solheim, O.; Jakola, A.S.; Giannadakis, C.; Vorhaug, A.; Padayachy, L.; Jensberg, H.; Dodick, D.; Salvesen, Ø.; et al. Functional outcome and survival following spontaneous intracerebral hemorrhage: A retrospective population-based study. Brain Behav. 2018, 8, e01113. [Google Scholar]
  7. Kuo, W.; Hane, C.; Mukherjee, P.; Malik, J.; Yuh, E.L. Expert-level detection of acute intracranial hemorrhage on head computed tomography using deep learning. Proc. Natl. Acad. Sci. USA 2019, 116, 22737–22745. [Google Scholar] [CrossRef]
  8. Wu, Y.; Supanich, M.P.; Jie, D. Ensembled deep neural network for intracranial hemorrhage detection and subtype classification on noncontrast CT images. J. Artif. Intell. Med Sci. 2021, 2, 12–20. [Google Scholar]
  9. Yeo, M.; Tahayori, B.; Kok, H.K.; Maingard, J.; Kutaiba, N.; Russell, J.; Thijs, V.; Jhamb, A.; Chandra, R.V.; Brooks, M.; et al. Review of deep learning algorithms for the automatic detection of intracranial hemorrhages on computed tomography head imaging. J. Neurointerventional Surg. 2021, 13, 369–378. [Google Scholar] [CrossRef]
  10. Bello, H.R.; Graves, J.A.; Rohatgi, S.; Vakil, M.; McCarty, J.; Van Hemert, R.L.; Geppert, S.; Peterson, R.B. Skull base–related lesions at routine head CT from the emergency department: Pearls, pitfalls, and lessons learned. Radiographics 2019, 39, 1161–1182. [Google Scholar]
  11. Imran, R.; Hassan, N.; Tariq, R.; Amjad, L.; Wali, A. Intracranial brain Haemorrhage segmentation and classification. iKSP J. Comput. Sci. Eng. 2021, 1, 52–56. [Google Scholar]
  12. Lee, H.; Yune, S.; Mansouri, M.; Kim, M.; Tajmir, S.H.; Guerrier, C.E.; Ebert, S.A.; Pomerantz, S.R.; Romero, J.M.; Kamalian, S.; et al. An explainable deep-learning algorithm for the detection of acute intracranial haemorrhage from small datasets. Nat. Biomed. Eng. 2019, 3, 173–182. [Google Scholar] [CrossRef] [PubMed]
  13. Rao, B.; Zohrabian, V.; Cedeno, P.; Saha, A.; Pahade, J.; Davis, M.A. Utility of artificial intelligence tool as a prospective radiology peer reviewer—Detection of unreported intracranial hemorrhage. Acad. Radiol. 2021, 28, 85–93. [Google Scholar] [CrossRef] [PubMed]
  14. Hostettler, I.C.; Muroi, C.; Richter, J.K.; Schmid, J.; Neidert, M.C.; Seule, M.; Boss, O.; Pangalu, A.; Germans, M.R.; Keller, E. Decision tree analysis in subarachnoid hemorrhage: Prediction of outcome parameters during the course of aneurysmal subarachnoid hemorrhage using decision tree analysis. J. Neurosurg. 2018, 129, 1499–1510. [Google Scholar] [CrossRef] [PubMed]
  15. Kuang, H.; Qiu, W.; Boers, A.M.; Brown, S.; Muir, K.; Majoie, C.B.; Dippel, D.W.; White, P.; Epstein, J.; Mitchell, P.J.; et al. Computed tomography perfusion–based machine learning model better predicts follow-up infarction in patients with acute ischemic stroke. Stroke 2021, 52, 223–231. [Google Scholar]
  16. Ramos, L.A.; van der Steen, W.E.; Barros, R.S.; Majoie, C.B.; van den Berg, R.; Verbaan, D.; Vandertop, W.P.; Zijlstra, I.J.A.J.; Zwinderman, A.; Strijkers, G.J.; et al. Machine learning improves prediction of delayed cerebral ischemia in patients with subarachnoid hemorrhage. J. Neurointerventional Surg. 2019, 11, 497–502. [Google Scholar]
  17. Keshavamurthy, K.N.; Leary, O.P.; Merck, L.H.; Kimia, B.; Collins, S.; Wright, D.W.; Allen, J.W.; Brock, J.F.; Merck, D. Machine learning algorithm for automatic detection of CT-identifiable hyperdense lesions associated with traumatic brain injury. In Proceedings of the Medical Imaging 2017: Computer-Aided Diagnosis SPIE, Orlando, FL, USA, 11–16 February 2017; Volume 10134, pp. 630–638. [Google Scholar]
  18. Guo, R.; Zhang, R.; Liu, R.; Liu, Y.; Li, H.; Ma, L.; He, M.; You, C.; Tian, R. Machine learning-based approaches for prediction of patients’ functional outcome and mortality after spontaneous intracerebral hemorrhage. J. Pers. Med. 2022, 12, 112. [Google Scholar] [CrossRef]
  19. Sage, A.; Badura, P. Intracranial hemorrhage detection in head CT using double-branch convolutional neural network, support vector machine, and random forest. Appl. Sci. 2020, 10, 7577. [Google Scholar] [CrossRef]
  20. Arbabshirani, M.R.; Fornwalt, B.K.; Mongelluzzo, G.J.; Suever, J.D.; Geise, B.D.; Patel, A.A.; Moore, G.J. Advanced machine learning in action: Identification of intracranial hemorrhage on computed tomography scans of the head with clinical workflow integration. NPJ Digit. Med. 2018, 1, 9. [Google Scholar]
  21. Grewal, M.; Srivastava, M.M.; Kumar, P.; Varadarajan, S. Radnet: Radiologist level accuracy using deep learning for hemorrhage detection in ct scans. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 281–284. [Google Scholar]
  22. Mansour, R.F.; Escorcia-Gutierrez, J.; Gamarra, M.; Díaz, V.G.; Gupta, D.; Kumar, S. Artificial intelligence with big data analytics-based brain intracranial hemorrhage e-diagnosis using CT images. Neural Comput. Appl. 2021, 1–13. [Google Scholar]
  23. Karki, M.; Cho, J.; Lee, E.; Hahm, M.H.; Yoon, S.Y.; Kim, M.; Ahn, J.Y.; Son, J.; Park, S.H.; Kim, K.H.; et al. CT window trainable neural network for improving intracranial hemorrhage detection by combining multiple settings. Artif. Intell. Med. 2020, 106, 101850. [Google Scholar]
  24. Anupama, C.; Sivaram, M.; Lydia, E.L.; Gupta, D.; Shankar, K. Synergic deep learning model–based automated detection and classification of brain intracranial hemorrhage images in wearable networks. Pers. Ubiquitous Comput. 2022, 26, 1–10. [Google Scholar] [CrossRef]
  25. Ye, H.; Gao, F.; Yin, Y.; Guo, D.; Zhao, P.; Lu, Y.; Wang, X.; Bai, J.; Cao, K.; Song, Q.; et al. Precise diagnosis of intracranial hemorrhage and subtypes using a three-dimensional joint convolutional and recurrent neural network. Eur. Radiol. 2019, 29, 6191–6201. [Google Scholar] [CrossRef] [PubMed]
  26. Burduja, M.; Ionescu, R.T.; Verga, N. Accurate and efficient intracranial hemorrhage detection and subtype classification in 3D CT scans with convolutional and long short-term memory neural networks. Sensors 2020, 20, 5611. [Google Scholar] [PubMed]
  27. Nguyen, N.T.; Tran, D.Q.; Nguyen, N.T.; Nguyen, H.Q. A CNN-LSTM architecture for detection of intracranial hemorrhage on CT scans. medRxiv 2020, 2020-04. [Google Scholar]
  28. Chang, P.D.; Kuoy, E.; Grinband, J.; Weinberg, B.D.; Thompson, M.; Homo, R.; Chen, J.; Abcede, H.; Shafie, M.; Sugrue, L.; et al. Hybrid 3D/2D convolutional neural network for hemorrhage evaluation on head CT. Am. J. Neuroradiol. 2018, 39, 1609–1616. [Google Scholar]
  29. Lee, J.Y.; Kim, J.S.; Kim, T.Y.; Kim, Y.S. Detection and classification of intracranial haemorrhage on CT images using a novel deep-learning algorithm. Sci. Rep. 2020, 10, 20546. [Google Scholar] [PubMed]
  30. RSNA. Intracranial Hemorrhage Detection Challenge; Technical report; Radiological Society of North America: Oak Brook, IL, USA, 2019. [Google Scholar]
  31. Flanders, A.E.; Prevedello, L.M.; Shih, G.; Halabi, S.S.; Kalpathy-Cramer, J.; Ball, R.; Mongan, J.T.; Stein, A.; Kitamura, F.C.; Lungren, M.P.; et al. Construction of a machine learning dataset through collaboration: The RSNA 2019 brain CT hemorrhage challenge. Radiol. Artif. Intell. 2020, 2, e190211. [Google Scholar] [CrossRef]
  32. Chilamkurthy, S.; Ghosh, R.; Tanamala, S.; Biviji, M.; Campeau, N.G.; Venugopal, V.K.; Mahajan, V.; Rao, P.; Warier, P. Deep learning algorithms for detection of critical findings in head CT scans: A retrospective study. Lancet 2018, 392, 2388–2396. [Google Scholar] [CrossRef]
  33. Das, D.; Kalita, S.K. Skull stripping of brain MRI for analysis of Alzheimer’s disease. Int. J. Biomed. Eng. Technol. 2021, 36, 331–349. [Google Scholar]
  34. Kumar, K.A.; Boda, R. A multi-objective randomly updated beetle swarm and multi-verse optimization for brain tumor segmentation and classification. Comput. J. 2022, 65, 1029–1052. [Google Scholar]
  35. Xue, Z.; Antani, S.; Long, L.R.; Demner-Fushman, D.; Thoma, G.R. Window classification of brain CT images in biomedical articles. AMIA Annu. Symp. Proc. 2012, 2012, 1023–1029. [Google Scholar] [PubMed]
  36. Dinsdale, N.K.; Bluemke, E.; Smith, S.M.; Arya, Z.; Vidaurre, D.; Jenkinson, M.; Namburete, A.I. Learning patterns of the ageing brain in MRI using deep convolutional networks. NeuroImage 2021, 224, 117401. [Google Scholar] [PubMed]
  37. Rahman, T.; Khandakar, A.; Kadir, M.A.; Islam, K.R.; Islam, K.F.; Mazhar, R.; Hamid, T.; Islam, M.T.; Kashem, S.; Mahbub, Z.B.; et al. Reliable tuberculosis detection using chest X-ray with deep learning, segmentation and visualization. IEEE Access 2020, 8, 191586–191601. [Google Scholar]
  38. Gunasinghe, H.; McKelvie, J.; Koay, A.; Mayo, M. Comparison of pretrained feature extractors for glaucoma detection. In Proceedings of the 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI), Nice, France, 13–16 April 2021; pp. 390–394. [Google Scholar]
  39. Islam, M.; Sanghani, P.; See, A.A.Q.; James, M.L.; King, N.K.K.; Ren, H. ICHNet: Intracerebral hemorrhage (ICH) segmentation using deep learning. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: 4th International Workshop, BrainLes 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 16 September 2018; Revised Selected Papers, Part I. pp. 456–463. [Google Scholar]
  40. Ni, J.; Zhang, F.; Ma, F.; Yin, Q.; Xiang, D. Random region matting for the high-resolution polsar image semantic segmentation. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2021, 14, 3040–3051. [Google Scholar]
  41. Chollet, F. Keras; GitHub, Inc.: San Francisco, CA, USA, 2015. [Google Scholar]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; pp. 630–645. [Google Scholar]
  43. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  44. Hou, F.; Liu, B.; Zhuo, L.; Zhuo, Z.; Zhang, J. Remote sensing image retrieval with deep features encoding of Inception V4 and largevis dimensionality reduction. Sens. Imaging 2021, 22, 1–14. [Google Scholar] [CrossRef]
  45. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 2017, 30, 52. [Google Scholar]
  46. Wang, M.; Yue, L.; Yang, X.; Wang, X.; Han, Y.; Yu, B. Fertility-LightGBM: A fertility-related protein prediction model by multi-information fusion and light gradient boosting machine. Biomed. Signal Process. Control. 2021, 68, 102630. [Google Scholar]
Figure 1. Systematic view of the proposed methodology for detection and subtype classification of ICH.
Figure 1. Systematic view of the proposed methodology for detection and subtype classification of ICH.
Diagnostics 13 00652 g001
Figure 2. Overview of subtypes of ICH.
Figure 2. Overview of subtypes of ICH.
Diagnostics 13 00652 g002
Figure 3. Representation of subtypes of ICH.
Figure 3. Representation of subtypes of ICH.
Diagnostics 13 00652 g003
Figure 4. Representation of three intensity windows.
Figure 4. Representation of three intensity windows.
Diagnostics 13 00652 g004
Figure 5. Loss analysis of Inception-V4 during training.
Figure 5. Loss analysis of Inception-V4 during training.
Diagnostics 13 00652 g005
Figure 6. Accuracy of Inception-V4 during training.
Figure 6. Accuracy of Inception-V4 during training.
Diagnostics 13 00652 g006
Figure 7. AUC during the training Inception-V4.
Figure 7. AUC during the training Inception-V4.
Diagnostics 13 00652 g007
Figure 8. AUPR of Inception-V4 during execution.
Figure 8. AUPR of Inception-V4 during execution.
Diagnostics 13 00652 g008
Figure 9. Loss analysis of ResNet101-V2 during training.
Figure 9. Loss analysis of ResNet101-V2 during training.
Diagnostics 13 00652 g009
Figure 10. Accuracy of ResNet101-V2 during training.
Figure 10. Accuracy of ResNet101-V2 during training.
Diagnostics 13 00652 g010
Figure 11. AUC during the training ResNet101-V2.
Figure 11. AUC during the training ResNet101-V2.
Diagnostics 13 00652 g011
Figure 12. AUPR of ResNet101-V2 during execution.
Figure 12. AUPR of ResNet101-V2 during execution.
Diagnostics 13 00652 g012
Figure 13. AUC-based performance of the proposed solution.
Figure 13. AUC-based performance of the proposed solution.
Diagnostics 13 00652 g013
Figure 14. Proposed solution’s performance based on AUPR.
Figure 14. Proposed solution’s performance based on AUPR.
Diagnostics 13 00652 g014
Table 1. Detailed information about the dataset.
Table 1. Detailed information about the dataset.
HemorrhageRSNA InstancesCQ500 Scans
Any011,399205
11934
IVH012,88128
1452
SAH012,70860
1625
EDH013,27913
155
SDH012,48853
1845
IPH012,666134
1668
Table 2. ICH detection and subtype classification results of the proposed solution using the IHDC dataset.
Table 2. ICH detection and subtype classification results of the proposed solution using the IHDC dataset.
SubtypesModelAUCSensitivitySpecificityPrecisionF1-ScoreAUPRAccuracy
anyRes-Inc-LGBM0.9850.9540.9720.9470.9630.9740.975
Res-LGBM0.9380.7640.9260.8250.8430.8920.926
Inc-LGBM0.9520.8030.9260.8630.8750.9520.936
SDHRes-Inc-LGBM0.9740.9650.9830.9630.9720.9710.987
Res-LGBM0.9510.9240.9860.9050.8920.9310.992
Inc-LGBM0.9570.9310.9630.9620.8930.9450.996
SAHRes-Inc-LGBM0.9760.9620.9730.9640.9630.9720.985
Res-LGBM0.9670.9440.9730.9250.9120.9210.972
Inc-LGBM0.9750.9010.9820.9750.9320.9520.991
EDHRes-Inc-LGBM0.9800.9720.9630.9660.9520.9640.975
Res-LGBM0.9270.8040.8960.8830.8320.8440.936
Inc-LGBM0.9350.8220.8850.8950.8750.9020.945
IPHRes-Inc-LGBM0.9790.9680.9820.9730.9720.9650.987
Res-LGBM0.9560.9230.9510.9160.9110.9030.954
Inc-LGBM0.9630.9350.9660.9530.9320.9360.967
IVHRes-Inc-LGBM0.9770.9620.9730.9720.9630.9680.985
Res-LGBM0.9530.9340.9620.9430.9320.9220.963
Inc-LGBM0.9670.9540.9730.9670.9590.9520.974
Table 3. Evaluation results of Res-Inc-LGBM for ICH detection using CQ500 dataset.
Table 3. Evaluation results of Res-Inc-LGBM for ICH detection using CQ500 dataset.
ICHRecallSpecificityF1-ScoreAccuracy
any0.9520.9620.9510.971
SDH0.9510.9530.9500.962
SAH0.9430.9510.9460.960
EDH0.9540.9420.9410.951
IPH0.9410.9350.9420.963
IVH0.9340.9520.9370.961
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Asif, M.; Shah, M.A.; Khattak, H.A.; Mussadiq, S.; Ahmed, E.; Nasr, E.A.; Rauf, H.T. Intracranial Hemorrhage Detection Using Parallel Deep Convolutional Models and Boosting Mechanism. Diagnostics 2023, 13, 652. https://doi.org/10.3390/diagnostics13040652

AMA Style

Asif M, Shah MA, Khattak HA, Mussadiq S, Ahmed E, Nasr EA, Rauf HT. Intracranial Hemorrhage Detection Using Parallel Deep Convolutional Models and Boosting Mechanism. Diagnostics. 2023; 13(4):652. https://doi.org/10.3390/diagnostics13040652

Chicago/Turabian Style

Asif, Muhammad, Munam Ali Shah, Hasan Ali Khattak, Shafaq Mussadiq, Ejaz Ahmed, Emad Abouel Nasr, and Hafiz Tayyab Rauf. 2023. "Intracranial Hemorrhage Detection Using Parallel Deep Convolutional Models and Boosting Mechanism" Diagnostics 13, no. 4: 652. https://doi.org/10.3390/diagnostics13040652

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop