Next Article in Journal
Qualitative Evaluation of Intracranial Pressure Slopes in Patients Undergoing Brain Death Protocol
Previous Article in Journal
Overlapping between Wound Healing Occurring in Tumor Growth and in Central Nervous System Neurodegenerative Diseases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Framework to Detect Intracranial Hemorrhage Using Hybrid Deep Neural Networks

by
Manikandan Rajagopal
1,
Suvarna Buradagunta
2,
Meshari Almeshari
3,
Yasser Alzamil
3,
Rajakumar Ramalingam
1 and
Vinayakumar Ravi
4,*
1
Department of CST, Madanapalle Institute of Technology & Science, Madanapalle 517325, India
2
Department of CSE, Vignan’s Foundation for Science, Technology, and Research Vadlamudi, Guntur 522213, India
3
Department of Diagnostic Radiology, College of Applied Medical Sciences, University of Ha’il, Ha’il 55476, Saudi Arabia
4
Center for Artificial Intelligence, Prince Mohammad Bin Fahd University, Khobar 34754, Saudi Arabia
*
Author to whom correspondence should be addressed.
Brain Sci. 2023, 13(3), 400; https://doi.org/10.3390/brainsci13030400
Submission received: 29 December 2022 / Revised: 30 January 2023 / Accepted: 23 February 2023 / Published: 25 February 2023
(This article belongs to the Section Computational Neuroscience and Neuroinformatics)

Abstract

:
Intracranial hemorrhage (ICH) is a serious medical condition that necessitates a prompt and exhaustive medical diagnosis. This paper presents a multi-label ICH classification issue with six different types of hemorrhages, namely epidural (EPD), intraparenchymal (ITP), intraventricular (ITV), subarachnoid (SBC), subdural (SBD), and Some. A patient may experience numerous hemorrhages at the same time in some situations. A CT scan of a patient’s skull is used to detect and classify the type of ICH hemorrhage(s) present. First, our model determines whether there is a hemorrhage or not; if there is a hemorrhage, the model attempts to identify the type of hemorrhage(s). In this paper, we present a hybrid deep learning approach that combines convolutional neural network (CNN) and Long-Short Term Memory (LSTM) approaches (Conv-LSTM). In addition, to propose viable solutions for the problem, we used a Systematic Windowing technique with a Conv-LSTM. To ensure the efficacy of the proposed model, experiments are conducted on the RSNA dataset. The suggested model provides higher sensitivity (93.87%), specificity (96.45%), precision (95.21%), and accuracy (95.14%). In addition, the obtained F1 score results outperform existing deep neural network-based algorithms.

1. Introduction

Damage or bleeding in the intracranial vault is referred to as intracranial hemorrhage (ICH). ICH detection is time-consuming even for highly experienced personnel [1]. Examining ICH diagnosis, on the other hand, is critical and time-consuming in order to pinpoint the bleeding site. As a result, an efficient and precise diagnosis is critical for the subsequent treatment process. To address this issue, some researchers hope to create neural networks capable of identifying hemorrhages based on a patient’s cranial CT scan. Neural network-based models may help radiologists diagnose ICH in a timely and accurate manner [2,3].
According to current studies, several research projects are being conducted that use deep learning algorithms for effective image categorization and segmentation. Most well-known deep learning systems, such as Convolutional Neural Networks (CNN), have piqued the interest of academics due to their ease of use and efficiency in dealing with medical diagnosis support [4,5]. CT scans are real-time 3D structures made up of a collection of 2D slices. Though working with image voxels is possible, it requires a high level of computing complexity. The latter can be avoided by processing slices independently or by utilizing the lowest 3D context complexity. Many deep-learning algorithms have been introduced in the literature to solve ICH detection and categorization. The authors of [6,7] presented a convolution neural network (CNN) to handle ICH challenges including segmentation and classification. However, the model’s precision is not adequate.
In general, ICH is divided into six categories of hemorrhages. Epidural (EPD) pain occurs as a result of an injury to the dura mater and skull. Intraparenchymal (ITP) happens when blood pools in brain tissues; intraventricular (ITV) occurs when bleeding enters the brain’s ventricular system. SBC arises when blood gathers beneath the arachnoid and the pia mater. Subdural (SBD) bleeding occurs when blood flows between the dura mater and the arachnoid mater, and some (SOME) bleeding is defined as any hemorrhage [8]. Figure 1 depicts the many forms of hemorrhages, with the sixth being classified as SOME or ANY hemorrhage. The different types of hemorrhage occur in different areas of the skull.
Figure 1 depicts the type of bleeding, as well as the location and specifics of the hemorrhage. Even though the output from the neural network does not provide any location information, this is used to pinpoint the bleeding. Epidural (EPD), Intraparenchymal (ITP), Intraventricular (ITV), Subarachnoid (SBC), Subdural (SBD), and more designations are conceivable (SOME). If there is some hemorrhage, the final label will be correct. Multiple hemorrhages can occur at the same moment in some real-life settings [10].
The input for this task will be a single slice from a DICOM CT scan of a patient’s skull with dimensions of 512 × 512. The Dataset section goes into greater detail regarding the input data. For each label, the result will be a probability (between 0 and 1). Multiple hemorrhages can occur at the same time in real-world settings. As a result, the output labels are noy mutually exclusive. Each label is handled independently.
Figure 2 is a CT scan slice of a patient’s cranium showing s subdural hemorrhage as seen through the axial subdural window. It is mapped to its output, where subdural and some labels are 1 and the others are 0. A CT scan reveals some hemorrhage as well as a subdural hematoma.
This research presents a precise and well-organized method for analyzing single-slice 3D CT scan pictures using a standard neural network with systematic windowing. Only a few works that we are aware of use the systematic windowing technique, employing convolutional and recurrent layers for the ICH detection process. Very hybrid deep neural networks, however, are used directly on single-slice 3D CT scans. To successfully control ICH hemorrhage and produce an accurate outcome, a simple and efficient mixed Conv-LSTM approach is provided. We ensure that the proposed paradigm is simple in dealing with ICH concerns. Our primary contribution in this work is as follows.
  • First, a hybrid deep neural network technique, convolution neural network and Long-short term memory (Conv-LSTM) for efficient and rapid identification of ICH in CT scans.
  • Next, a systematic windowing approach is used to generate a temporal sequence. Conv-LSTM then extracts the spatiotemporal data and predicts it.
  • Finally, the proposed model performance study shows that it is more accurate and has a lower false prediction rate than other existing learning approaches.
The next sections are structured as follows. The second section addresses recent research on ICH identification using deep learning algorithms. The dataset including RSNA Section 3 depicts intracranial hemorrhage and linear transformation, widowing and normalization, and the weighted sample method. The projected model, which resembles the process employed by radiologists in which the scan is systematically viewed under multiple window settings before an assessment is made, is covered in Section 4. Section 5 describes in depth the experimentation and findings after examining the recommended strategies. Finally, Section 6 concludes the study, followed by future directions.

2. Related Works

McCulloch and Pitts initially introduced artificial neurons in 1943 [11]. Later, this approach was enriched through growth in technology and system advances and has currently been utilized in various domains as a pioneering technology for image analysis.
The literature introduced numerous standard and profound learning approaches. Initially, many studies dealt with the ICH identification procedure using typical machine learning algorithms. The threshold-based technique for determining the ICH was introduced in [12]. The method identified ICH sub-types based on their position, volume, and contour. To optimize the threshold value, the author used retrospective instances of 33 CT scans and constructed their model using 210 CT scan images. The model achieved 59% specificity and 98% sensitivity for detecting ICH and moderate accuracy in classifying ICH sub-types. Li et al. [9,13] later introduced two ways to segment the SAH region and hemorrhage. One method uses elastic registration to deal with the SBC space atlas, while another employs the Bayesian decision strategy to extract distance transform features.
SAH space segmentation, entropy, mean grey value, energy, and variance, on the other hand, were defined and used to train the Support Vector Machine (SVM) classifier to detect SAH hemorrhage. The author used 60 CT images to train the model and 69 CT scans to test it. They discovered that the introduced model has a testing sensitivity of 100%, an accuracy of 91%, and a specificity of 92%.
Later, deep learning approaches drew the attention of various academics due to their ease of use and effectiveness in dealing with complicated problems. Arbabshirani et al. [14] used five layers and two fully linked CNN architecture models to detect the ICH over 46,583 records, which is one of the most practical strategies. These data are derived from official clinical radiology reports. The model’s performance achieves 96% accuracy and decreases the time required to diagnose new patient records. Prevedello et al. [15] presented two strategies based on the CNNs methodology. The early procedures focused on preventing ICH and hydrocephalus during a CT scan, while other methods were used to detect ICH sub-class kinds. For training and testing purposes, the author used a dataset of 319 records. For ICH detection, they specified 91% accuracy, 85% specificity, and 90% sensitivity. Grewal et al. [16] introduced CNN with an RNN model, RADnet, that uses standard DenseNet architecture for 3D slice level prediction. At the CT level, the author reported an 81.82% accuracy in ICH prediction. However, the model’s specificity and sensitivity are poor, and it can only handle a small dataset for prediction.
Jnawali et al. [17] created an ensemble learning model for ICH detection and segmentation using CNN architecture. The training model is trained using a dataset of 34,848 CT images, of which 8465 had ICH and 26,383 did not, though 5509 CT scans with and without ICH were used for testing purposes. The model’s performance showed 77% sensitivity, 80% precision, and 87% accuracy. However, after processing vast datasets, the model failed to improve ICH detection accuracy. [18] introduced a hybrid 3D CNN-RNN model that works directly on 3D CT images. They also used the Grad-CAM technique for image segmentation to process CNN attention maps. However, due to a restricted dataset for training purposes, the model faces significant challenges with a fresh CT scan.
Chan et al. [19] reported a knowledge-based classifier for detecting acute intracranial hemorrhage (AIH). A slice interpolation approach was developed to account for the spatial dependency among the closest slices. It reported 84% specificity, and 82% and 100% sensitivity at the slice and lesion levels, respectively. Shahangian et al. [20] created a method for image segmentation in CT scans called distance-regulated level set evolution (DRLSE). They first treated the segmentation to remove the irreverent forms that exist around the ventricles. They then carried out the ICH splitting utilizing DRLSE. On CT slices, the author reported 82.5% sensitivity and 90.5% specificity. Chang et al. [21] proposed convolutional neural networks based on region of interest (ROI). For training and testing, a dataset of 10159 and 862 records with and without ICH was used.
The majority of the studies published in the literature show inferior accuracy in the detection of some ICH subtypes, similar to the detection outcomes of low-grade radiology trainees [6]. However, recognizing SAH and EDH is a significant challenge for the various machine-learning methods [22,23]. To categorize the ICH sub-class type and improve classification accuracy, an efficient deep learning model is required.

3. Materials and Methods

3.1. Dataset: RSNA Intracranial Hemorrhage Detection

This dataset was provided by the RSNA (Radiological Society of North America) as part of a Kaggle competition called RSNA Intracranial Hemorrhage Detection [24]. The dataset is freely available for non-commercial and academic research purposes (see Competition Rules, point 7(A)).
The compressed data is 180 GB; the uncompressed data is 448 GB. It is divided into two sections: a labelled training set and an unlabeled test set. This study makes use of a dataset for training, validation and testing. Here, a single DICOM image file containing a single slice from a 3D CT scan is used. Exactly 269 files out of 752,803 had erroneous dimensions. They were narrowed down to 752,534 DICOM files. These are all the same size, 512 by 512 pixels. The statistics of the database are shown in Table 1.
Knowing how many files there are for each type of hemorrhage is also essential. We know the total number of files that show ICH hemorrhage, as shown in Table 2.
The important fact to remember here is that the total number of files (3145 + 36,115 + 26,204 + 35,675 + 47,060) equals 148,199. This is unusual because only 107,826 files reveal SOME bleeding (s). The 40,373-byte file has many hemorrhages. While counting, these files are counted many times, resulting in such an anomaly.
Now that we know that files display 1, 2, 3, 4, and 5 hemorrhages(s), we should look at how each is distributed in different files. There are now five “types” of files available (that show SOME hemorrhage). As a result, the class-wise distribution is as presented in Table 3.
The images were chosen carefully from this new data set for the neural networks we discuss in later sections.

3.2. Linear Transformation, Windowing and Normalization

DICOM files are not the same as standard files, such as JPEG or PNG images. DICOM files contain a lot of metadata, and the range of values for each pixel (voxel) is relatively broad when compared to normal image files.
The DICOM files were parsed using the Pydicom [25] package. To obtain the original Hounsfield values, a simple linear transformation is performed on the raw pixel array of the DICOM file.
Hounsfield   value = Raw   pixel   value × RescaleSlope + Rescale   Intercept
The Rescale Slope and Intercept are in the DICOM file’s metadata.
After this linear transformation, an optional windowing might be performed. The window width and center values must be known to apply a window.
Min = Window   Center Window   Width 2 Max = Window   Center + Window   Width 2
new   value = { Min   if   value < Min Max   if   value > Max
After the optional windowing, pixel values are normalized to obtain a value between 0 and 1.
These steps are applied in training, validation and testing. This linear transformation and normalization should be performed even when making predictions with the model.

3.3. Weighted Samples

As described in Section 3.1, there is a significant class imbalance among various forms of hemorrhages. Class weights were employed to compensate for this mismatch and to allow the neural network to train more aggressively on infrequently observed hemorrhages.

3.4. Problem Formulation

Because the ground truth values are known, this is a supervised learning problem. Because the output classes are not mutually exclusive, this is a multi-label classification problem rather than a multi-class classification problem. Each output label is self-contained and functions as its own binary classification problem.

3.5. Selection of Appropriate Loss Function

Binary cross-entropy [26] is the chosen loss function taken element-wise over the entire output. The loss formula is given in Equation (4).
loss = ( y log p + ( 1 y ) log ( 1 p ) )
where y is the ground truth value, and p is the predicted value.
This is taken element-wise over all the output labels and then averaged as per Equations (5) and (6).
loss EPD = ( y EPD × log ( p EPD ) + ( 1 y SBD ) × log ( 1 p SBD ) ) loss ITP = ( y ITP × log ( p ITP ) + ( 1 y ITP ) × log ( 1 p ITP ) ) loss ITV = ( y ITV × log ( p ITV ) + ( 1 y ITV ) × log ( 1 p ITV ) ) loss SBC = ( y SBC × log ( p SBC ) + ( 1 y SBC ) × log ( 1 p SBC ) ) loss SBD = ( y SBD × log ( p SBD ) + ( 1 y SBD ) × log ( 1 p SBD ) ) loss SOME = ( y SOME × log ( p SOME ) + ( 1 y SOME ) × log ( 1 p SOME ) )    
loss = loss EPD + loss ITP + loss ITV + loss SBC + loss SBD + loss SOME 6

3.6. Model-Specific Details

Both models have a six-unit dense layer with sigmoid activation to ensure the output value for each label stays between 0 and 1.
sigmoid   ( x ) = 1 ( 1 + e x )
Internal layers use ReLU as the activation function to make the training process faster.
relu   ( x ) = { 0   if   x 0 x   if   x > 0
We employed the Adam optimizer [27] in our model. Glorot Uniform [28] is the kernel and bias initializer that eliminates the issue of exploding and endangered gradients due to the network being too deep. To prevent overfitting, L2 [29] regularization is utilized. Keras [30] was chosen as the framework, with Tensor flow 2.0 [31] serving as the backend.

4. Proposed Approach

By combining convolutional and recurrent layers, this study presented a hybrid deep convolutional LSTM model. Furthermore, the training procedure is divided into three primary parts. In the first stage, we use the systematic windowing method to simplify detection. The second stage involves training our CNN model to predict ICH types at the CT slice level. The CNN model identifies the important features and feeds them into the LSTM model [32]. We add the LSTM model for predicting ICH types at the CT scan picture in the third stage of the training procedure.

4.1. Systematic Windowing

The kernel See like a Radiologist with Systematic Windowing [33] by David Tang discusses viewing DICOM files under various window settings before making a prognosis. This is the same strategy that radiologists utilize when reviewing CT scans in real-world scenarios. DICOM files vary depending on the window settings; certain hemorrhages are considerably simpler to identify under certain window settings. Windowing is used specifically for this purpose, reported in Table 4. Based on this kernel and our understanding of the process, ten window settings were chosen.
This methodology creates a temporal sequence. All window settings mentioned in Table 2 are applied, and the result is stacked. Figure 3 shows 10 different views of the DICOM image using different window settings. Figure 4 shows the creation of a temporal series taken as input by CNN. CNN extracts the spatiotemporal data and identifies the significant features. Later, the LSTM model utilizes the CNN model’s output and performs the accurate prediction level in the CT scan.

4.2. Convolutional Neural Network Model

The CNN model is broken into six main units. The first four units are convolution, max-pooling, and batch normalization layers, with the final layer being fully-connected or dense layers. Figure 5 depicts each unit; unit 1 accepts a 512 × 512 × 1 NumPy matrix as input and includes three separate convolutional layers with 4, 8, and 16 filters of dimensions 3 × 3, 3 × 3, and 3 × 3, respectively. By employing the “same” padding, all of these convolution processes preserve the original extent. A max-pooling layer was used to decrease the image to half of its original size by utilizing a pool size of 2 × 2 and a stride of 2 × 2.
According to Figure 5, unit 2 receives the preceding unit’s output of dimension 256 × 256 × 16 as input and has three separate convolutional layers with 4, 8, and 16 filters of dimensions 3 × 3, 3 × 3, and 5 × 5, respectively. By employing the “same” padding, all of these convolution processes preserve the original extent. A max pooling layer was used once more to decrease the image to half of its original size by employing a 2 × 2 pool size with a stride of 2 × 2. Unit 3 receives the preceding unit’s output of size 128 × 128 × 16 as input and has three distinct convolutional layers with 4, 8, and 16 filters of dimensions 5 × 5, 5 × 5, and 5 × 5, respectively. By employing the “same” padding, all of these convolution processes preserve the original extent. A max pooling layer was utilized once more to decrease the image to half of its original size by employing a 2 × 2 pool size with a stride of 2 × 2.
Unit 4 in Figure 5 takes the preceding unit’s output of dimension 64 × 64 × 16 as input and has two convolutional layers with 4 and 8 filters of dimensions 5 × 5 and 5 × 5, respectively. Using a 2 × 2 filter, a max-pooling layer was utilized to compress the image to half of its original size. Unit 5 accepts as input the previous unit’s output of dimensions 28 × 28 × 8. It contains two convolutional layers with 8 and 8 filters in 5 × 5 and 5 × 5 dimensions, respectively. A max-pooling layer helps to decrease the image to half of its original size when applying a 2 × 2 filter. By using unit 6, the output from the preceding unit is converted to an input of 800. Following that, there are four dense layers of 1024, 512, 256, and 128 units, respectively. The final/output layer has a thickness of 6 units. In the final layer, the activation function is sigmoid. The final shape is (6).

4.3. Deep Conv-LSTM Model

LSTMs are recurrent neural network variants that can handle data point sequences. LSTM networks have been used by several academics to handle complicated problems such as language modelling, machine translation, and others. The merging of CNN and LSTM is introduced in this study by mining key features from CT images using CNN and then providing input features to the LSTM. The LSTM then takes the feature sequences as input and predicts using the entire CT scan. Long short-term spatiotemporal relations can be mapped using LSTMs. As a result, it simulates a radiologist going through numerous windows before making an assessment.
Figure 6 depicts the entire architecture of our proposed model. The CNN model retrieves features from the CT image using the 10-window slice input. The features are then incorporated into the LSTM model, which appropriately classifies the ICH hemorrhage kinds.

4.4. Training and Testing

In this paper, we chose a training dataset of 72,516 CT scans for training purposes. The training dataset contains a variety of hemorrhages. Furthermore, the initial sample weights were used to account for the changing ratios in which the photos appeared. The model was trained on the chosen dataset for 500 epochs. Based on the results of the numerous runs, we decided on 500 epochs. because the model stops learning after 500 epochs and produces the same result. The performance of the proposed model is measured every 100 epochs to confirm that the learning process is improving. The suggested network is clearly correcting itself to recognize negatives more correctly than previously. In contrast, we chose 7515 CT scans for validation and 7512 CT scans for testing. The CT scan pictures chosen for training, validation and testing are used in other models to compare the performance of the proposed model.

5. Experimentation and Results Analysis

This section covers the performance indicators that will be used to validate the proposed approach. Furthermore, the experimental results and analysis of results with other existing models are discussed here.

5.1. Performance Metrics

To validate our proposed approach, we used major performance metrics. The classification performance was assessed using the metrics listed below.
(a)
Sensitivity (also known as True Positive rate (TPR)):
TPR = TP TP + FN × 100 %
(b)
Specificity (also known as True Negative rate (TNR)):
TNR = TN TN + FP × 100 %
(c)
Accuracy:
Acc = TP + TN TP + FP + TN + FN × 100 %
(d)
F1 Score:
F 1 = 2 × TP 2 × TP + FP + FN × 100 %
where TP and TN specify the true positive and negative rates, and FP and FN specify the false positive and negative rates, correspondingly.

5.2. Experimental Results

The experiment was carried out using a PC equipped with an i5 processor running at 2.3 GHz, 16 GB RAM, and 500 GB SSD file storage, and the dataset was obtained from the Kaggle repository. Furthermore, we employed the Python 3.8 tool together with significant libraries, such as Tensorflow, Keras, NumPy, Matplotlib, and OpenCV-python. A collection of 87,543 files contained 512 × 512 pixels from patients’ CT scans [24]. We chose 72,516 CT scans for training and 7515 and 7512 CT scans for validation and testing from the given dataset. The outcomes are measured using the standard performance measures described in Section 5.1. The suggested model’s performance is compared to that of state-of-the-art techniques, including ResNexT [34], SVM, CNN [35], AlextNet + SVM [36], U-Net, and WA-ANN [37]. The Adam optimizer was used, employing a binary cross entropy loss and a learning proportion of 0.0001.
Table 5 and Figure 7 show the results of the Conv-LSTM model after 500 epochs. We discovered that the proposed model has a maximum TPR of 92.09%, TNR of 94.90%, precision of 94.57%, and accuracy of 94.91% after analyzing the classifier output across 100 generations. On the other hand, the results of 200 epochs show that the suggested model achieves 93.18% TPR, 96.52% TNR, 95.10% precision, and 95.60% accuracy. Simultaneously, the results under 300 epoch show a TPR of 92.75%, TNR of 95.10%, precision of 95.79%, and accuracy of 94.43%. Under 400 generations, the suggested model inevitably obtains a TPR of 94.54%, TNR of 95.31%, precision of 94.90%, and accuracy of 95.70%. The suggested Conv-LSTM technique yields an optimal TPR of 95.20%, TNR of 96.80%, precision of 96.27%, and accuracy of 95.70%.
Figure 8 depicts the outcome of the proposed model’s ROC (receiver operating characteristic) curve. It depicts the network’s performance at various thresholds. The AUC (area under the curve) score was determined separately for each of the labels listed in Table 6. The proposed model has a better ROC-AUC score for some of 94.74%, EPD of 93.11%, ITP of 95.13%, ITV of 93.06%, SBC of 95.11%, and SBD of 95.37% for classifying the ICH sub-class.

5.3. Comparative Result Analysis

Table 7 presents a comprehensive compared result analysis of the proposed model with other state-of-the-art existing methodologies. Figure 9 depicts the sensitivity and specificity of the proposed model and alternative models. The WA-ANN model resulted in low performance in terms of TPR and TNR, with 64.87% and 72.15%, respectively, whereas U-Net performed somewhat better than WA-ANN, with sensitivities of 64.54% and 87.21%, respectively. As a result, the SVM technique produced better results, with a TPR of 75.12% and a TNR of 78.64%. Furthermore, the CNN model attempted to deliver superior results, with a TPR of 86.61% and a TNR of 88.94%. At the same time, the ResNexT model produced a reasonable classification result with a sensitivity of 87.84% and a specificity of 90.78%. Though the AlexNet+SVM model produces superior results with specificity and sensitivity of 92.45% and 93.74%, respectively, the suggested model outperforms all other models with sensitivity of 93.87% and 96.45%. It is stated unequivocally that the suggested model achieves correct classifications and achieves higher values of TPR, TNR, FPR, and FNR on the tested examples.
Figure 10 depicts the precision and accuracy attained by the proposed model and the other models compared. The graphic illustrates that the existing WA-ANN and SVM models do not perform better on the test case photos. WA-ANN and SVM models achieved precision of 71.94% and 79.27%, respectively, while WA-ANN and SVM projected accuracy of 70.27% and 78.24%, respectively. This demonstrates that classifying these two models does not produce superior results. Furthermore, the U-Net and CNN models produced results with precision of 77.39%, 87.98%, and accuracy of 87.50% and 87.63%, respectively. Although these two models attempted to produce better results, they were unable to categorize the photos containing the bleeding.
Furthermore, the ResNexT and AlexNet+SVM models tried to provide a competitive outcome with a precision of 94.86%, 93.67% and accuracy of 90.71% and 93.18%; the proposed model has surpassed the other existing approaches with a precision of 95.21% and accuracy of 95.14%, respectively. From the comparative results, we observed that the proposed model achieved better outcomes; thereby, the proposed Conv-LSTM model is a suitable technique for ICH diagnosis. The superior result of Conv-LSTM is due to the significant feature extraction and selection using the CNN model and LSTM-based classification.

5.4. Discussion

Conv-LSTM produces balanced outcomes with primarily minor differences. Conv-LSTM architecture is a hybrid of CNN and LSTM models, with EPD and ITV classification performing most and least efficiently, respectively. Except for the EPD, the CNN has a minor edge over the LSTM. Keep in mind, however, that the EPD and ITV provide much poorer classification results than the other subtypes. The subtype-classification-assessment ranking represents the order produced by individual ResNets. Nonetheless, the proposed architecture improves classification in each category. According to Table 6, the increases in ROC-AUC score vary by subtype, ranging from approximately 0.1–0.5% in Some, ITP, SBC, and SBD to more than 1–2% in EPD and ITV, which have the most space for advancement. The regularity of performance across distinct ICHs leads us to infer that using an external learning classifier with hybrid deep features is beneficial in ICH detection.
The usage of multi-source features for classification can be linked to improved performance of the proposed Conv-LSTM architecture. The combination of features retrieved by networks addressing specific picture data representations yields a more comprehensive set of data. A slice-by-slice neighborhood context appears to be very important. If the entire CT scan is arranged correctly, bleeding should be seen in the following slices. As a result, such geographical information boosts the automated system’s cognitive capacities.
Slice-wise brain hemorrhage detection frameworks typically operate on the full CT slice or, in the case of our technique, conduct some primary ROI extraction to prepare the data for analysis. The deep learning tool handles the majority of the processing, with the operator having little influence on feature extraction. Some focused localization of specific areas of the image may boost detection and classification results. The differences in hemorrhage subtypes are related to their location and morphology. Considering the foregoing in more detailed specialized processes may aid categorization accuracy improvement. Mechanical head traumas frequently result in hemorrhages. In addition to the deep learning analysis, the external area of the skull might be analyzed.
The lowest metrics obtained in EPD classification highlight concerns about class imbalance in a single classification challenge. The number of epidural cases is significantly lower than those of other kinds. That is most likely the case in this example of bleeding, and the class weighing used during training is insufficient to remedy the problem. However, no such association exists across all subtypes. The ITP with the highest classification accuracy (detecting 29 out of 30 instances correctly) is not the best-represented hemorrhage; the second-best ITV (14 out of 15 right) is the second-worst-represented ICH subtype. Thus, the presentation of a specific hemorrhage in CT has a significant impact on detection skills. Similarities between various bleeding symptoms and characteristics are also possible (recall that the dataset of negative samples in each subtype classification also contains images with all other hemorrhages). Some complex examples exhibit a mix of numerous kinds at the same time. As a result, correct classification is surely a difficult assignment.
Table 8 summarizes the F1 score values from relevant state-of-the-art ICH detection approaches, as well as our proposed Conv-LSTM. Because it ignores true negatives, the F1 score is a credible metric in severely imbalanced datasets. The approaches used addressed at least half of our bleeding subtype set detection. Most crucially, they found hemorrhages in CT slices rather than the complete 3D series, as we did. The RSNA competition database was used by Burduja et al. [8] and Danilov et al. [38]. We used this to supply data and evaluation results to calculate the F1 score when it was not clearly stated.

6. Conclusions

This work contributes to the investigation of CNN and LSTM techniques for determining intracranial hemorrhage (ICH) within the skull. The research demonstrates that the network can predict the type of ICH found in a CT scan. This is seen from the network’s sensitivity, specificity, and accuracy over predictions on the testing data. It is equally effective in avoiding false positives and false negatives, which is a major issue in medical testing. The ROC-AUC ratings for each label show that the proposed study detects certain types of hemorrhages. This network uses a 2D slice from a 3D scan to predict in all of the slices. It allows us to evaluate the full brain scan, examine each section separately, and determine the ICH class kinds. Experiment findings show that the proposed model has a sensitivity of 93.87%, a specificity of 96.45%, a precision of 95.21%, and an accuracy of 95.14%. The proposed Conv-LSTM model outperforms the competing models. The work can be expanded further by introducing an optimization algorithm to select the optimal features and to train the model to determine the optimal weight classifier.

Author Contributions

Conceptualization, M.R., S.B, R.R. and V.R; methodology, M.R., S.B. and R.R.; software, M.A. and Y.A.; validation, M.A., Y.A. and R.R.; formal analysis, M.R.; investigation, S.B. and V.R.; resources, V.R.; data curation, R.R.; writing—original draft preparation, M.R., S.B. and R.R; writing—review and editing, M.A., Y.A. and V.R; visualization, M.A., V.R. and Y.A.; supervision, R.R. and V.R. funding: V.R All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Dataset (https://www.kaggle.com/c/rsna-intracranial-hemorrhage-detection/data, accessed on 14 September 2022) is taken from RSNA Intracranial Hemorrhage Detection [1] competition on Kaggle. Used under point 7(A) of Competition Rules [2] (B. General Competition Rules) stated at https://www.kaggle.com/c/rsna-intracranial-hemorrhage-detection/rulesfor the purpose of academic research.

Acknowledgments

We thank Saktheeswar V.B., Department of English & Foreign Languages, MITS, Madanapalle for the proofread and English corrections.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Van Asch, C.J.; Luitse, M.J.; Rinkel, G.J.; van der Tweel, I.; Algra, A.; Klijn, C.J. Incidence, case fatality, and functional outcome of intracerebral haemorrhage over time, according to age, sex, and ethnic origin: A systematic review and meta-analysis. Lancet Neurol. 2010, 9, 167–176. [Google Scholar] [CrossRef]
  2. Huisman, T.A. Intracranial hemorrhage: Ultrasound, CT and MRI findings. Eur. Radiol. 2005, 15, 434–440. [Google Scholar] [CrossRef] [PubMed]
  3. Currie, S.; Saleem, N.; Straiton, J.A.; Macmullen-Price, J.; Warren, D.J.; Craven, I.J. Imaging assessment of traumatic brain injury. Postgrad. Med. J. 2015, 92, 41–50. [Google Scholar] [CrossRef] [PubMed]
  4. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  5. Olaf, R.; Fischer, P.; Brox, Y. U-Net: Convolutional Networks for Biomedical Image Segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  6. Guo, D.; Wei, H.; Zhao, P.; Pan, Y.; Yang, H.-Y.; Wang, X.; Bai, J.; Cao, K.; Song, Q.; Xia, J.; et al. Simultaneous classification and segmentation of intracranial hemorrhage using a fully convolutional neural network. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; pp. 118–121. [Google Scholar] [CrossRef]
  7. Majumdar, A.; Brattain, L.; Telfer, B.; Farris, C.; Scalera, J. Detecting intracranial hemorrhage with deep learning. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 583–587. [Google Scholar]
  8. Burduja, M.; Ionescu, R.T.; Verga, N. Accurate and efficient intracranial hemorrhage detection and subtype classi-fication in 3D CT scans with convolutional and long short-term memory neural networks. Sensors 2020, 20, 5611. [Google Scholar] [CrossRef]
  9. Li, Y.; Wu, J.; Li, H.; Li, D.; Du, X.; Chen, Z.; Jia, F.; Hu, Q. Automatic detection of the existence of subarachnoid hemorrhage from clinical CT images. J. Med. Syst. 2012, 36, 1259–1270. [Google Scholar] [CrossRef] [PubMed]
  10. Tenny, S.; Thorell, W. Intracranial Hemorrhage. In StatPearls [Internet]; StatPearls Publishing: St. Petersburg, FL, USA, 2021. [Google Scholar]
  11. McCulloch, W.S.; Pitts, W. A logical calculus of the ideas immanent in nervous activity. Bull Math Biophys. 1943, 5, 115–133. [Google Scholar] [CrossRef]
  12. Yuh, E.L.; Gean, A.D.; Manley, G.T.; Callen, A.L.; Wintermark, M. Computer-aided assessment of head computed tomography (CT) studies in patients with suspected traumatic brain injury. J. Neurotrauma 2008, 25, 1163–1172. [Google Scholar] [CrossRef]
  13. Li, Y.-H.; Zhang, L.; Hu, Q.-M.; Li, H.-W.; Jia, F.-C.; Wu, J.-H. Automatic subarachnoid space segmentation and hemorrhage detection in clinical head CT scans. Int. J. Comput. Assist. Radiol. Surg. 2011, 7, 507–516. [Google Scholar] [CrossRef]
  14. Arbabshirani, M.R.; Fornwalt, B.K.; Mongelluzzo, G.J.; Suever, J.D.; Geise, B.D.; Patel, A.A.; Moore, G.J. Advanced machine learning in action: Identification of intracranial hemorrhage on computed tomography scans of the head with clinical workflow integration. NPJ Digit. Med. 2018, 1, 9. [Google Scholar] [CrossRef] [Green Version]
  15. Prevedello, L.M.; Erdal, B.S.; Ryu, J.L.; Little, K.J.; Demirer, M.; Qian, S.; White, R.D. Automated critical test findings identi-fication and online notification system using artificial intelligence in imaging. Radiology 2017, 285, 923–931. [Google Scholar] [CrossRef]
  16. Grewal, M.; Srivastava, M.M.; Kumar, P.; Varadarajan, S. RADnet: Radiologist level accuracy using deep learning for hem-orrhage detection in CT scans. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 281–284. [Google Scholar]
  17. Jnawali, K.; Arbabshirani, M.R.; Rao, N.; Patel, A.A. Deep 3D Convolution Neural Network for CT Brain Hemorrhage Classifi-cation. In Medical Imaging 2018: Computer-Aided Diagnosis; International Society for Optics and Photonics: Washington, DC, USA, 2018; Volume 10575, p. 105751C. [Google Scholar]
  18. Alis, D.; Alis, C.; Yergin, M.; Topel, C.; Asmakutlu, O.; Bagcilar, O.; Senli, Y.D.; Ustundag, A.; Salt, V.; Dogan, S.N.; et al. A joint convolutional-recurrent neural network with an attention mechanism for detecting intracranial hemor-rhage on noncontrast head CT. Sci. Rep. 2022, 12, 2084. [Google Scholar] [CrossRef] [PubMed]
  19. Chan, T. Computer aided detection of small acute intracranial hemorrhage on computer tomography of brain. Comput. Med. Imaging Graph. 2007, 31, 285–298. [Google Scholar] [CrossRef]
  20. Shahangian, B.; Pourghassem, H. Automatic brain hemorrhage segmentation and classification algorithm based on weighted grayscale histogram feature in a hierarchical classification structure. Biocybern. Biomed. Eng. 2016, 36, 217–232. [Google Scholar] [CrossRef]
  21. Chang, P.; Kuoy, E.; Grinband, J.; Weinberg, B.; Thompson, M.; Homo, R.; Chen, J.; Abcede, H.; Shafie, M.; Sugrue, L.; et al. Hybrid 3D/2D convolutional neural network for hemorrhage evaluation on head CT. Am. J. Neuroradiol. 2018, 39, 1609–1616. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Lee, H.; Yune, S.; Mansouri, M.; Kim, M.; Tajmir, S.H.; Guerrier, C.E.; Ebert, S.A.; Pomerantz, S.R.; Romero, J.M.; Kamalian, S.; et al. An explainable deep-learning algorithm for the detection of acute intracranial haemorrhage from small datasets. Nat. Biomed. Eng. 2019, 3, 173–182. [Google Scholar] [CrossRef] [PubMed]
  23. Nag, M.K.; Chatterjee, S.; Sadhu, A.K.; Chatterjee, J.; Ghosh, N. Computer-assisted delineation of hematoma from CT volume using autoencoder and Chan Vese model. Int. J. Comput. Assist. Radiol. Surg. 2018, 14, 259–269. [Google Scholar] [CrossRef]
  24. Anouk Stein, M.D.; Wu, C.; Carr, C.; Shih, G.; Kalpathy-Cramer, J.; Elliott, J.; Luciano Prevedello, K.; Marc Kohli, M.D.; Lungren, M.; Culliton, P.; et al. RSNA Intracranial Hemorrhage Detection. Kaggle: San Francisco, CA, USA, 2019. Available online: https://www.kaggle.com/c/rsna-intracranial-hemorrhage-detection (accessed on 14 September 2022).
  25. Mason, D. SU-E-T-33: Pydicom: An open source DICOM library. Med. Phys. 2011, 38 Pt 10, 3493. [Google Scholar] [CrossRef]
  26. Ruby, U.; Yendapalli, V. Binary cross entropy with deep learning technique for image classification. Int. J. Adv. Trends Comput. Sci. Eng. 2020, 9, 5393–5397. [Google Scholar]
  27. Zhang, Z. Improved adam optimizer for deep neural networks. In Proceedings of the 2018 IEEE/ACM 26th International Symposium on Quality of Service (IWQoS), Banff, AB, Canada, 4–6 June 2018; pp. 1–2. [Google Scholar]
  28. Gao, Y.; Glowacka, D. Deep Gate Recurrent Neural Network. In Asian Conference on Machine Learning; PMLR: Birmingham, UK, 2016; pp. 350–365. [Google Scholar]
  29. Cortes, C.; Mohri, M.; Rostamizadeh, A. L2 regularization for learning kernels. arXiv 2012, arXiv:1205.2653. [Google Scholar]
  30. Ketkar, N. Introduction to Keras. In Deep Learning with Python; Apress: Berkeley, CA, USA, 2017; pp. 97–111. [Google Scholar]
  31. Goldsborough, P. A tour of tensorflow. arXiv 2016, arXiv:1610.01178. [Google Scholar]
  32. Staudemeyer, R.C.; Morris, E.R. Understanding LSTM—A tutorial into long short-term memory recurrent neural networks. arXiv 2019, arXiv:1909.09586. [Google Scholar]
  33. Kuzmak, P.M.; Dayhoff, R.E.; Gavrilov, S.; Cebelinski, G.; Shovestul, M.L.; Casertano, A. Streamlining importation of outside prior DICOM studies into an imaging system. J. Digit. Imaging 2011, 25, 70–77. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Bhat, N.; Vidyadevi, G.; Biradar, A.K.S.; Mallya, S.S.S.; Pareek, P.K. Identi-fication of Intracranial Hemorrhage using ResNeXt Model. In Proceedings of the 2022 IEEE 2nd Mysore Sub Section International Conference (MysuruCon), Mysuru, India, 16–17 October 2022; pp. 1–5. [Google Scholar]
  35. Castro, J.S.; Chabert, S.; Saavedra, C.; Salas, R. Convolutional neural networks for detection intracranial hemorrhage in CT images. CRoNe 2019, 2564, 37–43. [Google Scholar]
  36. Mohammed, B.A.; Senan, E.M.; Al-Mekhlafi, Z.G.; Rassem, T.H.; Makbol, N.M.; Alanazi, A.A.; Almurayziq, T.S.; Ghaleb, F.A.; Sallam, A.A. Multi-Method Diagnosis of CT Images for Rapid Detection of Intracranial Hemorrhages Based on Deep and Hybrid Learning. Electronics 2022, 11, 2460. [Google Scholar] [CrossRef]
  37. Davis, V.; Devane, S. Diagnosis & classification of brain hemorrhage. In Proceedings of the 2017 International Conference on Advances in Computing, Communication and Control (ICAC3), Mumbai, India, 1–2 December 2017; pp. 1–6. [Google Scholar]
  38. Danilov, G.; Kotik, K.; Negreeva, A.; Tsukanova, T.; Shifrin, M.; Zakharova, N.; Batalov, A.; Pronin, I.; Potapov, A. Classifi-cation of Intracranial Hemorrhage Subtypes Using Deep Learning on CT Scans. Stud. Health Technol. Inf. 2020, 272, 370–373. [Google Scholar]
Figure 1. Types of hemorrhages and their respective location inside the skull [9].
Figure 1. Types of hemorrhages and their respective location inside the skull [9].
Brainsci 13 00400 g001
Figure 2. Illustration showing an example of Input/Output for the Neural Networks.
Figure 2. Illustration showing an example of Input/Output for the Neural Networks.
Brainsci 13 00400 g002
Figure 3. DICOM image viewed in 10 different window settings.
Figure 3. DICOM image viewed in 10 different window settings.
Brainsci 13 00400 g003
Figure 4. DICOM files under different window settings create a temporal sequence.
Figure 4. DICOM files under different window settings create a temporal sequence.
Brainsci 13 00400 g004
Figure 5. The working architecture of the CNN approach.
Figure 5. The working architecture of the CNN approach.
Brainsci 13 00400 g005
Figure 6. The architecture of the Conv-LSTM Model for ICH Detection.
Figure 6. The architecture of the Conv-LSTM Model for ICH Detection.
Brainsci 13 00400 g006
Figure 7. Proposed Conv-LSTM model outcome under varying epochs.
Figure 7. Proposed Conv-LSTM model outcome under varying epochs.
Brainsci 13 00400 g007
Figure 8. ROC-AUC plot for ICH and its sub-types.
Figure 8. ROC-AUC plot for ICH and its sub-types.
Brainsci 13 00400 g008
Figure 9. Comparative result analysis of Conv-LSTM with Sensitivity and Specificity.
Figure 9. Comparative result analysis of Conv-LSTM with Sensitivity and Specificity.
Brainsci 13 00400 g009
Figure 10. Comparative result analysis of Conv-LSTM with Precision and Accuracy.
Figure 10. Comparative result analysis of Conv-LSTM with Precision and Accuracy.
Brainsci 13 00400 g010
Table 1. The total number of files with and without hemorrhage (s).
Table 1. The total number of files with and without hemorrhage (s).
Files that show some hemorrhage(s)107,826
Files that do not show any hemorrhage644,708
Total files752,534
Table 2. The number of files in ICH hemorrhage.
Table 2. The number of files in ICH hemorrhage.
HemorrhageTotal Number of File(s) with This Hemorrhage Present in Them
SOME107,826
EPD3145
ITP36,115
ITV26,204
SBC35,675
SDB47,060
Table 3. The number of files of each file type per hemorrhage.
Table 3. The number of files of each file type per hemorrhage.
HemorrhageNumber of Files of Each Type for This HemorrhageTotal
12345
EPD1694923368137233145
ITP15,66414,33851129782336,115
ITV987811,61737309562326,204
SBC16,42312,91153299892335,675
SBD32,096985741129722347,060
Table 4. Different window settings.
Table 4. Different window settings.
Window NameWindow Center (WL/WC)Window Width (WW)
Defaultfrom DICOM tagsfrom DICOM tags
Brain4080
Subdural (min)50130
Subdural (mid)75215
Subdural (max)100300
Tissue (min)20350
Tissue (mid)40375
Tissue (max)60400
Bone6002800
Grey-white differentiation328
Table 5. The outcome of the proposed Conv-LSTM model.
Table 5. The outcome of the proposed Conv-LSTM model.
EpochsSensitivity/TPR (%)Specificity/TNR (%)Precision (%)Accuracy (%)
10092.0994.9094.5794.91
20093.1896.5295.1095.60
30092.7595.1095.7994.43
40094.5495.3194.9095.70
50095.2096.8096.2796.58
Table 6. ROC-AUC score for different labels.
Table 6. ROC-AUC score for different labels.
MetricLabels
SomeEPDITPITVSBCSBD
ROC-AUC Score (%)94.7493.1195.1393.0695.1195.37
Table 7. Comparative result analysis of proposed model with other state-of-art techniques.
Table 7. Comparative result analysis of proposed model with other state-of-art techniques.
AlgorithmsSensitivity/TPR (%)Specificity/TNR (%)Precision (%)Accuracy (%)
ResNexT87.8490.7894.8690.71
SVM75.1278.6479.2778.24
CNN86.6188.9487.9887.63
AlextNet + SVM92.4593.7493.6793.18
U-Net64.5487.2188.3987.50
WA-ANN 64.8772.1571.9470.27
Conv-LSTM (Proposed)93.8796.4595.2195.14
Table 8. Comparison of classification F1 score achieved by proposed technique and other compared techniques.
Table 8. Comparison of classification F1 score achieved by proposed technique and other compared techniques.
AlgorithmsSomeEPDITPITVSBCSBD
ResNexT88.7484.5189.6985.1290.7191.62
SVM74.8872.9477.1773.8479.1479.67
CNN88.9183.7487.1683.5188.3787.38
AlextNet + SVM92.1790.4592.6791.4793.0892.87
U-Net84.1782.8886.7477.9787.1886.24
WA-ANN 72.1572.1472.7471.8672.6471.45
Conv-LSTM (Proposed)94.6393.1594.8793.1695.6195.37
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rajagopal, M.; Buradagunta, S.; Almeshari, M.; Alzamil, Y.; Ramalingam, R.; Ravi, V. An Efficient Framework to Detect Intracranial Hemorrhage Using Hybrid Deep Neural Networks. Brain Sci. 2023, 13, 400. https://doi.org/10.3390/brainsci13030400

AMA Style

Rajagopal M, Buradagunta S, Almeshari M, Alzamil Y, Ramalingam R, Ravi V. An Efficient Framework to Detect Intracranial Hemorrhage Using Hybrid Deep Neural Networks. Brain Sciences. 2023; 13(3):400. https://doi.org/10.3390/brainsci13030400

Chicago/Turabian Style

Rajagopal, Manikandan, Suvarna Buradagunta, Meshari Almeshari, Yasser Alzamil, Rajakumar Ramalingam, and Vinayakumar Ravi. 2023. "An Efficient Framework to Detect Intracranial Hemorrhage Using Hybrid Deep Neural Networks" Brain Sciences 13, no. 3: 400. https://doi.org/10.3390/brainsci13030400

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop