Next Article in Journal
An ELISA Test Able to Predict the Development of Oral Cancer: The Significance of the Interplay between Steroid Receptors and the EGF Receptor for Early Diagnosis
Next Article in Special Issue
Thyroid Nodule Detection and Region Estimation in Ultrasound Images: A Comparison between Physicians and an Automated Decision Support System Approach
Previous Article in Journal
Pathomorphological Manifestations and the Course of the Cervical Cancer Disease Determined by Variations in the TLR4 Gene
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Framework for the Detection of Abnormality in Cerebral Blood Flow Velocity Using Transcranial Doppler Ultrasound

by
Naima Nasrin Nisha
1,†,
Kanchon Kanti Podder
1,†,
Muhammad E. H. Chowdhury
2,*,
Mamun Rabbani
1,
Md. Sharjis Ibne Wadud
1,
Somaya Al-Maadeed
3,
Sakib Mahmud
2,
Amith Khandakar
2 and
Susu M. Zughaier
4
1
Department of Biomedical Physics & Technology, University of Dhaka, Dhaka 1000, Bangladesh
2
Department of Electrical Engineering, Qatar University, Doha 2713, Qatar
3
Department of Computer Science and Engineering, Qatar University, Doha 2713, Qatar
4
Department of Basic Medical Sciences, College of Medicine, Qatar University, Doha 2713, Qatar
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2023, 13(12), 2000; https://doi.org/10.3390/diagnostics13122000
Submission received: 3 April 2023 / Revised: 29 May 2023 / Accepted: 5 June 2023 / Published: 8 June 2023
(This article belongs to the Special Issue Application of Deep Learning in Medical Ultrasound)

Abstract

:
Transcranial doppler (TCD) ultrasound is a non-invasive imaging technique that can be used for continuous monitoring of blood flow in the brain through the major cerebral arteries by calculating the cerebral blood flow velocity (CBFV). Since the brain requires a consistent supply of blood to function properly and meet its metabolic demand, a change in CBVF can be an indication of neurological diseases. Depending on the severity of the disease, the symptoms may appear immediately or may appear weeks later. For the early detection of neurological diseases, a classification model is proposed in this study, with the ability to distinguish healthy subjects from critically ill subjects. The TCD ultrasound database used in this study contains signals from the middle cerebral artery (MCA) of 6 healthy subjects and 12 subjects with known neurocritical diseases. The classification model works based on the maximal blood flow velocity waveforms extracted from the TCD ultrasound. Since the signal quality of the recorded TCD ultrasound is highly dependent on the operator’s skillset, a noisy and corrupted signal can exist and can add biases to the classifier. Therefore, a deep learning classifier, trained on a curated and clean biomedical signal can reliably detect neurological diseases. For signal classification, this study proposes a Self-organized Operational Neural Network (Self-ONN)-based deep learning model Self-ResAttentioNet18, which achieves classification accuracy of 96.05% with precision, recall, f1 score, and specificity of 96.06%, 96.05%, 96.06%, and 96.09%, respectively. With an area under the ROC curve of 0.99, the model proves its feasibility to confidently classify middle cerebral artery (MCA) waveforms in near real-time.

1. Introduction

Non-invasive disease detection techniques are considered to be more convenient compared to invasive diagnostic procedures. Ultrasound is one of the most common non-invasive imaging techniques, and owing to its efficacy, low cost, and zero radiation hazard, it is considered the safest imaging modality [1]. The most common uses of diagnostic ultrasound involve fetal heart monitoring in pregnant women and observation of abdominal organs and heart valves. Pulsed doppler ultrasound is an imaging technique where short bursts of ultrasonic waves are applied to the insonation area for imaging the movements of the respective organs, especially for the measurement of blood flow through the blood vessels [2]. Owing to the brain’s inability to store energy, consistent blood flow is essential for keeping up with the metabolic demands of the brain and supporting cerebral function [3,4]. This implies that an alteration in cerebral blood flow velocity can be an indication of a disruption in the systemic flow caused by any neurological disorder [5].
Doppler ultrasound is a low-cost, non-invasive diagnostic tool commonly used to diagnose a wide spectrum of neurocritical disorders [6,7]. Primarily, the diagnosis is made by measuring the change in blood flow velocity [7,8,9]. Transcranial doppler (TCD) ultrasonography, which operates on the same concept as doppler ultrasound, can be a useful indication for diagnosing neurological diseases such as artery stenosis and occlusion in patients by observing a change in the blood flow velocity profile in the relevant arteries. For TCD scans, the basal arteries are used to quantify blood flow [10].

1.1. Related Work

Transcranial doppler ultrasound (TCD) is a non-invasive tool widely used for the detection of neurological diseases [7,10]. Using a high-frequency ultrasound signal, blood flow velocities in the major basal arteries, including the middle cerebral artery (MCA) and internal carotid artery (ICA), are measured. In recent years, the potentials of deep learning and neural networks are being explored for the efficient detection of arterial diseases from doppler spectrograms. Übeyli and Güler developed a technique for detecting abnormalities in the ophthalmic artery (OA) and the internal carotid artery (ICA) by decomposing the signals using wavelet transform to express them in time–frequency format [11]. Using the Levenberg–Marquardt (LM) optimization technique on a Multilayer Perceptron (MLP) network, they achieved classification accuracies of 95.52% and 97% for OA and ICA disorder, respectively [11]. Seddik and Shawky [12] described a cost-effective screening approach for carotid artery disorders. They extracted different signal features from the frequency spectrogram after pre-processing for noise elimination. The classification accuracies achieved for normal and occlusion patterns were 91.67% and 95.85%, respectively, using the MLP classifier [12]. Wavelet transform-based spectral analysis of the doppler signal from ICA was conducted in [13] using a technique similar to [11]. An MLP was trained with LM optimization for stenosis and occlusion detection. The model achieved accuracies of 96%, 96.15%, and 96.30% for healthy subjects, subjects with arterial stenosis, and subjects with blood vessel occlusion, respectively [13]. An intima–media thickness-based plaque identification technique for early detection of stroke was reported by [14]. They trained an MLBPN with LM optimization to classify ultrasound images of the carotid artery and achieved an accuracy of 89.43%. Uğuz proposed a classification algorithm based on the Learning Vector Quantization Neural Network (LVQ NN) for classifying ICA doppler data [15]. The Burg autoregressive spectrum was utilized to derive power spectral density (PSD) characteristics. LVQ NN achieved a classification accuracy of 97.91% using the five-fold cross-validation method.
One of the issues with the existing literature on the topic is that the majority of the literature focuses on doppler ultrasound taken from the carotid artery. Since cerebral arteries transport blood to deeper parts of the brain, ultrasound pictures of the middle cerebral artery have significant clinical significance [4]. Mei et al. [16] provide a similar study in which they evaluated TCD images captured from the middle cerebral artery. The CNN VGG16 model was utilized to classify the images into the stenosis and non-stenosis groups. They reported a classification accuracy of 80%, with a sensitivity of 84% and a specificity of 86%. Individual and Recurrent Neural Networks (RNNs) were utilized to classify the MCA doppler signal and data from other basal arteries in [17]. However, the model achieved accuracy ranging between only 71.1% and 75.89% for MCA ultrasound signals. Li et al. created a neural network-based system for detecting arterial stenosis [18]. They have utilized a synthetic peripheral pulse wave-based dataset that represents varying degrees of vascular stenosis. The disadvantage of their model, however, was that it was greatly dependent on stenosis severity. For less severe stenosis, the model’s accuracy is weak, rendering it unreliable for early diagnosis of stenosis.

1.2. Motivation

TCD signals are typically corrupted with speckle noise and motion artefacts since the quality of the signal depends heavily on the skill set of the operator [8,19]. Therefore, signal reconstruction techniques [20] or other noise reduction methods [21,22] can be employed to enhance the signal quality. For this reason, the successful classification of a variation in blood flow velocity requires the utilization of a signal-processing method that is both efficient and reliable. Owing to the efficiency and reliability of computer-aided detection and classification techniques, these are preferred over other detection techniques [23].
The unique morphologies of TCD waveform acquired from different subjects can be a significant indication of the presence of any neurological conditions [8]. This justifies our approach of categorizing TCD waveform being a classification problem. Biomedical signal classification using a decision-based algorithm [24,25] and deep learning [26,27,28] approach is gaining the attention of researchers for various tasks. Recent studies have proposed several approaches to address the TCD waveform classification problem [23,29,30,31]. However, the problem domain is still a work in progress due to the lack of comprehensiveness of the results. The use of neural networks is a significant method for signal processing, especially when it comes to the classification of TCD ultrasound waveforms [11,31,32,33].
By employing a deep learning model, signal classification can be made easy and reliable and be used in a clinical setting. With this in mind, we have designed a study that will contribute to the medical field as follows:
  • A novel TCD ultrasound waveform classification system that incorporates cerebral blood flow velocity (CBFV) waveform estimation, data cleaning and segmentation, and classification for classifying data from doppler ultrasounds in both healthy subjects and intensive care unit subjects.
  • Two novel SelfONN architectures, (a) a 1D version of ResNet18 and (b) a 1D version of ResNet architecture with a multiheaded attention layer, have been proposed for 1D binary classification, where TCD ultrasound can be used to identify ICU (in this study, “ICU” is an umbrella term used for representing MCA waveforms from hydrocephalus, traumatic brain injury, and intraparenchymal or subarachnoid hemorrhage patients) patients from healthy subjects by analyzing the maximal cerebral blood flow velocity (CBFV) waveforms.
This paper is organized into five sections and the rest of the paper is organized as follows. Section 2 elaborates on the methods and materials used along with the experimental setup. The experimental results and discussion are presented in Section 3. Section 4 presents the concluding remarks and finally, Section 5 discusses the limitations and future scopes of this study.

2. Materials and Methods

This study uses a range of different techniques for data pre-processing and classification of healthy subject and ICU patient data. This section presents a detailed discussion of the step-by-step procedure followed during the study. The overall procedure is summarized in Figure 1. The data-acquisition process is completed in the literature [9] and the other processes are investigated in this study.

2.1. Overview of the Dataset

This study uses a transcranial doppler (TCD) ultrasound dataset collected from IEEE Dataport (https://ieee-dataport.org/open-access/transcranial-doppler-ultrasound-database-philips-cx50-ultrasound-system, accessed on 3 April 2023), which was provided by Wadehn et al. [9]. The dataset comprises healthy subject data as well as data from patients with known neurocritical disorders. Table 1 contains the overview of the dataset, and the detailed protocols for clinical data collection can be found in [9,34].

2.2. Signal Extraction

To acquire the transcranial doppler (TCD) ultrasound signals, the segments of the MCA were insonated with short pulses of ultrasound waves with a carrier frequency of 1.75 MHz. The ultrasound pulses were reflected from the moving red blood cells and captured by the transducer as TCD echo signals. Wadehn et al. proposed a flow velocity estimation algorithm that takes the TCD ultrasound signals and returns a 1D maximal flow velocity waveform [9]. Using STFT (short-time Fourier transform), envelope tracking, and some post-processing, the TCD spectrogram was transformed into a 2D spectrogram. A black-and-white image of the spectrogram was created using an envelope tracing algorithm also developed by the same author. When the spectrogram was binarized, the speckle noise was eliminated using a 2D median filter with a kernel size of 0.03 s (horizontally) and 5 cm s−1 (vertically). The maximal flow velocity envelope was detected via an adaptive threshold method due to the overlap between the signal-carrying portion and the noise of the black-and-white spectrogram [35]. The maximal flow velocity envelope is traced on a time sample basis employing two physiological sanity checks, which were determined from the signal quality index. Different stages of signal extraction are shown in Figure 2.

2.3. Manual Data Annotation

In this investigation, we used the technique provided in [9] to extract MCA doppler ultrasound signals from TCD ultrasound images. Several artefacts and signal cuts of varying lengths were found in the retrieved data. After removing the signal segments containing signal cuts, the remaining signals were segmented in 1024 segment lengths at a sampling frequency of 217 Hz.
Afterwards, these segments were manually annotated by human supervision. The segments were annotated into two categories, clear segments and noisy segments, and Figure 3 represents four samples from the Healthy and ICU classes. The signal portions that have no visible artefacts or distortions were categorized as Clear Signals. Some corrupted segments with NaN values, plane lines, or no sign of physiological information were eliminated from the dataset during manual annotation. From the remaining signal portions, the cleanest or negligible distortions were labelled as Corrupted Signals. A total of 5468 segments were created from the Healthy and ICU classes during the segmentation process with 1024 segment lengths. For the Healthy class, the numbers of clear and corrupted segments were 1501 and 152, respectively, whereas they were 1742 and 2090 for the ICU class, respectively. Details of dataset segmentation and labelling can be seen in Figure 4.

2.4. Model Architecture

For the classification of signals into the two categories, “Healthy” and “ICU patients”, two Self-Organizing Operational Neural Network (Self-ONN)-based novel 1D classification models were proposed. The details about Self-ONN and the architecture of the 1D classification models are described in the following subsections.

2.4.1. Self-ONN

To surpass the drawbacks of Convolutional Neural Networks (CNNs), Operational Neural Networks (ONNs) were proposed in [36]. Conventional CNNs function through linear convolutional operators in their neurons and layers for feature learning. According to Equation (1) [36], the output of the k t h neuron in the l t h layer of an one-dimensional CNN can be expressed as:
x k l = b k l + i = 0 N l 1 x i k l
where, b k l is the bias associated with the neuron and x i k l is the l t h layer’s k t h neuron output, which can be further expressed as Equation (2):
x i k l ( m ) = C o n v 1 D w i k , y i l 1 r = 0 k 1 w i k l ( r ) y i l 1 ( m + r )
where, w i k l represents the weight of the kernel connecting the i t h neuron of the ( l 1 ) t h layer to the k t h neuron of the l t h layer, y i l 1 represents the ( l 1 ) t h layer’s i t h neuron output, and ‘m’ and ‘r’ are convolutional operators. ONNs can address the homogeneity issues of CNNs by employing heterogeneous neurons through non-linear convolutional operators which can learn complex model functions with minimal network complexity [37]. As reflected in Equation (3), every single neuron in the ONNs can be assigned unique nodal ( ψ ) and pool ( P ) operators.
x i k l ¯ ( m ) = P k l ψ k l w i k l ( r ) , y i l 1 ( m + r ) r = 0 k 1
Regardless of having non-linear operators, early ONNs were rigid due to dependence on a bunch of pre-set operators by the user. To solve this challenge, Kiranyaz et al. [38] proposed Self-Organizing ONNs or Self-ONNs in short, which can automatically determine the best set of problem-specific non-linear operators during training. Self-ONN uses Taylor series approximation for the non-linear transformation of each generative neuron to reach an even higher level of diversity and flexibility. Based on [38,39,40], the contribution of the i t h neuron in generating the feature map x i k l from the ( l 1 ) t h layer to the l t h layer of a Self-ONN model can be expressed by Equation (4) as follows:
x i k l ~ m = r = 0 K 1 q = 1 Q ψ k l ~ w i k l Q r , y i l 1 m + r = r = 0 K 1 q = 1 Q w i k l Q r , q y i l 1 m + r q q = 1 Q C o n v 1 D w i k l Q , y i l 1 q
where, w i k l Q is the K × Q dimensional kernel matrix between the ith neuron from the ( l 1 ) t h layer and the k t h neuron at the l t h layer. Here, the hyperparameter Q can be tweaked to control the degree of Taylor series approximation while w i k l Q is the learnable kernel, unlike CNNs and ONNs. Finally, the output of a single neuron can be formulated as in Equation (5):
x k l ~ = b k l + i = 0 N l 1 x i k l ~
Mentionable that, with the Q = 1 setting, a Self-ONN acts like a CNN as there is no non-linearity in the first term of the Taylor series approximation. SelfMLP, which is a variant of MLP, can also be designed from the SelfONN layer.

2.4.2. Self-ResNet18

The Self-ResNet18 model proposed in this study adopts the widely used CNN-based most lightweight version of the ResNet models proposed by He et al. [41] in 2015. Being one of the oldest deep learning architectures, ResNet models have been used in many studies to solve various types of problems. In this study, we replace the CNN layers of the ResNet18 model with the Self-ONN layers to construct the architecture of the proposed Self-ResNet18. The architecture of Self-ResNet18 has been drawn in Figure 5.
Our proposed Self-ResNet18 model developed in PyTorch contains 18 Self-ONN layers with Q = 3, as depicted in Figure 5. We started with 8 kernels or filters in the initial layer, which was doubled up after each Self-ONN layer with a stride of two. Based on the original implementation strategies [38], each Self-ONN block was followed by a “batch normalization” layer and a “tanh” activation function. The batch normalization layer was avoided during the down-sampling process. After the last residual block, which contained 64 filters, we implemented an average-pooling layer followed by a flattening layer to flatten the features into a single array. Then we passed the feature vectors through a fully connected (FC) dense layer that performed a linear transformation of the incoming features using a linear activation function [42]. This multi-layered perceptron (MLP)-based dense layer contained filters equal to the number of output classes (which is two in this case, “Healthy” and “ICU patients”, making it a binary classification problem) to aid the classification process based on refined features from previous layers. Finally, we implemented a “Log-Softmax” layer [43] as the final activation function to help in the classification process.

2.4.3. Self-ResAttentioNet18

This study proposes an attention-based version of the Self-ResNet18 model, which is called the Self-ResAttentioNet18. It is used to classify MCA signals acquired from subjects into “Healthy” and “ICU” categories. Figure 6 represents the illustration of Self-ResAttentioNet18, where every two SelfONN layers after the first one is a Residual Block. In this architecture, a multi-head attention layer of several heads of 2 was added to each alternative Self-Residual block. The attention layer requires query, key, and value. Every input feature of the residual block or the identity features was used as a query and value, while the output of the residual block was used as a key. The jointly attended feature was further added to the residual block output and propagated to the next block. The Self-ResAttentioNet18 is identical to the Self-ResNet18 architecture illustrated in Figure 5, with only four multi-head attention layers added to it to jointly attend to information from the identity feature and the output feature.

2.5. Experimental Setup

The deep learning models discussed above were investigated on the dataset provided by [9] after preprocessing. For the classification purpose, the healthy subjects were labelled as ‘Healthy’, while the subjects with hydrocephalus, traumatic brain injury, and intraparenchymal or subarachnoid hemorrhage were labelled as ‘ICU Patients’. Thus, the problem was considered a binary classification problem. Each study was conducted twice during training, validation, and testing, once with the standard SelfRes-Net18 model and once with an enhanced Self-ResAttentioNet18 model that included multiheaded-attention layers. Both models’ efficacies were evaluated by using q-values in the range [1,3,5].
To minimize bias and data scarcity, 5-fold cross-validation was employed with 80% signal overlapping. The splitting and fold creation was performed based on session IDs as shown in Table 2. A percentage of 10% of the overall training sample was moved to the validation set to validate the model learning process during training. The objective of session-wise splitting was to minimize data bias during training and validation and testing, meaning the data used for training will not be present in the test data. This method of the 5-fold training, validation, and test data splitting ensures data stratification [44].
The models used in this study use a Python 3.7 environment and several PyTorch libraries. For training and testing, Google-Colab was used along with its high-performing resources. A 16 GB Tesla T4 GPU was used for this study. For the classification of transcranial doppler (TCD) ultrasound signals, the following hyperparameters in Table 3 were used during the model training and test phases.

2.6. Performance Metrics

For the quantitative analysis of the classification model, several parameters were calculated. In this work, signal classification was mainly done from the waveform morphologies of Heathy and ICU subjects’ TCD scans. The performance of the proposed model was evaluated using measures such as precision, overall accuracy, recall, specificity, and f1 score, with a 90% confidence interval (CI) in addition to receiver operating characteristic (ROC) curves. Due to the difference in sample sizes between Healthy and ICU subjects, weighted recall, precision, f1 score, and specificity measures were determined. Equations (6)–(10) describe weighted precision, sensitivity or recall, specificity, F1 score, and overall accuracy quantitatively.
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
S p e c i f i c i t y = T N F P + T N
F 1   s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
O v e r a l l   A c c u r a c y = T P T P + F P + T N + F N
Here, TP = true positive, FP = false positive, TN = true negative, and FN = false negative.
A receiver operating characteristic (ROC) curve is a graph representing how well a binary classifier system works as the threshold for discrimination is varied. This graph shows the true positive rate (sensitivity) against the false positive rate (1-specificity) for different threshold settings.
F a s l e   P o s i t i v e   R a t e = F N T P + F N
The ROC curve is a useful tool for figuring out how well a binary classifier works because it shows how the true positive rate and the false positive rate work together. The AUC of a perfect classifier is 1, while the AUC of a classifier that guesses at random is 0.5. The ROC curve can be used to evaluate the performance of a classifier for imbalanced classes. The curve can also be used to find the optimum threshold for a classifier. The point on the ROC curve that is closest to the top left corner indicates the best balance between true positives and false positives.

3. Results and Discussion

The goal of the study is to identify the optimum model configuration for efficient classification. For evaluating the performance of the proposed models (Self-ResNet18 and Self-ResAttentioNet18), different performance metrics were calculated. The performance metrics achieved from different models are discussed in the following subsections.

3.1. Accuracy-Based Comparison

The five-fold cross-validation produces classification accuracy across the five folds. For similar experimental setups, different models produce different performance metrics. The accuracy-based comparison of the proposed Self-ResNet18 model without the attention layer and Self-ResAttentioNet18 with the attention layer is given in Figure 7. The best accuracy achieved was 96.05% which was from the Self-ResAttentioNet18_Q1 model. Additionally, the five-fold accuracy variance is also the minimum among the other models. The five-fold lowest accuracy 92.97% is achieved by Self-ResAttentioNet18_Q5. From Figure 7, it is evident that Self-ResAttentioNet18_Q1 achieved nearly the same accuracy across the five folds.
The other performance metrics used in this study are overall accuracy, weighted precision, weighted recall, weighted f1 score, and weighted specificity. The performance metrics are recorded in Table 4 for the six different configurations of the two model architectures. The efficiency in classifying the healthy subjects’ signals and ICU patients’ signals was evaluated for q-values of 1, 3, and 5. For the base Self-ResNet18 models, the Self-ResNet18_Q3 model shows superior performance among the three variations. This model achieved an accuracy of 94.91%, 94.92% precision, 94.91% f1 score, and 94.88% specificity. The experiments also were carried out by adding attention layers in the Self-ResNet18 model with the expectation that the overall accuracy would increase significantly. However, the accuracy achieved for the Self-ResAttentioNet18_Q1 model was 2.74% higher than the Self-ResNet18 models. This model achieved an accuracy of 96.05%, 96.06% precision, 96.06% f1 score, and a specificity of 96.09%. Based on the evaluation metrics such as overall accuracy, precision, recall, f1 score, and specificity (as shown in Table 4), Self-ResAttentioNet18 Q1 stands out as the top-performing model among the six variants of Self-ResNet18 and Self-ResAttentioNet18.
Learning curves of a deep learning model, especially the accuracy curves and loss curves, indicate how well the model is performing during training, validation, and testing. The accuracy curve plots the accuracy of the model for training, validation, and test data, while the loss curve plots the loss of the model for training, validation, and test data. The loss curve illustrates the model’s ability to minimize the gap between the predicted and the actual outputs, whilst the accuracy curve demonstrates the model’s ability in learning the features of the data. Both can indicate whether or not the model exhibits overfitting or underfitting. For our models, the top two best-performing variants are Self-ResAttentioNet18_Q1 and Self-ResNet18_Q3.
Figure 8 shows the learning curves of the Self-ResAttentioNet18_Q1 model, which achieved the highest accuracy for fold 2, among all five folds. The nature of the learning curves shows that the model was not overfitting or underfitting since the training accuracy reaches saturation after gradually improving up to a certain number of epochs. The absence of plateaus in training and validation accuracy points to a deep learning model that is well-fitted. Similar trends can be seen from all the other learning curves presented in Supplementary Tables S1–S6. After analyzing the learning curves for all the variants of our model, Self-ResAttentioNet18_Q1 was found to be the best performing model for session-independent classification of middle cerebral artery doppler ultrasound classification.

3.2. ROC Curve and Confusion Matrix-Based Comparison

The receiver operator characteristic (ROC) curve is an important metric for comparing the performance of different binary classification models. The area under the ROC curve (AUC) is a common way to measure how well a classifier does the classification. While a confusion matrix presents corresponding instances of all four TP, TN, FP, and FN, the ROC curve graphically represents the true positive rate against the false positive rate, which can be calculated from the instances of the confusion matrix. From the accuracy-based comparison, it was found that the Self-ResAttentioNet18_Q1 was the best performing model with 96.05% accuracy. The confusion matrix and the ROC curve of this model are presented in Figure 9. The confusion matrix and the ROC curve of the other models can be found in Supplementary Figures S1–S12. The figures show that compared to other models, Self-ResAttentioNet18_Q1 performs better for the classification of healthy subjects and ICU subjects based on the maximal blood flow velocity waveform. The model correctly predicted 95.51% of all the ICU instances with an area of 0.99 under the ROC curve.

3.3. Comparison with Respect to Previous Literature

To evaluate the performance of our model, we need to compare it with the existing literature. However, the dataset used in this study has no other associated literature that worked on signal classification. Therefore, the previous literature that worked on a similar problem was used for the comparison. From Table 5, the previous best performing model reported by [45] achieved 89.17% classification accuracy. However, the model proposed in this study surpasses the previous best performing model by achieving a classification accuracy of 96.05% with recall and specificity of 96.05% and 96.09%, respectively.

4. Conclusions

In the classification of biomedical signals, deep learning models have been shown to perform exceptionally well. The prognosis and diagnosis of diseases are greatly aided by this type of study. A deep learning model trained on the transcranial doppler (TCD) ultrasound signal, or more specifically MCA waveform, was proposed to be utilized in a binary classification system in this research. Self-ResNet18 and Self-ResAttentioNet18, both based on the SelfONN and ResNet architecture, were proposed in this research. Each deep learning application on biomedical signals needs rigorous testing across many evaluation matrices to guarantee its success. Among the six model versions, Self-ResAttentioNet18_Q1 had the highest classification accuracy at 96.05%, along with the highest recall (96.05%) and the highest specificity (96.09%). A comparative analysis of our proposed model with the existing literature [17,18,45,46] in the normal vs. abnormal classification using ICA or MCA waves has been done in this study to evaluate the performance of Self-AttentioNet18 against contemporary existing models. From that analysis, it can be concluded that the accuracy of the Self-AttentioNet18 in classifying healthy subjects and traumatic brain injured subjects is 6.88% greater than the existing state-of-the-art result [45]. Since both of these studies focus on ultrasound signals from the MCA, the generalizability of our study is also validated. These findings provide more evidence of Self-ResAttentioNet18 for the effectiveness of classifying MCA waveform in the ‘Healthy’ vs. ‘ICU’ classification task. To avoid the data-leakage issue, the entire investigation was planned out using a session-independent technique. This method enables a model to be evaluated using data that were not used during training. The Self-ResAttentioNet18 Q1 performed at an AUC of 0.99 and a classification accuracy of 95.51% when tested on data that had not previously been seen. Such a highly performing deep learning model has the prospect of being used for classifying MCA waveforms in real-time alongside the diagnosis of the targeted patients. This study thus concludes the medical importance of utilizing a Self-ONN-based classification model to classify TCD ultrasonography signals into ‘Healthy’ and ‘ICU’ classifications.

5. Limitations and Future Work

Although this study has the potential to detect neurological diseases in real-time, a larger, more diverse dataset can enhance the robustness and reliability of the outcomes. Further studies can be conducted on external datasets as well to assess the generalizability of the model. While this study considers only the signals from the MCA, as a future prospect of this work, signals from other basal arteries such as the internal carotid artery (ICA) can also be investigated and compared to the outcomes using only the MCA signals. The need for human supervision in manual annotation of signals can be eliminated by developing an automated annotator, which will categorize the signals based on their waveform morphologies. In order to improve the signal quality and hence boost the accuracy of the classifier, we plan to explore the signal reconstruction techniques by noise reduction with the use of the proposed model. The use of the proposed Self-AttentioNet18 model can also be explored in the domains of object detection [47] and segmentation [48]. While this study presents promising results and demonstrates the effectiveness of the proposed deep learning models for the classification of TCD ultrasound signals in neurological diseases, it also highlights avenues for further improvement and exploration.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/diagnostics13122000/s1, Figure S1: Confusion matrix for Self-ResNet18_Q1 model; Figure S2: ROC curve of Self-ResNet18_Q1 model; Figure S3: Confusion matrix for Self-ResNet18_Q3 model; Figure S4: ROC curve of Self-ResNet18_Q3 model; Figure S5: Confusion matrix of Self-ResNet18_Q5 model; Figure S6: ROC curve of Self-ResNet18_Q5 model; Figure S7: Confusion matrix of Self-ResAttentioNet18_Q1 model; Figure S8: ROC curve of Self-ResAttentioNet18_Q1 model; Figure S9: Confusion matrix of Self-ResAttentioNet18_Q3 model; Figure S10: ROC curve of Self-ResAttentioNet18_Q3 model; Figure S11: Confusion matrix of Self-ResAttentioNet18_Q5 model; Figure S12: ROC curve of Self-ResAttentioNet18_Q5 model; Table S1: Fold-wise learning curves of the Self-ResNet18_Q1 model; Table S2: Fold-wise learning curves of the Self-ResNet18_Q3 model; Table S3: Fold-wise learning curves of the Self-ResNet18_Q5 model; Table S4: Fold-wise learning curves of the Self-ResAttentionNet18_Q1 model; Table S5: Fold-wise learning curves of the Self-ResAttentionNet18_Q3 model; Table S6: Fold-wise learning curves of the Self-ResAttentionNet18_Q5 model.

Author Contributions

Conceptualization, N.N.N., K.K.P., M.E.H.C., M.R. and M.S.I.W.; Data curation, N.N.N., K.K.P., S.M. and A.K.; Formal analysis, N.N.N., K.K.P. and S.M.; Funding acquisition M.E.H.C., S.A.-M. and S.M.Z.; Investigation, N.N.N., K.K.P., S.M. and A.K.; Methodology, M.E.H.C., M.R., M.S.I.W. and S.A.-M.; Project administration, M.E.H.C., S.A.-M. and S.M.Z.; Resources, M.E.H.C., A.K., M.R. and M.S.I.W.; Supervision, M.E.H.C., M.R., M.S.I.W., S.A.-M. and S.M.Z.; Validation, M.E.H.C., S.A.-M. and S.M.Z. All authors equally contributed to the original paper drafting, reviewing, and editing of the manuscript and agreed to submit the current version of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was made possible by the High Impact grant of Qatar University # QUHI-CENG-22_23-548 and student grant: QUST-1-CENG-2023-796. The statements made herein are solely the responsibility of the authors.

Institutional Review Board Statement

This study involves a dataset that has been collected between the time period of 2016 and 2020. The authors of the dataset declared that they have acquired approval from the Institutional Review Board of Massachusetts Institute of Technology (MIT) and Boston Medical Center (BMC) [9], from where the data was collected.

Informed Consent Statement

The authors of the dataset declared that informed consent from the patients or their authorized guardians was obtained before data collection.

Data Availability Statement

The raw TCD data was acquired from IEEE Dataport from the following link: https://ieee-dataport.org/open-access/transcranial-doppler-ultrasound-database-philips-cx50-ultrasound-system (accessed on 3 April 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Szabo, T.L. Diagnostic Ultrasound Imaging: Inside Out; Academic Press: Cambridge, MA, USA, 2004. [Google Scholar]
  2. Shung, K.K. Diagnostic Ultrasound: Imaging and Blood Flow Measurements; CRC Press: Boca Raton, FL, USA, 2005. [Google Scholar]
  3. Birdsill, A.C.; Carlsson, C.M.; Willette, A.A.; Okonkwo, O.C.; Johnson, S.C.; Xu, G.; Oh, J.M.; Gallagher, C.L.; Koscik, R.L.; Jonaitis, E.M.; et al. Low cerebral blood flow is associated with lower memory function in metabolic syndrome. Obesity 2013, 21, 1313–1320. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Jarrett, C.L.; Shields, K.L.; Broxterman, R.M.; Hydren, J.R.; Park, S.H.; Gifford, J.R.; Richardson, R.S. Imaging transcranial Doppler ultrasound to measure middle cerebral artery blood flow: The importance of measuring vessel diameter. Am. J. Physiol. Integr. Comp. Physiol. 2020, 319, R33–R42. [Google Scholar] [CrossRef] [PubMed]
  5. Ozturk, E.D.; Tan, C.O. Human cerebrovascular function in health and disease: Insights from integrative approaches. J. Physiol. Anthropol. 2018, 37, 1–11. [Google Scholar] [CrossRef] [Green Version]
  6. Christou, I.; Felberg, R.A.; Demchuk, A.M.; Grotta, J.C.; Burgin, W.; Malkoff, M.; Alexandrov, A.V. A broad diagnostic battery for bedside transcranial Doppler to detect flow changes with internal carotid artery stenosis or occlusion. J. Neuroimaging 2001, 11, 236–242. [Google Scholar] [CrossRef]
  7. Thorpe, S.G.; Thibeault, C.M.; Wilk, S.J.; O’brien, M.; Canac, N.; Ranjbaran, M.; Devlin, C.; Devlin, T.; Hamilton, R.B. Velocity curvature index: A novel diagnostic biomarker for large vessel occlusion. Transl. Stroke Res. 2019, 10, 475–484. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Thorpe, S.G.; Thibeault, C.M.; Canac, N.; Jalaleddini, K.; Dorn, A.; Wilk, S.J.; Devlin, T.; Scalzo, F.; Hamilton, R.B. Toward automated classification of pathological transcranial Doppler waveform morphology via spectral clustering. PLoS ONE 2020, 15, e0228642. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Wadehn, F.; Heldt, T. Adaptive Maximal Blood Flow Velocity Estimation From Transcranial Doppler Echos. IEEE J. Transl. Eng. Health Med. 2020, 8, 1–11. [Google Scholar] [CrossRef] [PubMed]
  10. Aaslid, R.; Markwalder, T.-M.; Nornes, H. Noninvasive transcranial Doppler ultrasound recording of flow velocity in basal cerebral arteries. J. Neurosurg. 1982, 57, 769–774. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Übeyli, E.D.; Güler, İ. Improving medical diagnostic accuracy of ultrasound Doppler signals by combining neural network models. Comput. Biol. Med. 2005, 35, 533–554. [Google Scholar] [CrossRef]
  12. Seddik, A.F.; Shawky, D.M. A low-cost screening method for the detection of the carotid artery diseases. Knowl.-Based Syst. 2013, 52, 236–245. [Google Scholar] [CrossRef]
  13. Übeyli, E.D.; Güler, İ. Wavelet-based neural network analysis of internal carotid arterial Doppler signals. J. Med. Syst. 2006, 30, 221–229. [Google Scholar] [CrossRef] [PubMed]
  14. Samiappan, D.; Chakrapani, V. Classification of carotid artery abnormalities in ultrasound images using an artificial neural classifier. Int. Arab J. Inf. Technol. 2016, 13, 756–762. [Google Scholar]
  15. Uğuz, H. Detection of carotid artery disease by using learning vector quantization neural network. J. Med. Syst. 2012, 36, 533–540. [Google Scholar] [CrossRef] [PubMed]
  16. Mei, Y.J.; Hu, R.T.; Lin, J.; Xu, H.Y.; Wu, L.Y.; Li, H.P.; Ye, Z.M.; Qin, C. Diagnosis of Middle Cerebral Artery Stenosis Using Transcranial Doppler Images Based on Convolutional Neural Network. World Neurosurg. 2022, 161, e118–e125. [Google Scholar] [CrossRef]
  17. Seera, M.; Lim, C.P.; Tan, K.S.; Liew, W.S. Classification of transcranial Doppler signals using individual and ensemble recurrent neural networks. Neurocomputing 2017, 249, 337–344. [Google Scholar] [CrossRef]
  18. Li, Z.; He, W. Stenosis diagnosis based on peripheral arterial and artificial neural network. Netw. Model. Anal. Health Inform. Bioinform. 2021, 10, 1–21. [Google Scholar] [CrossRef]
  19. Sharma, S.; Lubrica, R.J.; Song, M.; Vandse, R.; Boling, W.; Pillai, P. The role of transcranial Doppler in cerebral vasospasm: A literature review. In Subarachnoid Hemorrhage: Neurological Care and Protection; Springer: Berlin/Heidelberg, Germany, 2020; pp. 201–205. [Google Scholar]
  20. Podder, K.K.; Chowdhury, M.E.; Al-Maadeed, S.; Nisha, N.N.; Mahmud, S.; Hamadelneil, F.; Almkhlef, T.; Aljofairi, H.; Mushtak, A.; Khandakar, A.; et al. Deep learning-based middle cerebral artery blood flow abnormality detection using flow velocity waveform derived from transcranial Doppler ultrasound. Biomed. Signal Process. Control. 2023, 85, 104882. [Google Scholar] [CrossRef]
  21. Hossain, S.; Mahmud, S.; Khandakar, A.; Al-Emadi, N.; Chowdhury, F.A.; Bin Mahbub, Z.; Reaz, M.B.I.; Chowdhury, M.E.H. MultiResUNet3+: A Full-Scale Connected Multi-Residual UNet Model to Denoise Electrooculogram and Electromyogram Artifacts from Corrupted Electroencephalogram Signals. Bioengineering 2023, 10, 579. [Google Scholar] [CrossRef]
  22. Mahmud, S.; Hossain, S.; Chowdhury, M.E.H.; Reaz, M.B.I. MLMRS-Net: Electroencephalography (EEG) motion artifacts removal using a multi-layer multi-resolution spatially pooled 1D signal reconstruction network. Neural Comput. Appl. 2023, 35, 8371–8388. [Google Scholar] [CrossRef]
  23. Alam, J.; Hassan, M.; Khan, A.; Chaudhry, A. Robust fuzzy RBF network based image segmentation and intelligent decision making system for carotid artery ultrasound images. Neurocomputing 2015, 151, 745–755. [Google Scholar] [CrossRef]
  24. Shuzan, N.I.; Chowdhury, M.E.; Reaz, M.B.I.; Khandakar, A.; Abir, F.F.; Faisal, A.A.; Ali, S.H.M.; Bakar, A.A.A.; Chowdhury, M.H.; Mahbub, Z.B.; et al. Machine learning-based classification of healthy and impaired gaits using 3D-GRF signals. Biomed. Signal Process. Control. 2023, 81, 104448. [Google Scholar] [CrossRef]
  25. Rahman, A.; Chowdhury, M.E.H.; Khandakar, A.; Kiranyaz, S.; Zaman, K.S.; Reaz, M.B.I.; Islam, M.T.; Ezeddin, M.; Kadir, M.A. Multimodal EEG and keystroke dynamics based biometric system using machine learning algorithms. IEEE Access 2021, 9, 94625–94643. [Google Scholar] [CrossRef]
  26. Abir, F.F.; Chowdhury, M.E.; Tapotee, M.I.; Mushtak, A.; Khandakar, A.; Mahmud, S.; Hasan, A. PCovNet+: A CNN-VAE anomaly detection framework with LSTM embeddings for smartwatch-based COVID-19 detection. Eng. Appl. Artif. Intell. 2023, 122, 106130. [Google Scholar] [CrossRef] [PubMed]
  27. Chowdhury, M.H.; Shuzan, N.I.; Chowdhury, M.E.H.; Reaz, M.B.I.; Mahmud, S.; Al Emadi, N.; Ayari, M.A.; Ali, S.H.M.; Bakar, A.A.A.; Rahman, S.M.; et al. Lightweight End-to-End Deep Learning Solution for Estimating the Respiration Rate from Photoplethysmogram Signal. Bioengineering 2022, 9, 558. [Google Scholar] [CrossRef]
  28. Faisal, A.A.; Chowdhury, M.E.H.; Bin Mahbub, Z.; Pedersen, S.; Ahmed, M.U.; Khandakar, A.; Alhatou, M.; Nabil, M.; Ara, I.; Bhuiyan, E.H.; et al. NDDNet: A deep learning model for predicting neurodegenerative diseases from gait pattern. Appl. Intell. 2023, 1–13. [Google Scholar] [CrossRef]
  29. Campadelli, P.; Casiraghi, E.; Pratissoli, S. A segmentation framework for abdominal organs from CT scans. Artif. Intell. Med. 2010, 50, 3–11. [Google Scholar] [CrossRef]
  30. Hassan, M.; Chaudhry, A.; Khan, A.; Kim, J.Y. Carotid artery image segmentation using modified spatial fuzzy c-means and ensemble clustering. Comput. Methods Programs Biomed. 2012, 108, 1261–1276. [Google Scholar] [CrossRef]
  31. Wright, I.A.; Gough, N.A.; Rakebrandt, F.; Wahab, M.; Woodcock, J.P. Neural network analysis of Doppler ultrasound blood flow signals: A pilot study. Ultrasound Med. Biol. 1997, 23, 683–690. [Google Scholar] [CrossRef]
  32. Ceylan, M.; Ceylan, R.; Dirgenali, F.; Kara, S.; Özbay, Y. Classification of carotid artery Doppler signals in the early phase of atherosclerosis using complex-valued artificial neural network. Comput. Biol. Med. 2007, 37, 28–36. [Google Scholar] [CrossRef]
  33. Ceylan, M.; Ceylan, R.; Özbay, Y.; Kara, S. Application of complex discrete wavelet transform in classification of Doppler signals using complex-valued artificial neural network. Artif. Intell. Med. 2008, 44, 65–76. [Google Scholar] [CrossRef]
  34. Jaishankar, R.; Fanelli, A.; Filippidis, A.; Vu, T.; Holsapple, J.; Heldt, T. A spectral approach to model-based noninvasive intracranial pressure estimation. IEEE J. Biomed. Health Informatics 2019, 24, 2398–2406. [Google Scholar] [CrossRef] [PubMed]
  35. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef] [Green Version]
  36. Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A survey of convolutional neural networks: Analysis, applications, and prospects. IEEE Trans. Neural Networks Learn. Syst. 2021, 33, 6999–7019. [Google Scholar] [CrossRef] [PubMed]
  37. Kiranyaz, S.; Ince, T.; Iosifidis, A.; Gabbouj, M. Operational neural networks. Neural Comput. Appl. 2020, 32, 6645–6668. [Google Scholar] [CrossRef] [Green Version]
  38. Kiranyaz, S.; Malik, J.; Ben Abdallah, H.; Ince, T.; Iosifidis, A.; Gabbouj, M. Self-organized operational neural networks with generative neurons. Neural Netw. 2021, 140, 294–308. [Google Scholar] [CrossRef]
  39. Malik, J.; Devecioglu, O.C.; Kiranyaz, S.; Ince, T.; Gabbouj, M. Real-time patient-specific ecg classification by 1d self-operational neural networks. IEEE Trans. Biomed. Eng. 2021, 69, 1788–1801. [Google Scholar] [CrossRef]
  40. Malik, J.; Kiranyaz, S.; Gabbouj, M. Self-organized operational neural networks for severe image restoration problems. Neural Networks 2021, 135, 201–211. [Google Scholar] [CrossRef]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  42. Linear. PyTorch 1.13 Documentation. 2022. Available online: https://pytorch.org/docs/stable/generated/torch.nn.Linear.html (accessed on 19 January 2023).
  43. Logsoftmax. PyTorch 1.13 Documentation. 2022. Available online: https://pytorch.org/docs/stable/generated/torch.nn.LogSoftmax.html (accessed on 19 January 2023).
  44. Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. In Proceedings of the Appears in the International Joint Conference on Articial Intelligence (IJCAI), Montreal, QC, Canada, 20–25 August 1995; pp. 1137–1145. [Google Scholar]
  45. Elzaafarany, K.; Aly, M.H.; Kumar, G.; Nakhmani, A. Cerebral artery vasospasm detection using transcranial doppler signal analysis. J. Ultrasound Med. 2019, 38, 2191–2202. [Google Scholar] [CrossRef]
  46. Hsu, K.-C.; Lin, C.-H.; Johnson, K.R.; Liu, C.-H.; Chang, T.-Y.; Huang, K.-L.; Fann, Y.-C.; Lee, T.-H. Autodetect extracranial and intracranial artery stenosis by machine learning using ultrasound. Comput. Biol. Med. 2020, 116, 103569. [Google Scholar] [CrossRef]
  47. Hang, S.; Miao, Y.; Chen, J.; Zhang, X.; Han, L.; Huang, Z.; Pei, N.; Liu, H.; An, C. CCS-Net: Cascade Detection Network with the Convolution Kernel Switch Block and Statistics Optimal Anchors Block in Hypopharyngeal Cancer MRI. IEEE J. Biomed. Health Inform. 2022, 27, 433–444. [Google Scholar]
  48. Zhang, S.; Miao, Y.; Chen, J.; Zhang, X.; Han, L.; Ran, D.; Huang, Z.; Pei, N.; Liu, H.; An, C. Twist-Net: A multi-modality transfer learning network with the hybrid bilateral encoder for hypopharyngeal cancer segmentation. Comput. Biol. Med. 2023, 154, 106555. [Google Scholar] [CrossRef] [PubMed]
Figure 1. A graphical representation of a novel TCD ultrasound waveform classification system.
Figure 1. A graphical representation of a novel TCD ultrasound waveform classification system.
Diagnostics 13 02000 g001
Figure 2. The outputs of the signal extraction process in the top–down view: (a) grayscale TCD spectrogram, (b) black-and-white spectrogram, and (c) envelope signal.
Figure 2. The outputs of the signal extraction process in the top–down view: (a) grayscale TCD spectrogram, (b) black-and-white spectrogram, and (c) envelope signal.
Diagnostics 13 02000 g002
Figure 3. Representative segments of clear and corrupted signals across the Healthy and ICU classes.
Figure 3. Representative segments of clear and corrupted signals across the Healthy and ICU classes.
Diagnostics 13 02000 g003
Figure 4. Details of data segmentation before and after data cleaning. From left to right, (a) a graphical representation of the number of segments across binary classes before data cleaning, (b) the graphical representation of clear and corrupted data numbers across the classes, and (c) a chart view of the number of the segments in Healthy and ICU classes used in this study.
Figure 4. Details of data segmentation before and after data cleaning. From left to right, (a) a graphical representation of the number of segments across binary classes before data cleaning, (b) the graphical representation of clear and corrupted data numbers across the classes, and (c) a chart view of the number of the segments in Healthy and ICU classes used in this study.
Diagnostics 13 02000 g004
Figure 5. The architecture of the proposed Self-ResNet18 model. The different colors of the boxes represent different layers and residual blocks in the proposed Self-ResNet18 model.
Figure 5. The architecture of the proposed Self-ResNet18 model. The different colors of the boxes represent different layers and residual blocks in the proposed Self-ResNet18 model.
Diagnostics 13 02000 g005
Figure 6. Model architecture of Self-ResAttentioNet18. The different colors of the boxes represent different layers and residual blocks in the proposed Self-ResAttentioNet18 model.
Figure 6. Model architecture of Self-ResAttentioNet18. The different colors of the boxes represent different layers and residual blocks in the proposed Self-ResAttentioNet18 model.
Diagnostics 13 02000 g006
Figure 7. Mean and standard deviation of fold-wise accuracy of different models for the six versions of models from Self-ResNet18 and Self-ResAttentioNet18 model architecture.
Figure 7. Mean and standard deviation of fold-wise accuracy of different models for the six versions of models from Self-ResNet18 and Self-ResAttentioNet18 model architecture.
Diagnostics 13 02000 g007
Figure 8. Learning curves of Self-ResAttentionResNet18_Q1. The left graphical representation is the (a) accuracy curve and the right graphical representation is the (b) loss curve.
Figure 8. Learning curves of Self-ResAttentionResNet18_Q1. The left graphical representation is the (a) accuracy curve and the right graphical representation is the (b) loss curve.
Diagnostics 13 02000 g008
Figure 9. (a) Confusion matrix where 0 indicates the Healthy class and 1 indicates the ICU class, and (b) ROC curve of Self-ResAttentioNet18_Q1 model.
Figure 9. (a) Confusion matrix where 0 indicates the Healthy class and 1 indicates the ICU class, and (b) ROC curve of Self-ResAttentioNet18_Q1 model.
Diagnostics 13 02000 g009
Table 1. Overview of the dataset used in this study.
Table 1. Overview of the dataset used in this study.
HealthyICU
Number of subjects612
DiagnosisHealthyIntraparenchymal hemorrhage, hydrocephalus, and traumatic brain injury (TBI)
Age group25–45 years23–74 years
Number of recordings1630
The total duration of all recordings2 h2 h 40 min
Targeted arteriesMCAMCA, ICA
Data of targeted artery used in this study MCAMCA
Ultrasound systemPhilips CX-50 with a 1.75 MHz transducer (S5-1)
Field of view90°
US probe placementTemporal region and the M1 segment of the MCA
Blow flow direction for MCATowards the probe
Blood flow velocity for MCAPositive
Table 2. Fold creation and stratification criteria.
Table 2. Fold creation and stratification criteria.
Fold NumSession ID
Train (Healthy)Test (Healthy)Train (ICU)Test (ICU)
1Rest“14–16”Rest“1, 2, 5”
2Rest“10–12”Rest“24–26”
3Rest“7–10”Rest“21–23”
4Rest“4–6”Rest“4, 5, 20”
5Rest“1–3”Rest“27–29”
Table 3. Training parameters used for binary classification.
Table 3. Training parameters used for binary classification.
Training ParametersValue
Number of folds5
Batch size4
Number of epochs100
Learning rate0.0001
Epoch patience7
Learning factor0.2
Loss typeSoftM_MSE
Optimization functionAdam
Table 4. Performance metrics for six variants of Self-ResNet18 and Self-ResAttentioNet18 different models.
Table 4. Performance metrics for six variants of Self-ResNet18 and Self-ResAttentioNet18 different models.
Model NameOverall AccuracyPrecisionRecallF1 ScoreSpecificity
Self-ResNet18_Q193.4693.4893.4693.4693.39
Self-ResNet18_Q394.9194.9294.9194.9194.88
Self-ResNet18_Q593.3193.3293.3193.393.26
Self-
ResAttentioNet18_Q1
96.0596.0696.0596.0696.09
Self-
ResAttentioNet18_Q3
95.3195.3495.3195.3195.37
Self-
ResAttentioNet18_Q5
92.9793.0192.9792.9793.03
Table 5. Evaluation of the proposed model against the background of previous research. The best results are highlighted in bold texts.
Table 5. Evaluation of the proposed model against the background of previous research. The best results are highlighted in bold texts.
Ref.DataClassesClassification ModelAccuracyRecallSpecificity
[45]TCD signal from MCAVasospasm & normalDecision tree89.1787.5%89.74%
[18]Peripheral pulse waveFive degrees of stenosis ANN88.7%--
[46]Cerebral ultrasoundStenosis & normalSVM80.8% to 81.9%70.9% to 73.1%90.7% to 90.8%
[17]TCD from basal arteriesStenosis & non-stenosisRNN71.1% to 75.89%74.6% to 75.53%71.15% to 74.89%
OursTCD signal from MCAHealthy & ICUSelf-
ResAttentioNet18_Q1
96.05%96.05%96.09%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nisha, N.N.; Podder, K.K.; Chowdhury, M.E.H.; Rabbani, M.; Wadud, M.S.I.; Al-Maadeed, S.; Mahmud, S.; Khandakar, A.; Zughaier, S.M. A Deep Learning Framework for the Detection of Abnormality in Cerebral Blood Flow Velocity Using Transcranial Doppler Ultrasound. Diagnostics 2023, 13, 2000. https://doi.org/10.3390/diagnostics13122000

AMA Style

Nisha NN, Podder KK, Chowdhury MEH, Rabbani M, Wadud MSI, Al-Maadeed S, Mahmud S, Khandakar A, Zughaier SM. A Deep Learning Framework for the Detection of Abnormality in Cerebral Blood Flow Velocity Using Transcranial Doppler Ultrasound. Diagnostics. 2023; 13(12):2000. https://doi.org/10.3390/diagnostics13122000

Chicago/Turabian Style

Nisha, Naima Nasrin, Kanchon Kanti Podder, Muhammad E. H. Chowdhury, Mamun Rabbani, Md. Sharjis Ibne Wadud, Somaya Al-Maadeed, Sakib Mahmud, Amith Khandakar, and Susu M. Zughaier. 2023. "A Deep Learning Framework for the Detection of Abnormality in Cerebral Blood Flow Velocity Using Transcranial Doppler Ultrasound" Diagnostics 13, no. 12: 2000. https://doi.org/10.3390/diagnostics13122000

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop