Next Article in Journal
Study of the Relationship between Pulmonary Artery Pressure and Heart Valve Vibration Sound Based on Mock Loop
Next Article in Special Issue
RCKD: Response-Based Cross-Task Knowledge Distillation for Pathological Image Analysis
Previous Article in Journal
Exploring New Therapeutic Avenues for Ophthalmic Disorders: Glaucoma-Related Molecular Docking Evaluation and Bibliometric Analysis for Improved Management of Ocular Diseases
Previous Article in Special Issue
Effect of Cilostazol on Delayed Cerebral Infarction in Aneurysmal Subarachnoid Hemorrhage Using Explainable Predictive Modeling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Machine Learning Framework for Multistage Parkinson’s Disease Classification Using Acoustic Features of Sustained Korean Vowels

1
Department of Electrical and Computer Engineering, Inha University, Incheon 22212, Republic of Korea
2
Department of Neurology, Inha University Hospital, Inha University College of Medicine, Incheon 22212, Republic of Korea
*
Authors to whom correspondence should be addressed.
Bioengineering 2023, 10(8), 984; https://doi.org/10.3390/bioengineering10080984
Submission received: 15 July 2023 / Revised: 17 August 2023 / Accepted: 18 August 2023 / Published: 20 August 2023
(This article belongs to the Special Issue Artificial Intelligence in Auto-Diagnosis and Clinical Applications)

Abstract

:
Recent research has achieved a great classification rate for separating healthy people from those with Parkinson’s disease (PD) using speech and the voice. However, these studies have primarily treated early and advanced stages of PD as equal entities, neglecting the distinctive speech impairments and other symptoms that vary across the different stages of the disease. To address this limitation, and improve diagnostic precision, this study assesses the selected acoustic features of dysphonia, as they relate to PD and the Hoehn and Yahr stages, by combining various preprocessing techniques and multiple classification algorithms, to create a comprehensive and robust solution for classification tasks. The dysphonia features extracted from the three sustained Korean vowels /아/(a), /이/(i), and /우/(u) exhibit diversity and strong correlations. To address this issue, the analysis of variance F-Value feature selection classifier from scikit-learn was employed, to identify the topmost relevant features. Additionally, to overcome the class imbalance problem, the synthetic minority over-sampling technique was utilized. To ensure fair comparisons, and mitigate the influence of individual classifiers, four commonly used machine learning classifiers, namely random forest (RF), support vector machine (SVM), k-nearest neighbor (kNN), and multi-layer perceptron (MLP), were employed. This approach enables a comprehensive evaluation of the feature extraction methods, and minimizes the variance in the final classification models. The proposed hybrid machine learning pipeline using the acoustic features of sustained vowels efficiently detects the early and mid-advanced stages of PD with a detection accuracy of 95.48%, and with a detection accuracy of 86.62% for the 4-stage, and a detection accuracy of 89.48% for the 3-stage classification of PD. This study successfully demonstrates the significance of utilizing the diverse acoustic features of dysphonia in the classification of PD and its stages.

1. Introduction

Parkinson’s disease (PD) is one of the most common neurodegenerative disorders, after Alzheimer’s disease [1]. At the initial stage of PD, symptoms typically include speech problems, tremors, and memory loss. As the disease progresses, patients may find it difficult to walk, run, communicate, and perform basic daily tasks [2]. The inability to reverse or cure the disease is its worst characteristic [3]; therefore, every effort has been made to discover it early, and take preventive steps to reduce its negative consequences. Recent studies suggest that there are currently more than 10 million people globally affected by PD [4], and it is expected to become a significant cause of mortality among the elderly population by 2040.
There is increasing evidence suggesting that individuals with PD may experience voice/speech problems during the prodromal phase [5,6,7,8,9], and such symptoms often persist into the early stages of the disease. As a result, the analysis of speech signals offers a greater possibility of detecting Parkinson’s in its early stages, and speech analysis can be used as a non-invasive and cost-effective tool in the early detection and monitoring of PD [10]. Recent studies on PD telediagnosis have focused on identifying vocal impairments through sustained vowel phonation, or running speech, in subjects [11,12,13,14,15,16,17,18,19,20,21,22,23]. These studies have employed various speech-signal-processing algorithms to extract clinically relevant data for the assessment of PD. The features derived from these data were then used to train learning algorithms, to build reliable decision support systems. Some studies have proposed the use of artificial neural networks to distinguish healthy individuals from those with PD, while others have suggested the use of simple speech-signal-processing algorithms.
For example, Sakar et al. [11] were the first to introduce the tunable Q-factor wavelet transform (TQWT) to extract features from the voice signals of PD patients. Using ensemble learning approaches that combine different machine learning classifiers, their results demonstrate that TQWT performs comparably better than state-of-the-art speech-signal-processing algorithms in PD classification. Avuçlu et al. [12] employed multiple classifiers to detect PD, using 22 vocal features from 195 sound samples in various training and testing instances. They utilized signal-processing techniques to extract important features from the acoustic signals of subjects with Parkinson’s, and control subjects. Bourouhou et al. [13] also evaluated a variety of classifiers to identify people who might have PD. Similarly, Zhang et al. [14] used naive Bayes, as well as other machine learning methods, to identify the presence of PD. Meghraoui et al. [15] suggested using Bernoulli and Multinomial Naive Bayes (BMNB) with harmonicity, pitch, and pulse features to identify PD. Braga et al. [16] proposed a methodology to detect early signs of PD through free speech in uncontrolled background conditions, using signal- and speech-processing techniques integrated with machine learning algorithms with a very high accuracy. Nilashi et al. [17] employed a new hybrid intelligent system for predicting PD progression, using noise removal, clustering, and prediction methods. They used the adaptive neuro-fuzzy inference system and support vector regression to predict PD progression. Dao et al. [18] proposed an approach that uses grey wolf optimization for feature selection, and the light gradient-boosting machine for classification. Kadiri et al. [19] presented a technique for identifying PD by utilizing SVM with single-frequency filtering cepstral coefficients and shifted delta cepstral features derived from the voice signals of both Parkinson’s patients, and control subjects. Jie et al. [20] employed a deep learning algorithm to learn from PD speech data. They used an embedded deep stack group sparse autoencoder for deep feature learning, and the resulting features were combined with the original speech features. Gunduz et al. [21] proposed two frameworks, based on convolutional neural networks, to classify PD using vocal features. The frameworks differ in how they combine feature sets. Deep features successfully distinguish PD patients from healthy individuals, and improve the classifier performance. While the results may appear impressive, their approach does not seem practical for several reasons. All these previous works focused on binary classification, treating the early and advanced stages of PD equally. However, the multistage classification of PD to assess new characteristics for neurocognitive assessment is important. It involves dividing patients into different stages, based on symptom severity, and developing separate classification models for each stage. Considering the different stages of PD and their associated symptoms, this approach could enable better treatment outcomes.
Only a few studies have investigated the use of a multistage classification in PD. For example, Hsu et al. [24] utilized photon emission computed tomography, and Ricciardi et al. [25] employed a three-dimensional gait analysis in their research into the multistage classification of PD. However, it is important to uncover whether voice/speech acoustic characteristics can be useful in a multistage classification. Suppa et al. [26] evaluated voice changes using machine learning algorithms in patients with PD at different stages of the disease, while receiving OFF and ON therapy. However, this research aimed to investigate the impact of the disease severity on the voice, to explore the effect of L-Dopa medication on groups of patients at the early stage (H&Y ≤ 2), and the mid–advanced stage (H&Y > 2). Similarly, Templeton et al. [27] conducted a study utilizing various neurocognitive functional tests, including speech, to classify individuals into the early stages (H&Y Stages 1 and 2) and advanced stages (H&Y Stages 3, 4, and 5) of PD. However, none of the approaches focused on a multistage classification. Instead, they were limited to a binary classification, specifically classifying between two groups of disease severity.
Therefore, the objective of this study is to use a hybrid machine learning pipeline (HMLP) to evaluate new characteristics obtained from neurocognitive assessments, also known as digital biomarkers, in relation to PD and its stages (H&Y Stages 1–5) [22]. The proposed HMLP was used in stage classification, to effectively classify the stages of PD as they are in.
The main contributions of this article are as follows:
(i)
A speech corpus of three sustained Korean vowels has been created (InhaPD) from Parkinson’s patients who were recruited at Inha University Hospital.
(ii)
A feature bank consisting of 43 acoustic features has been collaboratively created, using baseline features, vocal-fold features, and time–frequency features. The analysis of variance (ANOVA) F-Value feature selection classifier from scikit-learn was employed, to identify the top 20 most relevant features, resulting in a more effective and efficient feature subset.
(iii)
To address the dataset imbalance, we introduced a synthetic minority over-sampling technique (SMOTE) that generates new samples in the minority class and its neighboring samples, effectively balancing the distribution of classes, and reducing bias in the classification.
(iv)
A 10-fold cross-validation is performed for each model, using StratifiedKFold, demonstrating the robustness and effectiveness of our system compared to other recent approaches to stage classification.
(v)
The evaluation of the classification included the measuring accuracy, precision, recall, F1 score, and the area under the curve (AUC). This evaluation aimed to test the hypothesis proposing that a HMLP analysis of sustained vowels could effectively differentiate individuals with different stages of the disease.
The remainder of this paper is structured as follows. Section 2 explains our methodology, and the architecture of the proposed framework. Section 3 presents the experimental results. Section 4 discusses the findings and limitations of the research, and Section 5 concludes the paper.

2. Methodology

Currently, there is a lack of a standardized methodology that has undergone sufficient validation through clinical trials [23]. The primary focus of this research was not to develop a machine learning algorithm solely aimed at achieving the highest accuracy in automatically detecting PD. Instead, the study aimed to isolate multiple factors, and carry out experiments to determine whether these factors might lead to overly optimistic outcomes in the multistage classification of PD. The details of each part can be found in the subsequent sections.

2.1. Methods

This study examines speech signals, and extracts relevant features to correlate with responses, including several classes of PD. The disease detection model, depicted in Figure 1, involves multiple steps, from voice recording to the final classification. The primary steps include data collection, feature extraction and selection, model training for classification, and validation. The performance metrics of the classifier are evaluated, to determine the success of the detection system.

2.2. Dataset

In the areas of expanding datasets to include a larger corpus, and discovering novel objective biomarkers [28], a new corpus has been created, named InhaPD. The voice data for InhaPD were obtained from Parkinson’s patients who were recruited at Inha University Hospital, Incheon, South Korea. We enrolled a total of 101 individuals diagnosed with PD, who were between the ages of 43 and 81 years, with an average age of 68.11 ± 6.83 years. In all cases, the participants were native Korean speakers. None of the participants reported any respiratory disorders, hearing loss, or non-neurological disorders that may affect their vocal cords. All participants provided written informed consent, which was authorized by the institutional ethics committee. The clinical diagnosis of PD was based on the UK PD Society Brain Bank Diagnostic Criteria [29]. The symptoms and signs linked to Parkinson’s disease were evaluated and scored utilizing the H&Y scale [22]. Table 1 gives a comprehensive overview of the demographic details of the PD patients, including their gender, age, age onset, disease duration, and H&Y stage. The full data set can be found in the Supplementary Materials.

2.3. Voice Recordings

Consumer-focused devices, such as smartphones or tablets, offer an alternative that is widely available to patients, and has the potential to provide objective, frequent, and sensitive assessments. Additionally, the remote monitoring of PD using smart devices is gaining popularity, and numerous recent studies have explored the use of mobile devices for voice recording in PD research. For instance, Omberg et al. [30] used an iPhone for their voice recordings, while Farago et al. [31] employed Android smartphone devices (model not specified). On the other hand, Asci et al. [32] utilized various smartphones available on the market (Apple®, Samsung®, Huawei®, Xiaomi®, and Asus®). Lipsmeier et al. [33] used a Samsung Galaxy S3 mini; similarly, Vaiciukynas et al. [34] used a Samsung Galaxy Note 3, and Benba et al. [35] used a sound card in a desktop computer. These studies suggest that mobile devices can be a feasible and reliable alternative to traditional recording equipment for voice recording in Parkinson’s disease research.
In this study, the recordings were conducted in the inspection room of the hospital, using a Samsung Galaxy Tab S7 FE placed approximately 30 cm away from the mouth. The recording was not performed in an anechoic chamber, so a low level of ambient noise was present. The noise, however, was not significant enough to interfere with the analysis of the voice recordings. Similar to the approach followed by Omberg et al. [30] and Asci et al. [32], we also saved our voice samples in a compressed audio file format. The voice samples were recorded in mp3 format, with a sampling rate of 48 kHz, and a 32-bit sample size.
To obtain the voice recordings, participants were instructed to perform specific speech tasks using their normal voice intensity, pitch, and quality. The speech tasks included the sustained production of three Korean vowels, namely, /아/(a), /이/(i), and /우/(u). Each subject was recorded producing these vowels for a minimum of 10 s, resulting in a total of nine voice recordings for each subject (three samples for each vowel). The recorded Korean vowels were equivalent to the English /a/, /i/, and /u/, correspondingly.

2.4. Acoustic Measures

The research focusing on the impact of PD on the phonatory system mainly investigates the dysfunctions observed in the structures and muscles responsible for phonation. These include the diaphragm, muscles associated with the larynx, the vocal folds, and the resonant cavities above the glottis. Utilizing sustained vowels can produce simple acoustic patterns that have the potential to yield consistent and dependable assessments of voice quality [28]. These acoustic measures can be directly acquired using software and libraries that are readily accessible in the literature [30,36]. In this study, a set of 43 widely known acoustic measures were extracted for each vowel. All these features and their definitions are listed in Table 2.
Among these measures, all F0 features, all F1–4 features, all the jitter features, all the shimmer features, HNR, pF, F_Disp, avgF, Fitch_vtl, delta_F, and vtl_delta_F were obtained from Parselmouth 0.3.3 [37], which is a Python library for Praat software [36]. The mel-frequency cepstrum related features, for example, MFCC 1–4, and MFCC jitter 1–4, were obtained using MATLAB, as mentioned in [30].

2.5. Feature Selection

Most experiments, especially those involving larger corpora where the dimensionality of the feature vectors exceeded the number of recordings, utilized dimensionality-reduction techniques. For instance, Ozbolt et al. [23] used principal component analysis for dimensionality reduction in their study, utilizing the scikit-learn Python module. They also employed the ANOVA F-Value feature selection classifier from the scikit-learn SelectKBest module [38], to identify the most informative features for PD detection, selecting 10, 30, 50, or 70 features. In our case, k is set to 20, meaning that the top 20 features with the highest F-test scores were selected. Considering a total of 43 extracted acoustic features, along with the inclusion of gender information as an additional feature, the study utilized a set of 44 features for evaluation. Notably, no dimensionality-reduction techniques were applied in this regard. The ranking of the 20 selected features for the multistage (stage 1 to 3) classification is depicted in Figure 2. It should be noted that the feature ranking may vary for different classifications, as different stages of PD might have varying acoustic characteristics. However, it is important to emphasize that feature ranking is not the primary focus of this research; rather, it is utilized for analysis and classification purposes.

2.6. Data Balancing

Class imbalance is a common concern in clinical data, where there is a substantial difference in the number of patients and control group samples. This imbalance can create challenges in a machine learning classification, leading to a biased accuracy that favors the majority class. To address this issue, Chawla et al. [39] introduced the synthetic minority over-sampling technique (SMOTE), which generates new samples in the minority class and its neighboring samples, effectively balancing the class distribution, and mitigating classification bias. In this study, we also employed SMOTE to address our class-imbalance problem concerning the different stages of PD.

2.7. Machine Learning Algorithms and Hyperparameters

After selecting the top-ranked features, and addressing the issue of data imbalance, this study employed four commonly used machine learning classifiers from the scikit-learn Python module [38]: RF, SVM, kNN, and MLP. These groups of algorithms are common, due to their high model interpretability, effective minimization of misclassification, and strong diagnostic performance. While many studies do use DNN, simpler models are often preferred, due to the limited size of the available datasets [23].
Random forest (RF) is an ensemble learning algorithm that combines multiple decision trees to improve the performance and reduce overfitting. It works by randomly selecting a subset of features and samples to build each tree, and then aggregating the predictions of all the trees to make the final decision.
Support vector machine (SVM) is a supervised learning algorithm that can be used in both regression and classification tasks. It is particularly useful for datasets with complex features, and can handle both linear and nonlinear data. SVM works by finding the best separating hyperplane that maximizes the margin between the different classes.
K-nearest neighbor (kNN) is a simple and intuitive algorithm that can be used both in classification and in regression tasks. It works by finding the k-closest neighbors of a new sample, and then assigning it to the most common class among those neighbors. The value of k is usually chosen through cross-validation.
Multi-layer perceptron (MLP) is a type of artificial neural network that consists of multiple layers of interconnected nodes, also known as neurons. Each neuron applies a nonlinear activation function to the weighted sum of its inputs, allowing the network to learn complex relationships between the input features and the target variable.
A 10-fold cross-validation is performed for each model, using StratifiedKFold from the scikit-learn Python module. In 10-fold cross-validation, the data are divided into ten subsets, and each subset is used as a validation set exactly once, while the remaining nine subsets are used as the training set. This process is repeated ten times, with each of the ten subsets used exactly once, as the validation data. GridSearchCV is also used to find the best hyperparameters for each model, through searching the specified hyperparameter grid, and evaluating the model performance using cross-validation. The performance metrics are calculated for each fold, and the average performance metrics are reported for each model.

2.8. Performance Metrics

Choosing the appropriate metrics is crucial, because they affect the way in which performance is measured and compared. For instance, accuracy can be a useful metric when the classes are balanced but, when the classes are imbalanced, it can be misleading. In such cases, metrics such as precision, recall, F1 score, confusion matrix, and AUC can provide a better understanding of the model’s performance [40].
Accuracy measures the overall correctness of the model’s predictions. Precision (also known as the positive predictive value (PPV)) quantifies the proportion of correctly classified positive samples (PD stages) out of the total predicted positive samples, while recall (also known as sensitivity) measures the proportion of correctly classified positive samples out of the total actual positive samples. The F1 score provides a balanced measure, by taking into account both precision and recall.
To summarize the model’s performance, a confusion matrix is commonly utilized. It consists of four components: true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN). True positives represent cases where the model correctly predicts the positive class, or the correct stage of PD. True negatives refer to cases where the model correctly predicts the negative class, or correctly identifies a different PD stage. False positives occur when the model incorrectly predicts a positive class, or wrongly identifies a PD stage, while false negatives refer to cases where the model incorrectly predicts a negative class, or fails to identify a PD stage.
These performance metrics, including accuracy, precision, recall, and the F1 score, along with the confusion matrix, provide valuable insights into the algorithm’s effectiveness in classifying different stages of Parkinson’s disease. Additionally, AUC is a commonly used metric that quantifies the classifier’s discrimination ability in ranking and assigning higher probabilities to positive instances, compared to negative instances.

3. Results

This section presents the results obtained using the Parkinson’s disease (PD) datasets defined in Section 2.2. The machine learning classifiers used in this study included random forest (RF), support vector machine (SVM), k-nearest neighbor (kNN), and multi-layer perceptron (MLP). After preprocessing the dataset, which involved feature selection and data balancing techniques, we obtained the best hyperparameters for each model, and evaluated their performance. The best hyperparameters for each algorithm have also been listed in the tables. Five evaluation metrics were used: accuracy, precision, recall, F1 score, and AUC. We evaluated the results in two subsequent sections, namely Section 3.1 and Section 3.2, using the binary and multistage approaches, respectively.

3.1. Binary Classification

The classification of disease stages, specifically distinguishing between early-stage (H&Y Stages 1 and 2) and mid-stage (H&Y Stage 3) Parkinson’s disease, was conducted for all individuals with PD. Table 3 reports the measured metrics for the classification between individuals in the early stage and mid-stages of PD. SVM demonstrates the highest accuracy of 95.48%, and an F1 score of 0.9528, kNN produces the highest precision of 0.9956, and RF yields the highest recall of 0.9322, and an AUC score of 0.9873. However, MLP’s performance is relatively weaker, compared to the other models in the binary classification. Figure 3 resembles their corresponding confusion matrix.
Figure 3 represents the confusion matrix for the proposed HMLP models for HY ≤ 2 vs. HY > 2 classification. The confusion matrix also supports the analysis in Table 3.
Suppa et al. [26] and Templeton et al. [27] conducted an analysis comparing individuals with HY ≤ 2 (H&Y stages 1 and 2) to those with HY > 2 (H&Y stage 3 and/or more) in their respective research studies. Table 4 presents a comparative analysis of the results, highlighting the superior performance of our proposed HMLP approach utilizing the SVM algorithm. The results indicate that our approach achieved a higher accuracy, precision, and AUC, compared to the referenced studies. Notably, none of the studies were able to attain an accuracy level of 90% or higher.

3.2. Multistage Classification

The analysis mentioned above falls into the category of a binary classification. However, the primary objective of this study was to perform a multistage classification of PD. As a result, the classification of the disease stages (i.e., determining the stage of PD) was conducted for all individuals with Parkinson’s. The performance metrics in Table 5 correspond with the 4-stage classification, whereas Table 6 corresponds to the 3-stage classification. Figure 4 and Figure 5 represent their corresponding confusion matrices, respectively.
Figure 4 represents the confusion matrix for the proposed HMLP models for 4-stage (stage 1, 2, 2.5, and 3) classification. The confusion matrix also supports the analysis in Table 5.
Figure 5 represents the confusion matrix for the proposed HMLP models for a 3-stage (stage 1, 2, and 3) classification. The confusion matrix also supports the analysis in Table 6.
In all scenarios, the accuracy of the multistage classification was consistently lower compared to binary classification, with accuracy levels reaching approximately 90%. Nevertheless, it is important to highlight that MLP, despite exhibiting a lower performance in the binary classification, surpassed all the other algorithms in the multistage classification tasks.

4. Discussion

The primary objective of this research was not to develop a machine learning algorithm solely focused on achieving the highest accuracy in the multistage classification of Parkinson’s disease (PD). Instead, the study aimed to isolate multiple factors, and conduct experiments to determine whether these factors might lead to overly optimistic results in the multistage classification of PD. Using voice biomarkers in machine learning methods, we aim to explore a non-invasive and cost-effective diagnostic tool for better treatment outcomes, to improve the quality of the patient’s life.
Furthermore, machine-learning-based voice analysis may also hold potential for assessing the effectiveness of symptomatic treatments for PD. By comparing the pre- and post-treatment voice characteristics, machine learning algorithms can potentially determine whether a treatment has improved the patient’s motor function and overall quality of life. Recent advancements in speech analysis methodologies have shown promising results. We proposed a hybrid machine learning pipeline consisting of four commonly used ML classifiers, using the acoustic features of sustained Korean vowels, and found that the hybrid model outperforms the baseline models. We achieved an accuracy of 86.62%, 89.48%, and 95.48% for the 4-stage, 3-stage, and 2-stage classification of PD, respectively, which was higher than state-of-the-art methods. However, the multistage accuracy of our model may be limited by the size of the dataset used in the evaluation. Using a larger dataset could improve the model’s ability to generalize and classify the stages of PD more effectively. Previous studies have investigated the multistage classification of Parkinson’s disease using invasive methods, such as photon emission computed tomography [24] and three-dimensional gait analysis [25]. However, it is crucial to determine the potential of non-invasive methods, specifically voice/speech acoustic characteristics, in a multistage classification. Suppa et al. [26] examined voice changes in patients with Parkinson’s disease using machine learning algorithms, specifically investigating the impact of the disease severity on the voice and the effect of L-Dopa medication on patients at the early stage (H&Y ≤ 2) and the mid–advanced stage (H&Y > 2). Similarly, Templeton et al. [27] conducted a study utilizing neurocognitive functional tests, including speech, to classify individuals into the early stages (H&Y Stages 1 and 2) and the advanced stages (H&Y Stages 3, 4, and 5) of PD. However, both studies were focused on a binary classification, and did not address a multistage classification that distinguished between multiple levels of disease severity. In summary, our study presents a reliable model for detecting PD and its stages with higher accuracy, precision, recall, F1 score, and AUC rates.
Overall, the usefulness of machine-learning-based voice analysis in PD should be further explored and discussed. It should be noted that there are several limitations to the present study. Firstly, as we did not record vocal samples from each patient sequentially, it is possible that there may be daily fluctuations in the vocal characteristics in PD that we were unable to capture. Moreover, in terms of clinical–instrumental correlations, it is important to acknowledge that machine learning analysis requires a large number of data. Furthermore, the relationship between specific aspects of hypokinetic dysarthria in PD (such as hypophonia, mono-pitch, and mono-loudness speech), and the specific voice features selected by the machine learning algorithm, remains uncertain, and requires further investigation. While k-fold cross-validation offers advantages, such as optimizing the use of limited data, and ensuring model robustness across multiple subsets, there is the inherent risk of overfitting when the entire dataset is used. It is vital to understand that, while cross-validation might produce optimistic results, these may not always mirror the real-world performance in a new, independent cohort. Relying exclusively on cross-validation results could lead to inflated expectations. An alternative, more conservative method involves partitioning the dataset into distinct training and testing sets. This holdout method minimizes the overfitting impact by evaluating the model’s performance on an untouched test set. In contrast, using a holdout dataset offers a more genuine glimpse into how the model is likely to fare when presented with new, real-world data, despite our data limitations necessitating the use of cross-validation. To develop a dependable dataset for detecting PD using acoustic features, it is essential to consider some other factors, including the gender and age balance, the quality of the microphone, noise, robustness, the number of subjects, the disease severity, and the impact of PD medication.
The current understanding of speech disorders in PD is primarily based on perceptual speech assessment. The computerized analysis of speech or the voice has been suggested for the diagnosis and monitoring of PD. Despite the potential advantages of voice analysis in PD, it has not been widely adopted in standard clinical practice yet [10]. Besides, machine learning models for detecting Parkinson’s disease are usually evaluated based on standard machine learning parameters, such as accuracy, precision, recall, F1 score, and AUC rates. However, these parameters may not accurately reflect the clinical effectiveness and anticipated positive adjustments to patient treatment. Additionally, doctors require training on how to utilize AI-powered diagnostic tools, as some of these parameters can be complex to interpret.

5. Conclusions

We anticipate that machine-learning-based voice analysis could serve as a novel disease biomarker in the near future, facilitating the clinical management of PD. Furthermore, we believe that our study could encourage the future use of machine learning voice analysis in PD telemedicine approaches, including home-based applications. We also speculate that future studies with a more sensible dataset may report a higher multistage accuracy than that reported in this study.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/bioengineering10080984/s1, A “Dataset Description.txt” file has also been included mentioning all the details about the included files.

Author Contributions

Conceptualization, S.I.M.M.R.M. and S.L.; methodology, S.I.M.M.R.M.; software, S.I.M.M.R.M.; validation, S.I.M.M.R.M., R.K., and S.L.; formal analysis, S.I.M.M.R.M.; investigation, R.K.; resources, S.L.; data curation, R.K.; writing—original draft preparation, S.I.M.M.R.M.; writing—review and editing, S.I.M.M.R.M.; visualization, S.I.M.M.R.M.; supervision, R.K. and S.L.; project administration, S.L.; funding acquisition, R.K. and S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT (NRF-2020R1A2C2004624 and NRF-2021R1C1C1011822).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of Inha University Hospital (protocol code: 2021-03-008-006; date of approval: 4 October 2022).

Informed Consent Statement

Not applicable.

Data Availability Statement

The simulation experiment data used to support the findings of this study are available as supplementary material, attached to the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pringsheim, T.; Jette, N.; Frolkis, A.; Steeves, T.D.L. The Prevalence of Parkinson’s Disease: A Systematic Review and Meta-analysis. Mov. Disord. 2014, 29, 1583–1590. [Google Scholar] [CrossRef] [PubMed]
  2. Pramanik, M.; Pradhan, R.; Nandy, P.; Qaisar, S.M.; Bhoi, A.K. Assessment of Acoustic Features and Machine Learning for Parkinson’s Detection. J. Healthc. Eng. 2021, 9957132. [Google Scholar] [CrossRef] [PubMed]
  3. Agnihotri, A.; Aruoma, O.I. Alzheimer’s Disease and Parkinson’s Disease: A Nutritional Toxicology Perspective of the Impact of Oxidative Stress, Mitochondrial Dysfunction, Nutrigenomics and Environmental Chemicals. Taylor. Fr. 2019, 39, 16–27. [Google Scholar] [CrossRef]
  4. Dorsey, E.R.; Sherer, T.; Okun, M.S.; Bloemd, B.R. The Emerging Evidence of the Parkinson Pandemic. J. Park. Dis. 2018, 8, S3–S8. [Google Scholar] [CrossRef]
  5. Rusz, J.; Hlavnička, J.; Novotný, M.; Tykalov, T.; Pelletier, A.; Montplaisir, J.; Gagnon, J.-F.; Dušek, P.; Galbiati, A.; Marelli, S.; et al. Speech Biomarkers in Rapid Eye Movement Sleep Behavior Disorder and Parkinson Disease. Ann. Neurol. 2021, 90, 62–75. [Google Scholar] [CrossRef] [PubMed]
  6. Ma, A.; Lau, K.K.; Thyagarajan, D. Voice Changes in Parkinson’s Disease: What Are They Telling Us? J. Clin. Neurosci. 2020, 72, 1–7. [Google Scholar] [CrossRef]
  7. Fereshtehnejad, S.M.; Yao, C.; Pelletier, A.; Montplaisir, J.Y.; Gagnon, J.F.; Postuma, R.B. Evolution of Prodromal Parkinson’s Disease and Dementia with Lewy Bodies: A Prospective Study. Brain 2019, 142, 2051–2067. [Google Scholar] [CrossRef]
  8. Rusz, J.; Tykalová, T.; Novotný, M.; Zogala, D.; Růžička, E.; Dušek, P. Automated Speech Analysis in Early Untreated Parkinson’s Disease: Relation to Gender and Dopaminergic Transporter Imaging. Eur. J. Neurol. 2022, 29, 81–90. [Google Scholar] [CrossRef]
  9. Hlavnika, J.; Cmejla, R.; Tykalová, T.; Šonka, K.; Ruzicka, E.; Rusz, J. Automated Analysis of Connected Speech Reveals Early Biomarkers of Parkinson’s Disease in Patients with Rapid Eye Movement Sleep Behaviour Disorder. Sci. Rep. 2017, 7, 12. [Google Scholar] [CrossRef]
  10. Ngo, Q.C.; Motin, M.A.; Pah, N.D.; Drotár, P.; Kempster, P.; Kumar, D. Computerized Analysis of Speech and Voice for Parkinson’s Disease: A Systematic Review. Comput. Methods Programs Biomed. 2022, 226, 107133. [Google Scholar] [CrossRef]
  11. Sakar, C.O.; Serbes, G.; Gunduz, A.; Tunc, H.C.; Nizam, H.; Sakar, B.E.; Tutuncu, M.; Aydin, T.; Isenkul, M.E.; Apaydin, H. A Comparative Analysis of Speech Signal Processing Algorithms for Parkinson’s Disease Classification and the Use of the Tunable Q-Factor Wavelet Transform. Appl. Soft. Comput. 2019, 74, 255–263. [Google Scholar] [CrossRef]
  12. Avuçlu, E.; Elen, A. Evaluation of Train and Test Performance of Machine Learning Algorithms and Parkinson Diagnosis with Statistical Measurements. Med. Biol. Eng. Comput. 2020, 58, 2775–2788. [Google Scholar] [CrossRef]
  13. Bourouhou, A.; Jilbab, A.; Nacir, C.; Hammouch, A. Comparison of Classification Methods to Detect the Parkinson Disease. In Proceedings of the 2016 International Conference on Electrical and Information Technologies, ICEIT, Tangiers, Morocco, 4–7 May 2016; pp. 421–424. [Google Scholar] [CrossRef]
  14. Zhang, L.; Qu, Y.; Jin, B.; Jing, L.; Gao, Z.; Liang, Z. An Intelligent Mobile-Enabled System for Diagnosing Parkinson Disease: Development and Validation of a Speech Impairment Detection System. JMIR Med. Inform. 2020, 8, e18689. [Google Scholar] [CrossRef]
  15. Meghraoui, D.; Boudraa, B.; Merazi-Meksen, T.; Boudraa, M. Parkinson’s Disease Recognition by Speech Acoustic Parameters Classification. Lect. Notes Netw. Syst. 2016, 1, 165–173. [Google Scholar]
  16. Braga, D.; Madureira, A.M.; Coelho, L.; Ajith, R. Automatic Detection of Parkinson’s Disease Based on Acoustic Analysis of Speech. Eng. Appl. Artif. Intell. 2019, 77, 148–158. [Google Scholar] [CrossRef]
  17. Nilashi, M.; Ibrahim, O.; Ahani, A. Accuracy Improvement for Predicting Parkinson’s Disease Progression. Sci. Rep. 2016, 6, 34181. [Google Scholar] [CrossRef]
  18. Dao, S.V.T.; Yu, Z.; Tran, L.V.; Phan, P.N.K.; Huynh, T.T.M.; Le, T.M. An Analysis of Vocal Features for Parkinson’s Disease Classification Using Evolutionary Algorithms. Diagnostics 2022, 12, 1980. [Google Scholar] [CrossRef]
  19. Kadiri, S.R.; Kethireddy, R.; Alku, P. Parkinson’s Disease Detection from Speech Using Single Frequency Filtering Cepstral Coefficients. In Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, Shanghai, China, 25–29 October 2020; pp. 4971–4975. [Google Scholar] [CrossRef]
  20. Ma, J.; Zhang, Y.; Li, Y.; Zhou, L.; Qin, L.; Zeng, Y.; Wang, P.; Lei, Y. Deep Dual-Side Learning Ensemble Model for Parkinson Speech Recognition. Biomed Signal Process Control 2021, 69, 102849. [Google Scholar] [CrossRef]
  21. Gunduz, H. Deep Learning-Based Parkinson’s Disease Classification Using Vocal Feature Sets. IEEE Access 2019, 7, 115540–115551. [Google Scholar] [CrossRef]
  22. Hoehn, M.M.; Yahr, M.D. Parkinsonism. Neurology 1967, 17, 427. [Google Scholar] [CrossRef]
  23. Ozbolt, A.S.; Moro-Velazquez, L.; Lina, I.; Butala, A.A.; Dehak, N. Things to Consider When Automatically Detecting Parkinson’s Disease Using the Phonation of Sustained Vowels: Analysis of Methodological Issues. Appl. Sci. 2022, 12, 991. [Google Scholar] [CrossRef]
  24. Hsu, S.Y.; Yeh, L.R.; Chen, T.B.; Du, W.C.; Huang, Y.H.; Twan, W.H.; Lin, M.C.; Hsu, Y.H.; Wu, Y.C.; Chen, H.Y. Classification of the Multiple Stages of Parkinson’s Disease by a Deep Convolution Neural Network Based on 99mTc-TRODAT-1 SPECT Images. Molecules 2020, 25, 4792. [Google Scholar] [CrossRef] [PubMed]
  25. Ricciardi, C.; Amboni, M.; De Santis, C.; Ricciardelli, G.; Improta, G.; Iuppariello, L.; D’Addio, G.; Barone, P.; Cesarelli, M. Classifying Different Stages of Parkinson’s Disease Through Random Forests. IFMBE Proc. 2020, 76, 1155–1162. [Google Scholar]
  26. Suppa, A.; Costantini, G.; Asci, F.; Di Leo, P.; Al-Wardat, M.S.; Di Lazzaro, G.; Scalise, S.; Pisani, A.; Saggio, G. Voice in Parkinson’s Disease: A Machine Learning Study. Front Neurol. 2022, 13, 831428. [Google Scholar] [CrossRef]
  27. Templeton, J.M.; Poellabauer, C.; Schneider, S. Classification of Parkinson’s Disease and Its Stages Using Machine Learning. Sci. Rep. 2022, 12, 14036. [Google Scholar] [CrossRef]
  28. Moro-Velazquez, L.; Gomez-Garcia, J.A.; Arias-Londoño, J.D.; Dehak, N.; Godino-Llorente, J.I. Advances in Parkinson’s Disease Detection and Assessment Using Voice and Speech: A Review of the Articulatory and Phonatory Aspects. Biomed. Signal. Process. Control. 2021, 66, 102418. [Google Scholar] [CrossRef]
  29. Hughes, A.J.; Daniel, S.E.; Kilford, L.; Lees, A.J. Accuracy of Clinical Diagnosis of Idiopathic Parkinson’s Disease: A Clinico-Pathological Study of 100 Cases. J. Neurol. Neurosurg. Psychiatry 1992, 55, 181–184. [Google Scholar] [CrossRef] [PubMed]
  30. Omberg, L.; Chaibub Neto, E.; Perumal, T.M.; Pratap, A.; Tediarjo, A.; Adams, J.; Bloem, B.R.; Bot, B.M.; Elson, M.; Goldman, S.M.; et al. Remote Smartphone Monitoring of Parkinson’s Disease and Individual Response to Therapy. Nat. Biotechnol. 2022, 40, 480–487. [Google Scholar] [CrossRef]
  31. Faragó, P.; Ștefănigă, S.-A.; Cordoș, C.-G.; Mihăilă, L.-I.; Hintea, S.; Peștean, A.-S.; Beyer, M.; Perju-Dumbravă, L.; Ileșan, R.R. CNN-Based Identification of Parkinson’s Disease from Continuous Speech in Noisy Environments. Bioengineering 2023, 10, 531. [Google Scholar] [CrossRef]
  32. Asci, F.; Costantini, G.; Di Leo, P.; Zampogna, A.; Ruoppolo, G.; Berardelli, A.; Saggio, G.; Suppa, A. Machine-Learning Analysis of Voice Samples Recorded through Smartphones: The Combined Effect of Ageing and Gender. Sensors 2020, 20, 5022. [Google Scholar] [CrossRef]
  33. Lipsmeier, F.; Taylor, K.I.; Kilchenmann, T.; Wolf, D.; Scotland, A.; Schjodt-Eriksen, J.; Cheng, W.Y.; Fernandez-Garcia, I.; Siebourg-Polster, J.; Jin, L.; et al. Evaluation of Smartphone-Based Testing to Generate Exploratory Outcome Measures in a Phase 1 Parkinson’s Disease Clinical Trial. Mov. Disord. 2018, 33, 1287–1297. [Google Scholar] [CrossRef]
  34. Vaiciukynas, E.; Verikas, A.; Gelzinis, A.; Bacauskiene, M. Detecting Parkinson’s Disease from Sustained Phonation and Speech Signals. PLoS ONE 2017, 12, e0185613. [Google Scholar] [CrossRef]
  35. Benba, A.; Jilbab, A.; Hammouch, A. Detecting Patients with Parkinson ’ s Disease Using Mel Frequency Cepstral Coefficients and Support Vector Machines. Int. J. Electr. Eng. Inform. 2015, 7, 297–307. [Google Scholar]
  36. Boersma, P. Praat: Doing Phonetics by Computer. Ear Hear. 2011, 32, 266. [Google Scholar] [CrossRef]
  37. Jadoul, Y.; Thompson, B.; de Boer, B. Introducing Parselmouth: A Python Interface to Praat. J. Phon. 2018, 71, 1–15. [Google Scholar] [CrossRef]
  38. Pedregosa, F.; Michel, V.; Grisel Oliviergrisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Vanderplas, J.; Cournapeau, D.; Pedregosa, F.; Varoquaux, G.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  39. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-Sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  40. Japkowicz, N.; Stephen, S. The Class Imbalance Problem: A Systematic Study. Intell. Data Anal. 2002, 6, 429–449. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed PD classification system.
Figure 1. Overview of the proposed PD classification system.
Bioengineering 10 00984 g001
Figure 2. Feature ranking based on F-test scores.
Figure 2. Feature ranking based on F-test scores.
Bioengineering 10 00984 g002
Figure 3. Confusion matrix for the proposed HMLP models for HY ≤ 2 vs. HY > 2 classification.
Figure 3. Confusion matrix for the proposed HMLP models for HY ≤ 2 vs. HY > 2 classification.
Bioengineering 10 00984 g003aBioengineering 10 00984 g003b
Figure 4. Confusion matrix for 4-stage (stage 1, 2, 2.5, and 3) classification.
Figure 4. Confusion matrix for 4-stage (stage 1, 2, 2.5, and 3) classification.
Bioengineering 10 00984 g004
Figure 5. Confusion matrix for 3-stage (stage 1, 2, and 3) classification.
Figure 5. Confusion matrix for 3-stage (stage 1, 2, and 3) classification.
Bioengineering 10 00984 g005
Table 1. Demographic information of InhaPD.
Table 1. Demographic information of InhaPD.
PD Stages
Whole GroupH&Y 1H&Y 2H&Y 2.5H&Y 3
Number10123491217
Gender
Male601328118
Female41102119
Age
avg68.1167.0867.5771.0869
std6.8377.153.897.1
min4349436554
max8181808079
Age onset
avg62.8665.4762.5965.4158.29
std7.797.097.755.238.61
min4147415943
max7878767571
Disease duration
avg5.331.8655.7510.7
std4.141.013.292.984.35
min11111
max184151118
NB. avg (average), std (standard deviation), min (minimum), max (maximum).
Table 2. A summary of the feature sets used in this study.
Table 2. A summary of the feature sets used in this study.
Feature NameDefinitionsSymbol
F0 (mean)Mean of the fundamental frequencymeanF0
F0 (median)Median of the fundamental frequencymedianF0
F0 (stdv.)Standard deviation of the fundamental frequencystdevF0
F 1–4 (mean)Mean of the formant frequencies F1 to F4f1-4_mean
F 1–4 (median)Median of the formant frequencies F1 to F4f1-4_median
Jitter (%)Measure of the average absolute difference between consecutive periods, divided by the average periodlocalJitter
Jitter (abs)Measure of the cycle-to-cycle variation in the fundamental frequency, typically expressed in secondsabsJitter
Jitter: RAPMeasure of the average absolute difference between a period and the average of that period and its two neighboring periods, divided by the average period lengthrapJitter
Jitter: PPQ5Measure of the average absolute difference between a period and the average of that period and its four closest neighboring periods, divided by the average period lengthppq5Jitter
Jitter: DDPMeasure of the average absolute difference between consecutive differences of consecutive periods, divided by the average period lengthddpJitter
Jitter: meanMean of the jittermeanJt
Jitter: medianMedian of the jittermedianJt
Jitter: PCATwo-factor principal components analysis (PCA) on jitterjitterPCA
Shimmer (%)Measure of the average absolute difference between the amplitudes of consecutive periods, divided by the average amplitude, expressed as a percentagelocalShimmer
Shimmer (dB)Measure of the average absolute base-10 logarithm of the difference between the amplitudes of consecutive periods, multiplied by 20dbShimmer
Shimmer: APQ3Measure of the variability in the amplitude of a speech signal, as measured from the amplitude of a single period to the average amplitude of its three closest neighborsapq3Sh
Shimmer: APQ5Measure of the variability in the amplitude of a speech signal, as measured from the amplitude of a single period to the average amplitude of its five closest neighborsapq5Sh
Shimmer: APQ11Measure of the variability in the amplitude of a speech signal, as measured from the amplitude of a single period to the average amplitude of its eleven closest neighborsapq11Sh
Shimmer: DDAMeasure of the variability in the amplitude of a speech signal, as measured from the consecutive differences between the amplitudes of consecutive periodsddaSh
Shimmer: meanMean of the shimmermeanSh
Shimmer: medianMedian of the shimmermedianSh
Shimmer: PCATwo-factor PCA on shimmershimmerPCA
HNRAmplitude of the tonal relative to the noise components.hnr
MFCC 1–4Mel-frequency cepstrum band 1 to 4mfcc1-4
MFCC jitter 1–4Positive change in mel-frequency cepstrum band 1 to 4 over timemfccJt1-4
pFFormant positionpF
F_DispDispersion of formant frequencyfdisp
avgFAverage formantavgFormant
Fitch_vltFitch vocal tract lengthfitch_vtl
delta_FCumulated variation of formantsdelta_f
vtl_delta_FCumulated variation of vocal tract lengthvtl_delta_f
GenderGender information of each subject (male, female)gender
Table 3. Performance on the proposed HMLP models for HY ≤ 2 vs. HY > 2 classification.
Table 3. Performance on the proposed HMLP models for HY ≤ 2 vs. HY > 2 classification.
ModelBest HyperparametersAverage AccuracyAverage PrecisionAverage RecallAverage F1 ScoreAverage AUC
RF{‘max_depth’: None, ‘min_samples_split’: 2, ‘n_estimators’: 200}0.94480.95750.93220.94380.9873
SVM{‘C’: 50, ‘gamma’: ‘scale’}0.95480.99140.91760.95280.9846
kNN{‘n_neighbors’: 3,
‘weights’: ‘distance’}
0.91620.99560.83640.9080.9694
MLP{‘activation’: ‘relu’, ‘hidden_layer_sizes’: (1000, 500, 200), ‘learning_rate’: ‘constant’, ‘max_iter’: 3000, ‘solver’: ‘adam’}0.93690.9630.90970.93480.9785
Table 4. Comparison of the results for HY ≤ 2 vs. HY > 2 classification.
Table 4. Comparison of the results for HY ≤ 2 vs. HY > 2 classification.
ComparisonsAccuracyPrecision
(PPV)
Recall
(Sensitivity)
AUCAlgorithmAuthors
HY ≤ 2
vs.
HY > 2
0.95480.99140.91760.9848SVMProposed HMLP
0.8800.8890.8720.934SVM and ANNSuppa et al. [26]
0.89470.92860.9286-CARTTempleton et al. [27]
N.B.: HY ≤ 2 (early-stage patients with PD, H&Y Stage ≤ 2), HY > 2 (mid–advanced-stage patients with PD, H&Y Stage > 2), SVM (support vector machine), ANN (artificial neural network), CART (classification and regression tree), (-) not available.
Table 5. Performance on the proposed HMLP models for 4-stage (stage 1, 2, 2.5, and 3) classification.
Table 5. Performance on the proposed HMLP models for 4-stage (stage 1, 2, 2.5, and 3) classification.
ModelBest HyperparametersAverage AccuracyAverage PrecisionAverage RecallAverage F1 ScoreAverage AUC
RF{‘max_depth’: None, ‘min_samples_split’: 2, ‘n_estimators’: 200}0.85030.85140.85030.84740.9694
SVM{‘C’: 50, ‘gamma’: ‘scale’}0.86510.86710.86510.86040.9625
kNN{‘n_neighbors’: 3,
‘weights’: ‘distance’}
0.86110.86960.86110.85130.9469
MLP{‘activation’: ‘relu’, ‘hidden_layer_sizes’: (1000, 500, 200), ‘learning_rate’: ‘constant’, ‘max_iter’: 3000, ‘solver’: ‘adam’}0.86620.86620.86620.86350.9584
Table 6. Performance on the proposed HMLP models for 3-stage (stage 1, 2, and 3) classification.
Table 6. Performance on the proposed HMLP models for 3-stage (stage 1, 2, and 3) classification.
ModelBest HyperparametersAverage AccuracyAverage PrecisionAverage RecallAverage F1 ScoreAverage AUC
RF{‘max_depth’: None, ‘min_samples_split’: 2, ‘n_estimators’: 200}0.88680.88870.88680.8860.9764
SVM{‘C’: 50, ‘gamma’: ‘scale’}0.87640.88190.87640.87390.9603
kNN{‘n_neighbors’: 3,
‘weights’: ‘distance’}
0.87520.88980.87520.86950.9533
MLP{‘activation’: ‘relu’, ‘hidden_layer_sizes’: (1000, 500, 200), ‘learning_rate’: ‘constant’, ‘max_iter’: 3000, ‘solver’: ‘adam’}0.89480.89640.89480.89330.9602
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mondol, S.I.M.M.R.; Kim, R.; Lee, S. Hybrid Machine Learning Framework for Multistage Parkinson’s Disease Classification Using Acoustic Features of Sustained Korean Vowels. Bioengineering 2023, 10, 984. https://doi.org/10.3390/bioengineering10080984

AMA Style

Mondol SIMMR, Kim R, Lee S. Hybrid Machine Learning Framework for Multistage Parkinson’s Disease Classification Using Acoustic Features of Sustained Korean Vowels. Bioengineering. 2023; 10(8):984. https://doi.org/10.3390/bioengineering10080984

Chicago/Turabian Style

Mondol, S. I. M. M. Raton, Ryul Kim, and Sangmin Lee. 2023. "Hybrid Machine Learning Framework for Multistage Parkinson’s Disease Classification Using Acoustic Features of Sustained Korean Vowels" Bioengineering 10, no. 8: 984. https://doi.org/10.3390/bioengineering10080984

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop