Next Article in Journal
High Expressed Emotion and Warmth among Families of Patients with Schizophrenia in Greece
Next Article in Special Issue
Using Deep Neural Network Approach for Multiple-Class Assessment of Digital Mammography
Previous Article in Journal
The Relationships between Physical Activity, Screen Time and Sleep Time According to the Adolescents’ Sex and the Day of the Week
Previous Article in Special Issue
Customized Deep Learning Classifier for Detection of Acute Lymphoblastic Leukemia Using Blood Smear Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dysarthria Speech Detection Using Convolutional Neural Networks with Gated Recurrent Unit

1
Department of Information Management, National Yunlin University of Science and Technology, Douliu 64002, Taiwan
2
Department of Electrical and Computer Engineering, Iowa State University, 2520 Osborn Drive, Ames, IA 50011, USA
*
Author to whom correspondence should be addressed.
Healthcare 2022, 10(10), 1956; https://doi.org/10.3390/healthcare10101956
Submission received: 12 September 2022 / Revised: 2 October 2022 / Accepted: 5 October 2022 / Published: 7 October 2022

Abstract

:
In recent years, due to the rise in the population and aging, the prevalence of neurological diseases is also increasing year by year. Among these patients with Parkinson’s disease, stroke, cerebral palsy, and other neurological symptoms, dysarthria often appears. If these dysarthria patients are not quickly detected and treated, it is easy to cause difficulties in disease course management. When the symptoms worsen, they can also affect the patient’s psychology and physiology. Most of the past studies on dysarthria detection used machine learning or deep learning models as classification models. This study proposes an integrated CNN-GRU model with convolutional neural networks and gated recurrent units to detect dysarthria. The experimental results show that the CNN-GRU model proposed in this study has the highest accuracy of 98.38%, which is superior to other research models.

1. Introduction

Speech is an essential medium of communication between people. Once the medium of communication is abnormal, it increases the difficulty of communication. Furthermore, many people with neurological diseases often have this condition, which is called dysarthria. Dysarthria is mainly a symptom caused by neuromuscular control disorders that affect breathing, vocalization, resonance, articulation, and prosody [1]. Sounds can be too loud or too low due to damage to the central and peripheral nervous system, such as stroke, Parkinson’s disease, brain trauma, brain tumors, cerebral palsy, amyotrophic lateral sclerosis, multiple sclerosis, muscular dystrophy, and other neurological diseases. The voice will also appear hoarse and lack tone changes. Hence, dysarthria patients are more likely to have abnormal speech characteristics [2].
Dysarthria may lead to social difficulties, a sense of isolation, and even world-weariness, depression, and other psychological problems [3]. Therefore, if there is no timely intervention and early rehabilitation training, it is easy to cause difficulties in disease course management, and the disease will continue to worsen. Doctors can subjectively diagnose dysarthria, but it is generally considered an expensive, laborious, and time-consuming test [4]. Therefore, having an objective and immediate automatic backing test is extremely important.
Deep learning has recently been popular and widely used in medical treatment. In order to objectively and accurately diagnose patients with dysarthria, more and more researchers are using deep learning to develop automatic detection of dysarthria. Many researchers use words for speech detection and different feature extraction methods to extract features from speech signals. For example, Vashkevich et al. [5] use pitch period entropy (PPE) based on acoustic features. Muhammad et al. [6] use glottal to noise excitation (GNE) and formant frequency or use spectrum and cepstrum for feature extraction. Other examples are mel-frequency cepstral coefficients (MFCC) [7], perception linear predictive coefficients (PLP), etc. [8]. After that, deep learning methods are used to detect dysarthria, such as convolutional neural network (CNN), CNN-LSTM (long short-term memory), and other models [9,10].
In the previous studies on the detection of dysarthria using the UA-Speech database, Narendra [10] selected the CNN-LSTM hybrid model as the classification model, but the accuracy of this model was only 77.57%. In order to improve the accuracy of the dysarthria detection model, this study used speech signals recorded by dysarthria patients and healthy people to undergo a short-time Fourier transform (STFT) and then convert the signals into spectrograms. After that, the signals were transformed into a spectral map, and mel-frequency cepstral coefficients (MFCC) were used to select the features. Finally, the accuracy in detecting dysarthria of the proposed CNN-GRU (gated recurrent unit) deep learning model was compared with three other models (CNN, LSTM, and CNN-LSTM).

2. Materials and Methods

2.1. Data Collection

Schlauch et al. [11] pointed out in their study that patients with dysarthria use words to make judgments with high recognition and low error rates. Therefore, this study chose words as the input audio samples for the subsequent studies. Our dataset was collected from the UA research database [12] (http://www.isle.illinois.edu/sst/data/UASpeech/, accessed on 18 February 2022). This database mainly contains the voice recordings of 15 dysarthria patients (4 women and 11 men) and 13 healthy subjects (4 women and 9 men), all of which were recorded by microphone and processed by noise removal. The subjects ranged in age from 18 to 58. A total of 455 words were recorded for each subject in the database, consisting of the numbers 1 to 10, the 26 letters, 19 computer command words, the 100 most common words from the Brown Corpus, and 300 words selected from the Project Gutenberg novel.

2.2. Method

The method proposed in this study consists of three stages, as shown in Figure 1. In the first stage, the original speech signal is transformed from the time domain to the frequency domain by a short-time Fourier transform. Second, the frequency domain data are extracted by mel-frequency cepstral coefficients. In the third stage, the features extracted from the mel spectrogram are used to detect and classify dysarthria patients and healthy people using the CNN-GRU model used in this study. In order to verify the excellence of the CNN-GRU deep learning model, this study also used the CNN model, LSTM model, and CNN-LSTM model to detect dysarthria and compare their results.

2.3. Data Preprocessing

The audio could identify the amplitude waveform differences from the audio images of patients with dysarthria and healthy people through waveform images because people with dysarthria pronounce words more slowly and with a less steady pitch than healthy people. In general, the waveforms of the dysarthria patient (ID: dysarthria01) in Figure 2a are more irregular than the healthy subject (ID: healthy01) in Figure 2b. Audio waveforms can only show the relationship between amplitude and time. This study used Python Librosa to perform a short-time Fourier transform (STFT) of the audio. The short-time Fourier spectrograms of a dysarthria patient and a healthy subject are shown in Figure 3. From the spectrograms in Figure 3, it can be observed that the spectrum of subject dysarthria01 (Figure 3a) has more irregular frequencies and sudden higher decibels than the spectrum of subject healthy01 (Figure 3b).
Short-time Fourier transform (STFT) was used to transform speech signals from the time domain to the frequency domain. The frame length of the speech in this study was between 10 and 30 ms, the sampling frequency was set to 8 KHz, and the window length was set to 128 to improve the resolution. The transformation of the STFT voice signal x(T) to the frequency domain is shown in Equation (1).
X t , f = ω t T x T e j 2 π f T   dT  

2.4. Feature Selection

Mel-frequency cepstral coefficients (MFCC) are widely used in speech recognition. Mel is the scale of tone frequencies picked up by the human ear. The relationship between the mel spectrum (M) and frequency (Hz) is shown in Equations (2) and (3).
M f = 2595 × log 10 1 + f / 700
f = 700 10 m 2595 1
The power of spectrum, P(k), can be obtained from Equation (4).
P k = 1 N     X k 2
The power of the spectrum, P(k), is passed through a series of mel-scale triangular filter windows to obtain the mel spectrum. The frequency, H m k , of the triangular filter is calculated as shown in Equation (5).
H m k = 0 , k f m 1 f m f m 1 f m + 1 k f m + 1 f m , 0 ,   ,
f(m) is the central frequency of the mel triangle filter. The logarithmic energy spectrum of each frame is S(m), which is obtained using a logarithmic process, as shown in Equation (6).
S m = ln [ N 1 K = 0 P k H m k ] , 0 m M
P k is the power spectrum, H m k is the filter window, and M is the number of filter windows.
This study used Librosa in Python software to extract the feature of the inverse coefficient of the mel frequency. Figure 4a is the voice signal. After extracting the speech signal samples and features, the mel spectrum is shown in Figure 4b.

2.5. Deep Learning Algorithms

2.5.1. CNN Model

The CNN model can be used to detect the critical features of the audio in the audio message [13]. The CNN model’s output and input architecture is shown in Figure 5, and the core CNN model is explained as follows.
The CNN model uses the convolution layer to retain the original feature arrangement of the image and obtain some essential features from the image. Then, the max pooling layer is used to select the more intense feature values from the essential features and shave the weak ones. This study adopted a rectified linear unit (Relu) to shave off the eigenvalues less than 0 at the site to speed up model training between the convolution layer and the max pooling layer. Then, the feature values are converted into one-dimensional data through the flatten layer to facilitate the subsequent use of the fully connected layer. Finally, the activation function of softmax is connected to the classification output. Table 1 shows the parameter settings of the CNN model in this study.

2.5.2. LSTM Model

Speech is a typical temporal signal because the LSTM (long short-term memory) model has a solid temporal ability [14]. The output and input architecture of the LSTM model is shown in Figure 6. In this study, a four-layer LSTM was used as the input layer, and a four-layer dropout was added to prevent the over-fitting problem of the model in the training process. A dense layer was used for dimensional transformation, and softmax was used for the classification output. Table 2 shows the parameter settings of the LSTM model in this study.

2.5.3. CNN-LSTM Model

CNN combined with LSTM for speech detection is an efficient and accurate hybrid model [15]. The CNN-LSTM model uses a CNN convolution layer to retain the original feature arrangement of an image and obtain some essential features from the image. Then, the max pooling layer is used to select the more intense feature values from the essential features and shave the weak ones. Between the convolution layer and the max pooling layer, the rectified linear unit (Relu) is provided to shave off feature values less than 0 to speed up model training. Then, the LSTM is connected to capture the temporal dynamics of the sequence, and the flatten layer is connected to convert the feature values into one-dimensional data. Finally, the activation function of softmax is connected for classification output. The output and input architecture of the CNN-LSTM model is shown in Figure 7. Table 3 shows the parameter settings of the CNN-LSTM model in this study.

2.5.4. CNN-GRU Model

CNN combined with GRU was used as a classifier in the study of speech enhancement [16] and android botnet detection [17]. In GRU architecture, fewer parameters need to be set, and it is simpler than LSTM architecture [18]. Therefore, it becomes natural to use GRU to optimize the CNN model. However, the combination of CNN and GRU is not always the same. This study combines the studies by Hasannezhad et al. [16] and Yerima et al. [17] into a different CNN-GRU model and writes programs through Python’s Keras package for experiments. The output and input architecture of the CNN-GRU model proposed in this study is shown in Figure 8.
The CNN-GRU model proposed in this study uses a CNN convolutional layer to retain the original feature arrangement of an image and obtains some essential features from the image. In addition, the max pooling layer is used to select more intense feature values from important features and shave the weak feature values, which can prevent the problem of over-fitting the model. Between the convolutional layer and the max pooling layer, this study also used a rectified linear unit to shave off the eigenvalues less than 0 to accelerate model training. Then, the eigenvalues are passed through the update gate and reset gate of the gated recurrent unit (GRU) to increase the calculation speed of the model so that the model can be more accurate. Then, the flattened layer is connected to convert the feature value into one-dimensional data, which is convenient for the subsequent use of the fully connected layer. Finally, softmax’s activation function is connected as the output to determine whether the speech audio is dysarthria. Table 4 shows the parameter settings of the CNN-GRU model in this study.

2.6. Experimental Design

In this study, the word audio of dysarthria patients and healthy subjects was converted into the frequency domain by short-time Fourier transformation, and the mel spectrum image was extracted by the mel-frequency cepstral coefficient as the input of the four models, including CNN, LSTM, CNN-LSTM, and CNN-GRU, proposed in this study. The pros and cons of each model of dysarthria detection were compared by the training and validation sets and the test results. The dataset was divided into the training set, validation set, and test set, which were used for the training and testing of the four deep learning models. The distribution of data was based on the ratio of 0.7:0.15:0.15. In this study, after comprehensive area testing, these parameters had different batch sizes and learning rates to obtain the ideal solution. In the case of batch sizes of 32, 64, and 128, learning rates of 0.01, 0.001 and 0.0001, and Epoch = 10, the experimental results are described in detail in Section 3.

2.7. Model Evaluation

In this study, the effectiveness of the deep learning models was evaluated by the following evaluation indicators, which are generally divided into four types: (1) true positive (TP); (2) true negative (TN); (3) false positive (FP); and false negative (FN). The following evaluation metrics of the models can be calculated and learned based on those four results: accuracy, precision, recall, f1-score, and ROC curve [19,20].

3. Experimental Results

3.1. Experimental Results of CNN Model

The CNN model in this study adopted Keras in Python software for model training [21], and the classification results of the CNN model are shown in Table 5. If the CNN model parameter value of the batch size was set as 128 and the learning rate was set as 0.01, the highest accuracy of 94.36% of the CNN model could be obtained. In this study, Scikit-Learn [20] in Python software was used to draw the ROC curve of the CNN model in Figure 9, from which it can be seen that the AUC of the CNN model was 0.871 and the classification result of the model was good.
In this study, the epoch parameter value of the CNN model was set as 10, and the execution time, loss function, and accuracy of the CNN model can be observed from the training process in Table 6. The accuracy of the test set was 94.36%. It only took about 3 ms/epoch to train the CNN model. The accuracy of the final training set was 97.88%, and the loss function was 0.0638.

3.2. Experimental Results of LSTM Model

The LSTM model in this study adopted Keras in Python software for model training [21]. The classification results of the LSTM model are shown in Table 7. In the LSTM model, if the parameter value of the batch size was set as 64 and the learning rate was set as 0.001, the LSTM model could achieve the highest accuracy of 56.61%. In this study, Scikit-Learn [22] in Python software was used to draw the ROC curve of the LSTM model in Figure 10, from which it can be seen that the AUC of the LSTM model was 0.670 and the classification result of the model was, in general, poor. According to the area under an ROC curve (https://darwin.unmc.edu/dxtests/roc3.htm, accessed on 27 September 2022), AUC is divided into five grades: 0.9–1 = excellent (A), 0.80–0.90 = good (B), 0.70–0.80 = fair (C), 0.60–0.70 = poor (D), and 0.50–0.60 = fail (F). Since the AUC result of LSTM was 0.67, the effect of LSTM was considered poor in this study.
In this study, the epoch parameter values of the LSTM model were set as 10, and the execution time, loss function, and accuracy of the LSTM model can be observed from the training process data in Table 8. The training time of the LSTM model was only about 2 ms/epoch. The accuracy of the final training set was 56.60%, and the loss function was 0.7562. The accuracy of the test set was 56.61%.

3.3. Experimental Results of CNN-LSTM

The CNN-LSTM model in this study adopted Keras in Python software for model training [21]. The classification results of the CNN-LSTM model are shown in Table 9. In the CNN-LSTM model, if the parameter value of the batch size was set as 128 and the learning rate was set as 0.01, the CNN-LSTM model could obtain the highest accuracy of 78.57%. In this study, Scikit-Learn [22] in Python software was used to draw the ROC curve of the CNN-LSTM model in Figure 11. It can be seen from Figure 11 that the AUC of the CNN-LSTM model was 0.758 and the classification result of the model was above medium.
In this study, the epoch parameter value of the CNN-LSTM model was set as 10, and the execution time, loss function, and accuracy of the CNN-LSTM model can be observed from the training process data in Table 10. The training time of the CNN-LSTM model was only about 4 to 8 ms/epoch, and the accuracy of the final training set was 84.21%. The loss function was 0.2745, and the test set accuracy was 78.57%.

3.4. Experimental Results of CNN-GRU Model

The CNN-GRU model in this study adopted Keras in Python software for model training [21]. The classification results of the CNN-GRU model are shown in Table 11. In the CNN-GRU model, if the parameter value of the batch size was set as 128 and the learning rate was set as 0.001, the highest accuracy of 98.88% of the CNN-GRU model could be obtained. In this study, Scikit-Learn [22] in Python software was used to draw the research results of the ROC curve of the CNN-GRU model in Figure 12. It can be seen that the AUC of the CNN-GRU model was 0.916 and the model classification results were excellent.
In this study, the epoch parameter value of the CNN-GRU model was set as 10, and the execution time, loss function, and accuracy of the CNN-GRU model can be observed from the training process data in Table 12. It only took about 2 ms/epoch to train the CNN-GRU model, and the accuracy of the final training set was 98.14%. The loss function was 0.1621, and its test set accuracy was 98.38%.

4. Discussion of Results

According to the experimental results in Section 3, the accuracy values of the CNN model training set and test set were 97.88% and 94.36%, respectively (Table 6). The accuracy of the LSTM model training set and test set was 56.61% (Table 8). The accuracy of the CNN-LSTM model training set was 84.21%, and the accuracy of the test set was 78.57% (Table 10). Finally, the accuracy values of the proposed CNN-GRU model training set and test set were 98.14% and 98.38%, respectively (Table 12). Regardless of the perspective of the training set and test set, the CNN-GRU model had the highest accuracy. In the judgment of the AUC value, the AUC = 0.916 of the CNN-GRU model was also the highest, which was better than the other three models. Various evaluation metrics show that the proposed CNN-GRU model can obtain more accurate judgment results in dysarthria detection.
The results of this study are compared with other methods used in previous studies and summarized in Table 13. Hernandez et al. [23] used a method based on fricative sounds in audio messages and machine learning to detect dysarthria. The average spectral peak in the spectral moment was used to extract the fricatives in the audio as the input features of the SVM model, and the final SVM accuracy was 72%. Narendra et al. [24] trained an SVM with acoustic and glottic features extracted from coded speech utterances and their corresponding dysarthria/health labels and finally achieved an accuracy of 96.38% from the SVM. Narendra et al. [25] developed an end-to-end system that mainly used raw speech signals and raw glottal flow waveforms to detect dysarthria in two deep learning architectures: CNN-MLP and CNN-LSTM. The results showed that the original glottal flow waveform is more suitable for model training than the original speech signal, and the accuracy of CNN-MLP and CNN-LSTM were 87.93% and 77.57%, respectively. Rajeswari et al. [26] enhanced the speech by variational mode decomposition and fed the reconstructed signal to CNN for model training, and the final result achieved 95.95% accuracy. The accuracy of the CNN-GRU model proposed in this study was 98.38%, which is the highest in all studies. However, our approach may take a longer time to execute. After a survey, previous studies have not reported the execution times in their articles. Therefore, an execution time of 2 ms for our approach is appended in Table 13 for further investigation or comparison in the future.

5. Conclusions

Although dysarthria testing can be based on the subjective judgment of doctors, it is also regarded as a costly and time-consuming test, which can easily cause a medical burden. Therefore, if dysarthria testing can be conducted objectively, it can assist doctors in making an immediate judgment. This study used a CNN-GRU classification model for dysarthria detection. The results showed that the proposed CNN-GRU model can achieve the highest accuracy of 98.38%, which is better than the CNN, LSTM, CNN-LSTM models and those of other scholars.
The results can be used as an auxiliary diagnostic procedure for detecting dysarthria in the future. In future studies, it may be possible to take more eigenvalues from audio to analyze the severity level of dysarthria symptoms so that dysarthria detection can be further studied. In addition, others can also use the CNN-GRU model to detect other speech pathologies, such as Parkinson’s disease, amyotrophic lateral sclerosis (ALS), and other symptoms of speech detection. The proposed architecture can also be used for image identification, just as Priyanka and Ganesan [26] used different data preprocessing methods combined with machine learning to classify the severity of dementia. Better prediction results may be achieved if the research is conducted through deep learning architecture.
In addition, most of the existing freely available dysarthric speech databases, including [12], contain speech data recorded from a small number of patients [24]. The volume of speech samples recorded in the dataset used in this study is quite large. However, the number of samples included is not immense, and there has been no continuous addition of samples, which makes it challenging to ensure that the results of this study can be adequately transferred to other clinical trials of dysarthria.

Author Contributions

Conceptualization, D.-H.S. and M.-H.S.; Data curation, C.-H.L. and X.-Y.X.; Formal analysis, T.-W.W. and X.-Y.X.; Funding acquisition, D.-H.S.; Investigation, C.-H.L., T.-W.W. and M.-H.S.; Methodology, D.-H.S. and X.-Y.X.; Project administration, D.-H.S.; Resources, C.-H.L.; Software, X.-Y.X.; Validation, C.-H.L.; Visualization, M.-H.S.; Writing—original draft, T.-W.W.; Writing—review and editing, M.-H.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the Taiwan Ministry of Science and Technology (grant MOST 111-2410-H-224-006). The funder had no role in study design, data collection and analysis, the decision to publish, or the preparation of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gentil, M.; Pollak, P.; Perret, J. Parkinsonian dysarthria. Rev. Neurol. 1995, 151, 105–112. [Google Scholar] [PubMed]
  2. Rampello, L.; Rampello, L.; Patti, F.; Zappia, M. When the word doesn’t come out: A synthetic overview of dysarthria. J. Neurol. Sci. 2016, 369, 354–360. [Google Scholar] [CrossRef] [PubMed]
  3. Marmor, S.; Horvath, K.J.; Lim, K.O.; Misono, S. Voice problems and depression among adults in the U nited S tates. Laryngoscope 2016, 126, 1859–1864. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Van Nuffelen, G.; Middag, C.; De Bodt, M.; Martens, J. Speech technology-based assessment of phoneme intelligibility in dysarthria. Int. J. Lang. Commun. Disord. 2009, 44, 716–730. [Google Scholar] [CrossRef] [PubMed]
  5. Vashkevich, M.; Rushkevich, Y. Classification of ALS patients based on acoustic analysis of sustained vowel phonations. Biomed. Signal Process. Control 2020, 65, 102350. [Google Scholar] [CrossRef]
  6. Muhammad, G.; Alsulaiman, M.; Ali, Z.; Mesallam, T.A.; Farahat, M.; Malki, K.H.; Al-Nasheri, A.; Bencherif, M.A. Voice pathology detection using interlaced derivative pattern on glottal source excitation. Biomed. Signal Process. Control 2017, 31, 156–164. [Google Scholar] [CrossRef]
  7. Karan, B.; Sahu, S.S.; Mahto, K. Parkinson disease prediction using intrinsic mode function based features from speech signal. Biocybern. Biomed. Eng. 2019, 40, 249–264. [Google Scholar] [CrossRef]
  8. Moro-Velazquez, L.; Gómez-García, J.A.; Godino-Llorente, J.I.; Villalba, J.; Orozco-Arroyave, J.R.; Dehak, N. Analysis of speaker recognition methodologies and the influence of kinetic changes to automatically detect Parkinson’s Disease. Appl. Soft Comput. 2018, 62, 649–666. [Google Scholar] [CrossRef]
  9. Albaqshi, H.; Sagheer, A. Dysarthric Speech Recognition using Convolutional Recurrent Neural Networks. Int. J. Intell. Eng. Syst. 2020, 13, 384–392. [Google Scholar] [CrossRef]
  10. Narendra, N.P.; Alku, P. Glottal source information for pathological voice detection. IEEE Access 2020, 8, 67745–67755. [Google Scholar] [CrossRef]
  11. Schlauch, R.S.; Anderson, E.S.; Micheyl, C. A demonstration of improved precision of word recognition scores. J. Speech, Lang. Heart Res. 2014, 57, 543–555. [Google Scholar] [CrossRef]
  12. Kim, H.; Hasegawa-Johnson, M.; Perlman, A.; Gunderson, J.; Huang, T.S.; Watkin, K.; Frame, S. Dysarthric speech database for universal access research. Interspeech 2008, 2008, 480. [Google Scholar] [CrossRef]
  13. Dumane, P.; Hungund, B.; Chavan, S. Dysarthria Detection Using Convolutional Neural Network. Techno-Soc. 2021, 2020, 449–457. [Google Scholar] [CrossRef]
  14. Gers, F.; Schmidhuber, E. LSTM recurrent networks learn simple context-free and context-sensitive languages. IEEE Trans. Neural. Netw. 2001, 12, 1333–1340. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Chaiani, M.; Selouani, S.A.; Boudraa, M.; Yakoub, M.S. Voice disorder classification using speech enhancement and deep learning models. Biocybern. Biomed. Eng. 2022, 42, 463–480. [Google Scholar] [CrossRef]
  16. Hasannezhad, M.; Ouyang, Z.; Zhu, W.P.; Champagne, B. An integrated CNN-GRU framework for complex ratio mask estimation in speech enhancement. In Proceedings of the 2020 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Auckland, New Zealand, 7–10 December 2020; pp. 764–768. [Google Scholar]
  17. Yerima, S.; Alzaylaee, M.; Shajan, A.; Vinod, P. Deep learning techniques for android botnet detection. Electronics 2021, 10, 519. [Google Scholar] [CrossRef]
  18. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint 2014, arXiv:1412.3555. [Google Scholar]
  19. Goutte, C.; Gaussier, E. A probabilistic interpretation of precision, recall and F-score, with implication for evaluation. In European Conference on Information Retrieval; Springer: Berlin/Heidelberg, Germany, 2005; pp. 345–359. [Google Scholar]
  20. Fawcett, T. An Introduction to ROC analysis. Pattern Recogn. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  21. Vasilev, I.; Slater, D.; Spacagna, G.; Roelants, P.; Zocca, V. Python Deep Learning: Exploring Deep Learning Techniques and Neural Network Architectures with Pytorch, Keras, and TensorFlow; Packt Publishing Ltd.: Birmingham, UK, 2019. [Google Scholar]
  22. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Duchesnay, E. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  23. Hernandez, A.; Chung, M. Dysarthria classification using acoustic properties of fricatives. In Proceedings of Seoul International Conference on Speech Sciences (SICSS) 2019, Seoul, Korea, 15–16 November 2019. [Google Scholar]
  24. Narendra, N.; Alku, P. Dysarthric speech classification from coded telephone speech using glottal features. Speech Commun. 2019, 110, 47–55. [Google Scholar] [CrossRef]
  25. Rajeswari, R.; Devi, T.; Shalini, S. Dysarthric Speech Recognition Using Variational Mode Decomposition and Convolutional Neural Networks. Wirel. Pers. Commun. 2021, 122, 293–307. [Google Scholar] [CrossRef]
  26. Priyanka, A.; Ganesan, K. Radiomic features based severity prediction in dementia MR images using hybrid SSA-PSO optimizer and multi-class SVM classifier. IRBM, 2022; in press. [Google Scholar]
Figure 1. Flowchart of dysarthria detection.
Figure 1. Flowchart of dysarthria detection.
Healthcare 10 01956 g001
Figure 2. Amplitude waveforms of (a) dysarthria01 and (b) healthy01 subjects.
Figure 2. Amplitude waveforms of (a) dysarthria01 and (b) healthy01 subjects.
Healthcare 10 01956 g002
Figure 3. Short-time Fourier spectra of (a) dysarthria01 and (b) healthy01 subjects.
Figure 3. Short-time Fourier spectra of (a) dysarthria01 and (b) healthy01 subjects.
Healthcare 10 01956 g003
Figure 4. Feature extraction: (a) voice signal and (b) mel spectrum.
Figure 4. Feature extraction: (a) voice signal and (b) mel spectrum.
Healthcare 10 01956 g004
Figure 5. Architecture of CNN model.
Figure 5. Architecture of CNN model.
Healthcare 10 01956 g005
Figure 6. Architecture of LSTM model.
Figure 6. Architecture of LSTM model.
Healthcare 10 01956 g006
Figure 7. Architecture of CNN-LSTM model.
Figure 7. Architecture of CNN-LSTM model.
Healthcare 10 01956 g007
Figure 8. Architecture of CNN-GRU model.
Figure 8. Architecture of CNN-GRU model.
Healthcare 10 01956 g008
Figure 9. ROC curve of CNN model.
Figure 9. ROC curve of CNN model.
Healthcare 10 01956 g009
Figure 10. ROC curve of LSTM model.
Figure 10. ROC curve of LSTM model.
Healthcare 10 01956 g010
Figure 11. ROC curve of CNN-LSTM model.
Figure 11. ROC curve of CNN-LSTM model.
Healthcare 10 01956 g011
Figure 12. ROC curve of CNN-GRU model.
Figure 12. ROC curve of CNN-GRU model.
Healthcare 10 01956 g012
Table 1. Parameters of CNN model.
Table 1. Parameters of CNN model.
NetworkLayerNo. of ActivationsNo. of Parameters
CNNCov1(27,27,32)160
Maxpooling1(13,13,32)0
Cov2(12,12,64)8256
Maxpooling2(6,6,64)0
Flatten(2304)0
Dense(3)387
Table 2. Parameters of LSTM model.
Table 2. Parameters of LSTM model.
NetworkLayerNo. of ActivationsNo. of Parameters
LSTMLSTM(26,10)480
Dropout(26,10)0
LSTM(26,10)840
Dropout(26,10)0
LSTM(26,10)840
Dropout(26,10)0
LSTM(26,10)840
Dropout(26,10)0
Dense(26,2)22
Table 3. Parameters of CNN-LSTM model.
Table 3. Parameters of CNN-LSTM model.
NetworkLayerNo. of ActivationsNo. of Parameters
CNN-LSTMCov1(23,32)320
Maxpooling1(11,32)0
Cov2(11,64)14,400
Maxpooling2(5,64)0
LSTM(2,128)20,608
Flatten(64)0
Dense(44)2860
Table 4. Parameters of CNN-GRU model.
Table 4. Parameters of CNN-GRU model.
NetworkLayerNo. of ActivationsNo. of Parameters
CNN-GRUCov1(23,32)320
Maxpooling1(11,32)0
Cov2(11,64)14,400
Maxpooling2(5,64)0
GRU(2,32)15,552
Flatten(64)0
Dense(44)2860
Table 5. Classification results of CNN model.
Table 5. Classification results of CNN model.
ModelBatch SizeLearning RateAccuracy%Precision%RecallF1-Score
CNN320.170.0070.010.70020.7010
0.0188.8969.440.83340.7575
0.00175.0075.500.75000.7506
640.188.8886.660.83330.8148
0.0189.8987.500.83330.8285
0.00186.6779.990.6660.6249
1280.193.3387.500.83330.8285
0.0194.3690.390.89130.8896
0.00194.3586.660.83330.8148
Table 6. Execution time, loss function, and accuracy of CNN model.
Table 6. Execution time, loss function, and accuracy of CNN model.
EpochExecution Time (ms)Accuracy
(Training) (%)
Loss FunctionAccuracy
(Validation) (%)
Accuracy
(Testing) (%)
1382.230.497245.4079.20
2391.260.212983.4190.27
3393.420.167490.3591.27
4394.180.144492.6492.30
5394.860.131493.9093.24
6395.770.113494.8793.50
7396.600.093195.4893.52
8396.520.089496.3794.27
9397.210.075797.1394.20
10397.880.063897.5394.36
Table 7. Classification results of LSTM model.
Table 7. Classification results of LSTM model.
ModelBatch SizeLearning RateAccuracy%Precision%RecallF1-Score
LSTM320.150.3250.100.50010.5002
0.0154.2953.210.53210.5321
0.00154.6754.600.54600.5430
640.155.6044.210.44210.4421
0.0154.3054.120.54200.5411
0.00156.6153.430.54350.5324
1280.155.2144.250.65500.5220
0.0156.6043.210.55210.4201
0.00155.3750.200.50200.5020
Table 8. Execution time, loss function, and accuracy of LSTM model.
Table 8. Execution time, loss function, and accuracy of LSTM model.
EpochExecution Time (ms)Accuracy
(Training) (%)
Loss FunctionAccuracy
(Validation) (%)
Accuracy
(Testing) (%)
1253.320.734654.8953.20
2253.630.736055.3253.39
3254.290.737555.3653.65
4254.420.739454.2254.30
5254.680.715355.8954.39
6254.560.716355.9054.37
7254.040.730955.9154.49
8255.600.731656.0155.60
9256.010.755756.4355.97
10256.600.756256.4256.61
Table 9. Classification results of CNN-LSTM model.
Table 9. Classification results of CNN-LSTM model.
ModelBatch SizeLearning RateAccuracy%Precision%RecallF1-Score
CNN-LSTM320.162.4965.990.66660.6549
0.0166.6664.440.66560.6333
0.00173.2068.540.67560.6723
640.170.2169.450.62300.7165
0.0169.2069.440.73330.7175
0.00173.2167.540.67400.6740
1280.175.3070.150.65630.7490
0.0178.5770.330.66600.7500
0.00177.3369.440.74750.7375
Table 10. Execution time, loss function, and accuracy of CNN-LSTM model.
Table 10. Execution time, loss function, and accuracy of CNN-LSTM model.
EpochExecution Time (ms)Accuracy
(Training) (%)
Loss FunctionAccuracy
(Validation) (%)
Accuracy
(Testing) (%)
1542.110.949350.0043.50
2557.890.801066.6750.65
3563.160.672066.6751.27
4873.680.561766.6765.90
5484.210.336783.3366.37
6884.210.325683.3367.47
7584.210.310283.3370.30
8684.210.306083.3375.98
9584.210.266583.3376.35
10584.210.274583.3378.57
Table 11. Classification results of CNN-GRU model.
Table 11. Classification results of CNN-GRU model.
ModelBatch SizeLearning RateAccuracy%Precision%RecallF1-Score
CNN-GRU320.192.2793.210.91210.9220
0.0194.5294.230.94220.9420
0.00195.2193.200.92200.9231
640.196.4195.510.94210.9412
0.0196.7090.240.90260.9633
0.00196.3896.310.94270.9532
1280.197.7196.210.96210.9621
0.0198.0290.470.90300.9021
0.00198.8891.470.91470.9147
Table 12. Execution time, loss function, and accuracy of CNN-GRU model.
Table 12. Execution time, loss function, and accuracy of CNN-GRU model.
EpochExecution Time (ms)Accuracy
(Training) (%)
Loss FunctionAccuracy
(Validation) (%)
Accuracy
(Testing) (%)
1279.200.15790.7789.21
2292.270.126791.0890.20
3294.520.335390.9793.45
4295.210.293791.1694.60
5296.410.155390.7795.88
6296.700.127491.3697.56
7296.830.102991.2096.30
8297.710.239691.4097.13
9298.020.208491.6397.79
10298.140.162191.5298.38
Table 13. Performance comparison.
Table 13. Performance comparison.
AuthorClassification MethodDatasetAccuracy (%)Execution Time
Hernandez et al. (2019) [23]SVMUA-Speech72%-
Narendra et al. (2019) [24]SVMUA-Speech96.38%-
Narendra et al. (2020) [10]CNN-MLPUA-Speech87.93%-
CNN-LSTM77.57%-
Rajeswari et al. (2022) [25]CNNUA-Speech95.95%-
Our ApproachCNN-GRUUA-Speech98.38%2 ms
(-: indicates unknown or uncertainty).
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shih, D.-H.; Liao, C.-H.; Wu, T.-W.; Xu, X.-Y.; Shih, M.-H. Dysarthria Speech Detection Using Convolutional Neural Networks with Gated Recurrent Unit. Healthcare 2022, 10, 1956. https://doi.org/10.3390/healthcare10101956

AMA Style

Shih D-H, Liao C-H, Wu T-W, Xu X-Y, Shih M-H. Dysarthria Speech Detection Using Convolutional Neural Networks with Gated Recurrent Unit. Healthcare. 2022; 10(10):1956. https://doi.org/10.3390/healthcare10101956

Chicago/Turabian Style

Shih, Dong-Her, Ching-Hsien Liao, Ting-Wei Wu, Xiao-Yin Xu, and Ming-Hung Shih. 2022. "Dysarthria Speech Detection Using Convolutional Neural Networks with Gated Recurrent Unit" Healthcare 10, no. 10: 1956. https://doi.org/10.3390/healthcare10101956

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop