Next Article in Journal
Deep Learning and Cloud-Based Computation for Cervical Spine Fracture Detection System
Next Article in Special Issue
Monocular Depth Estimation for 3D Map Construction at Underground Parking Structures
Previous Article in Journal
Learning Peer Recommendation Based on Weighted Heterogeneous Information Networks on Online Learning Platforms
Previous Article in Special Issue
Self-Supervised Skill Learning for Semi-Supervised Long-Horizon Instruction Following
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Knowledge-Based Features for Speech Analysis and Classification: Pronunciation Diagnoses

1
Department of Electrical Engineering, Northern Illinois University, DeKalb, IL 60115, USA
2
Departmment of Electrical and Computer Engineering, University of Victoria, Victoria, BC V8W 2Y2, Canada
3
School of Allied Health and Communicative Disorders, Northern Illinois University, DeKalb, IL 60115, USA
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(9), 2055; https://doi.org/10.3390/electronics12092055
Submission received: 27 March 2023 / Revised: 22 April 2023 / Accepted: 25 April 2023 / Published: 29 April 2023
(This article belongs to the Special Issue AI in Knowledge-Based Information and Decision Support Systems)

Abstract

:
Accurate pronunciation of speech sounds is essential in communication. As children learn their native language, they refine the movements necessary for intelligible speech. While there is variability in the order of acquisition of speech sounds, there are some sounds that are more complex and are later developing. The rhotic /r/ is a later-developing sound in English, and some children require intervention to achieve accurate production. Additionally, individuals learning English as a second language may have difficulty learning accurate /r/ production, especially if their native language does not have an /r/, or the /r/ they produce is at a different place of articulation. The goal of this research is to provide a novel approach on how a knowledge-based intelligence program can provide immediate feedback on the accuracy of productions. In the proposed approach, the audio signals will first be detected, after which features of audio signals will be extracted, and finally, knowledge-based intelligent classification will be performed. Based on the obtained knowledge and application scenarios, novel features are proposed and used to classify various speaker scenarios.

1. Introduction of English Pronunciation

1.1. Native Speakers and Second Language Speakers

Typically, native speakers of a language can quickly identify second language learners due to variations in the stress patterns and prosody of speech and the articulation of individual sounds [1]. These variations occur because the language learner must overcome interference patterns from their first language. In addition to learning which syllable must be stressed in a word, the learner must refine and create additional phonemic categories, focusing on acoustic characteristics of sounds that may not have been distinctive in their first language. Once the learner can hear the acoustic differences between the new phonemes, they must determine the articulatory placements required to pronounce them. This is not an easy task, and characteristics of the native and second language can make it even more difficult. A study indicates that it is challenging to change the accent of non-standard English speakers, even after engaging in a 10-month full-time English class [2]. Further, not all phonemes are created equally, with some being more frequent cross-linguistically, making them easier for language learners to incorporate in their speech. Other phonemes may be articulatorily complex or may be acoustically very similar to or distinct from other sounds in their native phonetic inventory, making them more challenging to incorporate in their phonemic repertoire [3].

1.2. Method to Pronounce Phonetic /r/

In English, there are two main methods to produce the /r/, and while the location of the tongue is very different between the two, the resulting sound is similar. One method of production is referred to as bunched. The back of the tongue is lifted, thereby putting the sides of the tongue in line with the back teeth. The middle of the tongue’s back is lower and the air flows to build the resonance across this groove. The tip of the tongue will point upwards or stay down. The second method is considered a retroflex. The tip of the tongue is elevated and curves backward behind the tooth ridge while the back of the tongue remains low [4].
When it comes to Asian languages, pronunciation of the English /r/ is challenging because /r/ is not common in these languages. For example, in Japanese, there are five characters in the Romanization of Japanese “ra, ri, ru, re, ro.” In Japanese, /r/ and /l/ are allophones in free variation, meaning they can be substituted with each other without changing the meaning of the words. Thus, Japanese speakers do not need to focus on the acoustic distinctions between /r/ and /l/ because they are not distinct phonemic categories; /li/ and /ri/ can both be understood as “ri.” When Japanese speakers learn English, they are challenged to reconfigure their underlying representations of sound categories and differentiate /r/ and /l/ when speaking English and ignore the distinction when speaking Japanese. The complexity of production and many phonetic contexts of the American English /r/ make accurate production challenging even after years of speaking English.

1.3. Speech and Language Disorder and Assessment

Speech disorders can affect pronunciation, which influences their everyday lives. While distortion errors may not affect the understanding of speech as severely as omission or substitution, it does not mean distortion errors should be ignored, because they can result in negative academic [5] as well as emotional and social consequences in children [6,7].
Further, while the cause of the delay is not always known, there are other diagnoses that indicate intervention is necessary before the child even starts to speak. For example, children with hearing loss are deprived from the acoustic signal. Without amplification, they will have extreme difficulty creating the phonemic categories of their language. Thus, early intervention is critical [8].
Because some children with speech sound errors also have a language impairment [9], diagnosis of a speech sound disorder typically occurs after a comprehensive speech-language evaluation. The evaluation will consist of a hearing test and an oral-peripheral test to rule out organic causes for the speech errors. A variety of standardized assessments will be utilized to test the child’s understanding and use all of the domains of language: semantics (vocabulary), syntax (word order), morphology (base words, prefixes, and suffixes), phonology (sound system) and pragmatics (social communication). A conversational speech sample will allow for voice and fluency to be screened as well as additional conversational speech-language analyses to be completed. Common articulation and phonology measures extracted from the conversational speech-language sample are phonetic and syllable inventories, as well as a calculation of percent consonants correct (PCC). An important component of all speech sound evaluations is stimulability testing, which will indicate if the individual has the capacity to produce the sound with perceptual, instructional, or tactile cueing [10].

1.4. Purpose and Proposed Work

The current project aims to evaluate the effectiveness of using signal processing and pattern recognition in remediating /r/ mispronunciation. A crucial aspect of addressing /r/ errors is training individuals to perceptually distinguish between correct and incorrect productions. Once individuals can differentiate between the two, they can start modifying their tongue movements to approximate accurate /r/ pronunciation. Speech-language pathologists play a crucial role in guiding individuals on how to move their tongue correctly and provide feedback on the accuracy of their production. They may also assign homework to facilitate practice and reinforce new motor movements. However, practice may prove counterproductive if the speaker continues to produce the sound incorrectly, thereby increasing the habitual nature of the distorted sound.
To address this challenge, a pronunciation analysis and classification system is proposed to provide individuals with immediate feedback on their productions, enabling them to identify inaccuracies and make necessary modifications. In this project, we propose using speech signal processing, analysis, and pattern recognition techniques to classify pronunciation accuracy. We plan to use /r/ as the pilot study to demonstrate the feasibility of using advanced signal processing and pattern recognition approaches to improve speech production.
Similar to automatic speech recognition (ASR), the proposed analyzing/classifying system includes speech detection, pattern or feature extraction, and classification parts. However, the objective of ASR is to figure out the contents embedded in the speech waveform, whereas the purpose of this study is to identify correct and incorrect pronunciation. Therefore, the typical features used for ASR cannot be directly used without modification. To address this issue, we propose engineering new features for this task.
The mechanism for generating speech in humans involves a complex interaction between various biological structures, including the lungs, larynx, vocal cords, and articulators. The process begins when air is exhaled from the lungs and passes through the larynx, where it causes the vocal cords to vibrate when producing voiced sounds. These vibrations produce a sound wave that travels up the throat and into the mouth (and nose for nasal sounds), where it is shaped into speech sounds by the movement of the tongue, lips, and other articulators. Different speech sounds are created by manipulating the shape of the vocal tract. The movement of the articulators create different filters that modify the sound wave. In order to capture the collaboration and the changes of those structures. We propose the Pitch and Partials Amplitude Gradient (PPAG) feature, which tracks the changes in the fundamental frequency and its harmonics in their magnitude, capturing the subtle variations in the process of speech sound production.
Achieving correct pronunciation requires the precise collaboration of all the biological systems in time and space, resulting in the production of an accurate acoustic waveform. In contrast, incorrect pronunciation arises from errors in one or more of these systems, or from incorrect collaboration among them, resulting in the production of an inaccurate waveform. Incorrect wave forms can differ from one another in terms of their position, momentum, and energy, and are characterized by different probabilities. In contrast, correct wave forms are similar to one another and are produced with greater probability when all the biological systems are properly calibrated and functioning together. These observations suggest that the production of speech may be best understood as a complex quantum mechanical process, where the correct pronunciation corresponds to the quantum state with the highest probability, resulting from the coherent interaction of multiple physical systems. To explore this hypothesis further, we propose borrowing the concept of the Schrödinger equation to model the phenomenon of speech production. We designed the second innovative feature, a speech wave function-based probability amplitude (SWPA) feature from speech signals.
This paper is organized as follows: Section 2 describes the speech sound detection methods; Section 3 reviews the typical features used for ASR; Section 4 proposes innovative features based on calculating the gradients of the pitch, its partials from speech signals, and SWPA features based on wave function; Section 5 presents the classifiers for recognizing the correctness of the speech signals. The experiments and results are shown in Section 6. Section 7 concludes this paper.

2. Speech Signal Detection

Speech signal processing has been a dynamic and constantly evolving field for a couple of decades now. It involves a range of techniques and methods for analyzing, manipulating, synthesizing, interpreting, and recognizing speech signals. Some of the main areas of speech signal processing include speech analysis, speech synthesis, speech coding and compression, speech enhancement, speech recognition and transcription, speaker identification and verification, and language modeling and understanding.
Speech processing has a wide range of real-world uses in diverse aspects of our everyday lives, including communications, consumer electronics, education, entertainment, and healthcare services. For example, speech recognition is used in virtual assistants, speech-to-text transcription is used for dictation and subtitling, speech processing is used in telecommunications and call centers, and speech therapy uses real-time feedback and guidance to help individuals with speech disorders. Additionally, speech processing is used in biometric identification, forensic analysis, and in education to teach foreign languages and improve pronunciation. In entertainment, speech processing is used in voice acting and dubbing, as well as creating synthetic voices for characters in movies and video games. Overall, speech processing has a broad range of applications and continues to play an increasingly important role in our daily lives as technology advances.
Speech signals are non-stationary, meaning that many algorithms designed for stationary signal processing cannot be directly applied to speech without modification. However, research has shown that for short time intervals of 10–30 milliseconds (ms) in length, speech signals can be viewed as stationary signals for certain applications [11]. Additionally, speech signals contain voiced, voiceless, and silent portions, each with different short time characteristics such as power and frequency. Therefore, speech detection and segmentation are widely used in speech signal processing.
Typically, the first step in speech signal processing is speech signal detection, followed by segmentation and other processing steps such as speech coding, feature extraction, and more.

2.1. Short-Time Energy

Short-time energy (STE) is one of the most popular methods for speech detection, as the voiced portion of speech typically has a higher STE, while the voiceless component has a significantly lower STE, and the silent portion has the lowest STE. Additionally, STE can be used to determine the beginning and ending points of each word.
STE is defined as the average of the squared values of a signal’s samples within a given window. The mathematical representation of a window may be defined as follows [12]:
S T E = 1 N m = 0 N 1 [ W ( m ) x ( n m ) ] 2
where w ( m ) is the window coefficient correspond, m is the sample index, and N is the length of a window.
A normalized STE > 0.07 was set as a threshold to remove the silent portion, such as whispering and breathing. The second threshold was set as 0.14 seconds in duration in order to remove some impulsive voiced artifact such as a cough or sneeze.

2.2. Short-Time Magnitude

Additionally, the short-time magnitude is commonly used as a time-frequency representation for audio signals, and usually used to detect speech signals as well, particularly when a signal has a wide dynamic range. Short-time magnitude is described as follows [12]:
M n = m = | x ( m ) | w ( n m )
with | x ( n ) | denoting the signal’s magnitude,

3. General Feature Extraction for Speech Signals

Speech features refer to the characteristics of the speech signal that are extracted and analyzed to represent the speech signal in a more concise and meaningful way. These features are used in various speech processing applications such as speech recognition, speaker identification, and speech synthesis. Some commonly used abstract-based features in speech signal analysis include Linear Predictive Coding (LPC), Linear Predictive Cepstral Coefficients (LPCCs), the Mel-frequency Cepstrum Coefficient (MFCC), and Bark Frequency Cepstral Coefficients (BFCCs).

3.1. Linear Predictive Coding (LPC)

Linear predictive coding (LPC) is a common technique used for speech coding and speech recognition. It models the spectral envelope of speech using a linear predictive model and represents it in digital form [13].
A LPC algorithm computes a vector of coefficients which presents an all-pole filter [14].
H ( z ) = 1 1 i = 1 M a i z i
where a i are the LPC coefficients, and M is the order of the filter.
A speech sample can be predicated by using a linear combination of the past M speech samples
x ^ ( n ) = a 1 x ( n 1 ) + a 2 x ( n 2 ) + + a M x ( n M ) = i = 1 M a i x ( n i )
An autocorrelation method is used to calculate the LPC coefficients a i because it is capable of minimizing the prediction error. The autocorrelation for a speech signal is
R ( k ) = 1 N n = 0 N 1 x ( n ) x ( n k )
where k is the sample delay interval, and N is the number of speech samples used for evaluating. Then, the LPC coefficients can be calculated by solving equation
i = 1 M a i R ( k i ) = R ( k )

3.2. Linear Predictive Cepstral Coefficients (LPCCs)

The cepstrum is a series of numbers that can be used to describe a single frame of an audio signal. The cepstrum can be used in pitch monitoring and speech recognition.
Linear predictive cepstral coefficients (LPCCs) are the modified version of LPC coefficients in the cepstral domain. The cepstral sequence function is an estimate of the “envelope” of the signal [15].

3.3. Mel-Frequency Cepstrum Coefficient

The Mel-frequency Cepstrum Coefficient (MFCC) is a widely used cepstral feature in speech processing and interpretation. It represents the short-term power spectrum of a sound by applying a linear cosine transform to a log power spectrum on a nonlinear Mel-scale of frequency [15].
The calculation of the MFCC typically involves the following steps: pre-emphasis of the high frequency portion, segmentation of the signal into short-duration frames, multiplication of the frames with a desired window, calculation of fast Fourier transform, application of a Mel filter bank, and calculation of discrete cosine transform, as shown in Figure 1 [16]. This procedure is similar to the human auditory system’s perception, which is more sensitive to low-frequency components.
The transfer function that maps the frequency to the Mel-scale frequency can be expressed as follows [17]:
M e l ( f ) = 2595 l o g 10 ( 1 + f 700 )
Figure 2 shows the nonlinear Mel-frequency curve.
Generally, high frequency components in speech signals get attenuated during propagation and transmission. To compensate for this loss, a high pass Finite Impulse Response (FIR) filter is typically used. The rational transfer function for this filter is shown as (8)
y ( n ) = x ( n ) α x ( n 1 )
where α = 0.9 to 1.0 .
After pre-emphasis, the signal is divided into many short frames during the framing process by applying different windows that can reduce frequency leakage [18].
Fast Fourier Transform(FFT): discrete-time Fourier transform converts a speech signal from the time domain to its frequency domain [19]. It is shown as (9).
Y ( k ) = j = 1 n x ( j ) W n ( j 1 ) ( k 1 )
where,
W n = e ( 2 π i ) / n
In the Mel-frequency analysis, a bank of triangular filters is used to compute the power spectrum. Each filter is a triangle-shaped band pass filter with decreasing magnitude towards its edges. The central frequency of each filter is distributed according to the Mel-scale. After passing the signal through the filter bank, the output values are converted to a logarithmic scale to obtain the logarithmic power spectrum [20]. The logarithm formula used for this purpose is
L m = l o g ( k = 0 N 1 | Y ( k ) | 2 H m ( k ) ) , 0 m M
where N is the number of discrete Fourier transform points, and M is the MFCC order.
The last step is using discrete cosine transform (DCT) to transfer the logarithmic power spectrum from a Mel-scaled frequency domain back into its time domain [21]. Equation (11) represents the procedure.
C m ( k ) = 2 N n = 1 N L m ( n ) 1 1 + δ k 1 c o s ( π 2 N ( 2 n 1 ) ( k 1 ) )
For a signal x of length N, and with δ the Kronecker delta, k is the order of the coefficient [22].

3.4. Bark Frequency Cepstral Coefficients (BFCCs)

This technique employs the concept of Bark Frequency Cepstral Coefficients (BFCCs), which is a pitch-perception-based feature extraction process used to classify voice features [23]. The Bark-scale has a scaling factor that is modified by multiplying the initial frequency scale. The formula for converting the frequency to Bark is:
B a r k ( f ) = 13 a r c t a n ( 0.00076 f ) + 3.5 a r c t a n ( ( f 7500 ) 2 )
In BFCCs, filter banks are used in the Bark-scale, and the overall procedure is similar to that used in MFCCs, except for the filter bank transition from the Mel-scale to the Bark-scale.

4. Innovative Features

4.1. Pitch and Partials Amplitude Gradient Feature

We have proposed an innovative feature that enables the tracking of amplitude changes in the fundamental frequencies of speech signals. This algorithm calculates the first-order temporal derivatives of the fundamental frequency and its harmonics, creating a new feature.

4.1.1. Fundamental Frequency and Its Harmonics for Speech Signal

To prepare the speech signal for analysis, we first conduct pre-emphasis and framing. Next, we applied a Hamming window to reduce frequency leakage [24].
w ( n ) = 0.54 0.46 cos ( 2 π n N )
Following the fast Fourier transform of the signal, the amplitude of the pitch and partials can be obtained by analyzing the spectrum of the signal in one frame. We selected the 10 most significant frequencies, sorted them from low to high, and generated a two-dimensional amplitude-frequency matrix, as shown in matrix (14). The matrix contains the amplitudes a 1 a 10 and their corresponding frequencies f 1 f 10 , which are sorted from small to large.
a 1 f 1 a 2 f 2 a 10 f 10

4.1.2. Derivation

During articulation, muscles supporting the articulators (e.g., the lips, tongue, and vocal cords) move, resulting in the production of speech sounds, and the amplitude of each frequency in the speech signal changes accordingly. Therefore, the pronunciation procedure can be monitored by inspecting the amplitude changes. Based on this, we propose a hypothesis that when people pronounce the same phoneme, the amplitude in each frequency has the same trend of change.
To track and quantify this trend, we reshape the amplitude-frequency matrix into a one-dimensional vector and combine all the vectors generated in different frames together. Then, we take the derivative of the amplitude to obtain the derivation matrix. As shown in matrix (15), every column (superscript) represents the elements in the same frame, and every row (subscript) represents elements in the same pitch frequency. Here, n is the number of frames, and d t is the step size used in the framing process, which is relatively short in our case.
d d t a 1 ( 1 ) a 1 ( 2 ) a 1 ( n ) a 2 ( 1 ) a 2 ( 2 ) a 2 ( n ) a 10 ( 1 ) a 10 ( 2 ) a 10 ( n ) 10 × n =
d 1 ( 1 ) ( 2 ) d 1 ( 2 ) ( 3 ) d 1 ( n 1 ) ( n ) d 2 ( 1 ) ( 2 ) d 2 ( 2 ) ( 3 ) d 2 ( n 1 ) ( n ) d 10 ( 1 ) ( 2 ) d 10 ( 2 ) ( 3 ) d 10 ( n 1 ) ( n ) 10 × n 1
We start by normalizing the derivation matrix and the amplitude-frequency matrix separately. We then combine the derivation matrix with the second column of the amplitude-frequency matrix, representing the frequency, into a one-dimensional feature vector. We refer to this feature as the Pitch and Partials Amplitude Gradient Feature.
[ d 1 ( 1 ) ( 2 ) d 1 ( n 1 ) ( n ) f 1 ( 1 ) f 1 ( n ) d 10 ( 1 ) ( 2 ) d 10 ( n 1 ) ( n ) f 10 ( 1 ) f 10 ( n ) ]

4.2. Speech Wave Function Probability Amplitude Feature

4.2.1. Model Speech Signal by Using Concept of Schrödinger Equation

This section is a brief introduction to the Schrödinger equation. The Schrödinger equation is a fundamental equation in quantum mechanics that describes the time evolution of a quantum state.
The Schrödinger equation is a partial differential equation, and its solution gives the probability amplitude of finding the quantum system in a particular state at a given time. The wave function ψ can be used to calculate various physical quantities, such as the probability density of finding a particle in a particular location, or the expectation value of an observable physical-like energy or momentum.
The Schrödinger equation has revolutionized our understanding of the behavior of particles at the quantum level and is essential in many areas of physics, chemistry, and engineering.
Quantum mechanics describes the properties of systems by using a wave function, Ψ ( x , t ) and we get it by solving the Schrödinger equation [25]:
i Ψ t = 2 2 m 2 Ψ x 2 + V Ψ
Here, i is the square root of −1, and is Planck’s constant, which is the original constant h divided by 2 π :
= h 2 π = 1.054573 × 10 34 J s .
Centered on Porn’s mathematical analysis of the wave function, | Ψ ( x , t ) | 2 is the probability of discovering a particle x and the time t [25]. When we look at exponential activity in the time coordinate for the first time, we guess Ψ ( x , t ) = e i ω t f ( x ) and use the method of separation of variables to solve the Schrödinger equation:
Ψ ( x , t ) = ψ ( x ) f ( t )
Ψ t = ψ d f d t , 2 Ψ x 2 = d 2 ψ d x 2 f
So, the Schrödinger equation changes the form to:
i ψ d f d t = 2 2 m d 2 ψ d x 2 f + V ψ f
Dividing by ψ f :
i 1 f d f d t = 2 2 m 1 ψ d 2 ψ d x 2 + V
The left side is the time function t, and the right side is the location function x. It only occurs when all sides are constant and when the divergence constant is labeled E in quantum mechanics.
i ψ d f d t = E ,
or
d f d t = i E f
and
2 2 m 1 ψ d 2 ψ d x 2 + V = E
or
2 2 m d 2 ψ d x 2 + V ψ = E ψ
The equation describing a particle with a definite energy, known as the time-independent Schrödinger equation, is more stringent than the initial Schrödinger time-dependent equation. This is because the time-independent equation culminates in two ordinary differential equations instead of one partial differential equation, with the constant E representing the particle’s total energy (kinetic plus potential) [25]. The general solution of the time-dependent Schrödinger equation, represented by Equation (23), is given by C exp ( i E t / ) , where C is a constant that is absorbed into ψ . Then, we apply
f ( t ) = e i E t / .

4.2.2. Infinite Square Well Model

To solve the time-dependent Schrödinger equation, we can make use of the infinite square well model. This model system in quantum mechanics consists of a particle that is confined to a one-dimensional box with infinite potential energy at the boundaries. It is used to study the behavior of a particle in a potential well with fixed boundaries, which can be an idealization of certain real-world systems, such as electrons in a solid or a gas trapped in a container.
The wave function of the particle in the infinite square well can be found analytically, and it exhibits some interesting properties, such as the quantization of energy levels and the absence of probability density outside of the box. By studying this simple model, we can gain insight into more complex quantum systems, such as atoms and molecules.
V ( x ) = 0 , if 0 x a , , otherwise .
In quantum mechanics, this type of configuration is often referred to as a “particle in a box” (a 1D box) or “square”. The particle is free to move within the specified region, but has zero probability of escaping, much like a particle trapped inside a box. Therefore, the wave function ψ ( x ) is equal to zero outside of the region.
The potential energy is infinite outside of the two boundaries that prevent particles from escaping. This is similar to an elastic ball being placed inside a square well, where it can move freely within the well but cannot escape due to the infinitely high walls of the well [25].
In fact, the particle has zero probability of being detected beyond the range of x values that vary from 0 to L . This is because the particle would need to overcome the exponentially tall and infinitely high theoretical wall at the edge of the box. When the particle is outside of the specified range, ψ ( x ) is equal to zero. Within the range, where V ( x ) = 0 , the time-independent Schrödinger Equation (25) takes the form:
2 2 m d 2 ψ d x 2 = E ψ
or
d 2 ψ d x 2 = k 2 ψ , where k = 2 m E
We assume that E > 0 because the reciprocal of E < 0 is not physically meaningful. Thus, we can approach the problem dynamically, interpreting it as an oscillatory function that can be described using trigonometric functions. The general solution of Equation (29) is given by:
ψ ( x ) = A sin ( k x ) + B cos ( k x )
However, in the context of quantum mechanics, it is not immediately clear what this solution means physically. If the probability distribution were discontinuous, it would be highly problematic to say that there was something fundamentally wrong with the distribution. However, in this case (and in most other cases that are not pathological), it is reasonable to assume that the probability distribution is continuous. Therefore, we can proceed under the assumption that the solution is valid.
Applying the boundary conditions ψ ( 0 ) = ψ ( a ) = 0 , the equation can be rewritten as:
ψ ( 0 ) = A s i n 0 + B c o s 0 = B
So B = 0 , and
ψ ( x ) = A s i n k x .
Then ψ ( a ) = A s i n k a , this only happens when
k a = 0 , ± π , ± 2 π , ± 3 π ,
k n = n π a , with n = 1 , 2 , 3 ,
We can use Equation (29) to determine that k = 2 m E . Since we assume E > 0 , we can write the possible values of E as:
E n = 2 k n 2 2 m = n 2 π 2 2 2 m a 2
where n is a positive integer representing the energy level of the particle.
To ensure that the total probability of finding the particle in the range 0 x a is 1, we must normalize the wave function ψ by fixing the constant ψ as follows:
Ψ n ( x , t ) = 2 a s i n ( n π a x ) e i ( n 2 π 2 / 2 m a 2 ) t

4.3. Speech Wave Function-Based Probability Amplitude Feature

During the process of speech production, the vibrations generated by the vocal cords pass through the vocal tract to create specific sounds. Although every individual is unique, the changes over time and space of these vibrations should be similar when making the same sound. Therefore, the shape and movement of the interaction among all involved biological structures generate an acoustic waveform. This means that all the information is embedded in the time-domain and frequency-domain characteristics with different probabilities. However, the commonly used features for speech signals, such as Mel-frequency cepstral coefficients, do not explicitly capture the probability information of the signal.
To address this issue, we introduce the concept of a frame probability feature in speech signal processing. Frames with a high probability receive a greater coefficient, resulting in a higher degree of discrimination in the feature vector.
To improve the Mel-frequency cepstrum coefficient, we apply the infinite square well model in feature extraction. We can use the concept of probability from the wave function. Just as the infinite square well restricts the particle, the speed of sound limits the audio to a particular range if we determine the length of time. The stationary states Equation (36) is used to calculate the probability in each frame. To calculate this equation, we need to know every frame’s time, position in space, and energy. Figure 3 shows an audio sample of the phoneme /r/ in the time domain, which we recorded near the speaker’s mouth. The characteristics of the audio signal contain all the information embedded in the time domain and frequency domain with different probabilities.
We also need to consider how the audio is distributed in space. Assuming the speed of sound is 340 m/s, we can calculate that the sound wave can travel a distance of 34 m during the 0.1-second duration. Therefore, the sound wave originating from the speaker’s mouth will be located 34 m away at the beginning of the recording. Similarly, the sound wave at the end of the recording will be located at the microphone’s position, which we can assume is at position zero. Figure 4 shows the audio distribution of the same sample in space.
This new feature can be interpreted as a probability-based feature, which is calculated based on the time, position, and energy of each sampling point in a complex manner and provides a way to incorporate the probability information of each frame into the Mel-frequency cepstrum coefficient (MFCC) feature. The procedure consists of pre-emphasis and framing, calculating the wave function Ψ ( x , t ) , and obtaining the probability | Ψ ( x , t ) | 2 . We then add up all probabilities of samples within one frame to obtain the total probability of that frame. We continue with the MFCC procedure, which includes windowing, applying Fast Fourier Transform, applying a Mel-scaled filter bank, and applying Discrete Cosine Transform to obtain the original MFCC features of that frame. Finally, we multiply the MFCC features by the probability of that frame to obtain the new feature.

5. Classification Methods

In this paper, we use a hypothesis H = 1 to represent the classification result of standard English pronunciation of /r/, and use hypothesis H = 0 to represent the classification result of a non-standard English /r/. We use L = 1 to represent speech labeled as standard English speakers, and L = 0 to represent speech labeled as non-standard English speakers.

5.1. K-Nearest Neighbor (KNN)

One of the most basic and straightforward classification techniques is K-nearest neighbor classification. The KNN algorithm works based on the principle that if the majority of the most similar samples in the feature space (i.e., the sample’s closest neighbor in the feature space) belong to the same group, then the sample does as well [26]. The Euclidean distance between a reference sample and the training samples is typically used as the basis for the K-nearest neighbor classifier.
Data in the training set L, are classified by this algorithm into several separate categories, C 1 , , C M , with M as the number of categories. In the jth category, there are samples { ( x 1 ( j ) ) , , ( x n ( j ) ) } , where n is the number of samples and features are N dimension vectors. The expectation of each category will be
μ ( j ) = 1 n i = 1 n x i ( j )
Then, we can calculate the distance of each sample in the testing dataset, X 1 , , X n , to each category expectation, μ ( j ) , using Equation (38).
d j = ( X i μ ( j ) ) 2
Because this is a binary classification, the value of the threshold affects the classification rate. We can investigate the performance of the classifier by changing the threshold and using a receiver operating characteristic curve. The classification logic for a given threshold V i is shown in (39).
H = 1 , if d 1 d 2 V i 0 , if d 1 d 2 > V i

5.2. Gaussian Multivariate Model (GMM)

A Gaussian Mixture Model (GMM) is a statistical model that assumes that the probability distribution of the observed data points comes from a mixture of several Gaussian distributions, each with its own mean and covariance. The model is trained to learn the parameters of these Gaussian distributions and the probabilities of each data point belonging to each of the Gaussian distributions. Once trained, the GMM can be used for a variety of tasks such as clustering, density estimation, and classification (40).
The mean and the covariance matrix of the GMM can be obtained from the training dataset [27].
μ = 1 n i = 1 n x i
Σ = E [ ( x μ ) ( x μ ) T ]
In order to calculate fast, we extract the diagonal elements of Σ and generate a diagonal matrix to replace the original covariance matrix. To expedite calculations, we can extract the diagonal elements of the covariance matrix, denoted by D , and create a diagonal matrix by placing these elements along the main diagonal. This diagonal matrix can then be used to replace the original covariance matrix in computations, leading to faster processing times.
D = σ 1 2 σ 2 2 σ N 2
The probability density function (PDF) of the Gaussian Multivariate Model is shown by the following Equation (43):
p i ( x ; μ , Σ ) = 1 ( 2 π ) N 2 Σ 1 2 e ( 1 2 ( X i μ ) T Σ 1 ( X i μ ) )
After calculating the PDF of a testing sample using Gaussian Mixture Models (GMMs), the testing sample can be classified as belonging to the Gaussian component with the highest posterior probability. In other words, for each Gaussian component in the GMM, we can compute the posterior probability that the testing sample belongs to that component given its observed features. The testing sample is then classified as belonging to the component with the highest posterior probability.

5.3. Artificial Neural Network

An Artificial Neural Network (ANN) is a popular classifier used in various fields such as image recognition, natural language processing, speech recognition, and many others. An ANN is a type of machine learning model that is designed to simulate the behavior of the human brain’s neural networks.
The basic structure of an ANN consists of an input layer, one or more hidden layers, and an output layer. The nodes in the input layer receive the input data, which are then processed by the nodes in the hidden layers to extract important features. The output layer then produces the final output based on the learned features.
The learning process of an ANN involves training the network on a labeled dataset, where the network adjusts its weights and biases to minimize the difference between the predicted output and the true output. This process is typically performed using an optimization algorithm such as backpropagation.
One of the strengths of an ANN is its ability to learn complex patterns and relationships in data, making it suitable for a wide range of classification tasks. Additionally, an ANN can handle noisy or incomplete data, and can generalize unseen data well, making it a powerful tool for machine learning. The activation function in an ANN is (44),
S ( x ) = 1 1 + e x
The ability to learn from training data and generalize results using active function is one of the main characteristics of an artificial neural network; it stores the knowledge gained during the training process in the synaptic weights of the neurons [28].
The synaptic weight of a connection between two nodes is a value that represents the connection between the two nodes. The way nodes are connected, connection weight values, and the activation function all affect the output of a neural network model.
For an ANN, the input layer is the input signal, and the hidden layer and output layer are assembled by cells [29].
For m samples { x ( 1 ) y ( 1 ) x ( m ) y ( m ) }, the cost function of applying a neural network is Equation (45).
J ( θ ) = 1 m [ i = 1 m k = 1 k y k ( i ) l o g ( h θ ( x ( i ) ) ) k + ( 1 y k ( i ) ) l o g ( 1 h θ ( x ( i ) ) ) k ) ] + λ 2 m l = 1 L 1 i = 1 S L j = 1 S L + 1 ( θ j i ( l ) ) 2
where the input data x are represented as features, the output data y represent the label of the training sample, and the weight matrix between the network’s cells is denoted by θ . To prevent overfitting, a regularization parameter λ is introduced. The network typically consists of L layers, with S l representing the number of units in layer l. The objective of the ANN is to find the weight matrix θ that minimizes the cost function.
The backpropagation algorithm is a supervised learning algorithm used to train Artificial Neural Networks (ANNs) to make accurate predictions for a given input. The name “backpropagation” refers to the fact that the algorithm works by propagating the error from the output layer back through the network, adjusting the weights of the connections between neurons as it goes. During the training process, the backpropagation algorithm calculates the error between the network’s output and the desired output, and then adjusts the weights of the connections between neurons in the network to minimize this error. This is typically performed using a gradient descent optimization algorithm, which calculates the gradients of the error with respect to each weight in the network, and then adjusts the weights in the direction of the steepest descent.

6. Experiment and Result

6.1. Data Collection and Labeling

In this project, we recruited 31 human subjects, including 19 native English speakers and 12 nonnative English speakers. All participants were asked to read a file containing 40 words with the phoneme ‘r’ sound in word initial position followed by a variety of vowels. Their speech signals were recorded using a SONY PCM-D50 two-channel linear digital recorder. The sample of a part of the file was shown in the Table 1. The recordings were conducted either in the communication lab at the School of Allied Health and Communicative Disorders, or at the Digital Signal Processing lab, Northern Illinois University. To ensure a quiet recording environment, we aimed to maintain an average ambient noise level of less than 45 dB during the recording procedure. The sampling rate was set at 48 kHz, and a data file with two channels was obtained for each participant.
The duration of the phoneme /r/ in a word is approximately 0.085 s, so we extracted the first 0.1 s of the speech segment from each word as a data sample.
In this research, which is a binary classification problem, we defined two categories for each data instance: ’correct 1’ and ’incorrect 0’. The initial labeling process for the data instances was conducted by a professional speech-language pathologist, who determined whether each instance represented the correct or incorrect pronunciation of the phoneme /r/. Then, we divided the labeled data into training and testing sets. The training set is used to train the machine learning model, while the testing set is used to evaluate the model’s performance.
The first 0.1 s part from an original waveform from a native English speaker is shown in Figure 5.

6.2. Pre-Processing

The pre-emphasizing high pass FIR filter we used is y ( n ) = x ( n ) 0.98 x ( n 1 ) . The audio wave after passing the filter is shown as Figure 6.
We chose the frame length as 40 ms with an overlap rate of 6.3% and we obtained 10 frames from each data sample for analysis. Figure 7 shows the difference in signal between the first frame and tenth frame.
Then, we used a hamming window, shown in Figure 8, to multiply each frame, as shown in Figure 9.

6.3. Feature Extraction

6.3.1. MFCC Features

We extract the MFCC feature, which is one of the the most popular features used for speech recognition, as a comparison baseline. The frame length is 0.025 s, the frame step is 0.01 s and the overlap rate is 60%. For each /r/ production, we get a total of 8 frames and we select 13 coefficients as features in 1 frame. Finally, the feature vector is a vector of 104 elements.

6.3.2. Pitch and Partials Amplitude Gradient Feature

We applied the fast Fourier transform to each frame of the speech signal, and then we searched for the peak amplitude in the frequency domain. As shown in Figure 10, when we compare the FFT figures obtained from the 1st and 10th frames, we can see that the pitch frequency does not change significantly, but the amplitude in each frequency changes dramatically.
We then obtained the amplitude-frequency matrix shown in Equation (14), which is a 10 × 2 matrix. For a particular frequency, we can analyze the trend of amplitude changes across different frames. Figure 11 shows the amplitude changing trend for the pitch frequency of 225 Hz.
Next we calculated the derivation of the amplitude part from each frame with the same pitch frequency as matrix (15). In this case, we have 10 frames, so the derivation matrix will be a 10 × 9 matrix. The derivation matrix quantifies the amplitude changes with the time changes. Figure 12 shows the derivation representing the amplitude changing trend in the 225 Hz pitch frequency.
Finally we calculated the derivation matrix and obtained the pitch and partials amplitude gradient features which include 190 elements.

6.3.3. Speech Wave Function Probability Features

In this experiment, we have 2480 audio samples of the phonetic /r/ pronunciation, each with a duration of 0.1 s. These samples include 1520 labeled ‘1’ and 960 labeled ‘0’. The recorder’s sampling rate is 48,000 Hz. The rational transfer function index α is 0.93. We use a frame length of 0.025 s and a frame step of 0.01 s, resulting in an overlap rate of 60% and the generation of eight frames. Each frame contains 1200 sampling points, with each sampling point having a value of amplitude, whose square value represents the energy, time, and corresponding spatial location. We can then use the stationary state wave function (37) for analysis.
Ψ n ( x , t ) = 2 a s i n ( n π a x ) e i ( n 2 π 2 / 2 m a 2 ) t
In this experiment, x represents the sampling point position in the spatial domain, while t represents time. The length of the total 0.1 s signal, denoted as a, is 34 m. We can calculate n, where E n is the square value of amplitude. The constant m is assumed to be 1 in order to reduce computational complexity. Then, we can calculate the total probability of this frame and continue with the calculation of the MFCC. A triangle window is applied afterwards. We then perform FFT and pass the resulting signal through a Mel-scaled filter bank, which is a 13-order filter bank, to obtain 13 coefficients from 1 frame. After applying Discrete Cosine Transform, we can obtain the original MFCC, which we then multiply by the probability. The probability of each frame from one standard English speaker’s pronunciation is shown in Figure 13.
The new feature is a 8 × 13 matrix that represent the message of one production. Reshape the matrix to a vector of 104 elements.

6.4. Classification Results

In the classification part, the first algorithm used is KNN. We randomly select an audio feature for a correct and an incorrect pronunciation as the reference. Using Equation (37), we calculate the Euclidean distance for both reference features in the feature space, called μ 1 and μ 2 . Then, we use all other data samples as the testing sample and calculate the distances ( d 1 and d 2 ). To optimize the classification performance, we modify the original KNN algorithm by using the classification logic for a given threshold V i :
H = 1 , if d 1 d 2 V i 0 , if d 1 d 2 > V i
the second classification algorithm is the Gaussian model. We select four native English speakers’ feature samples to calculate the expectation ( μ ) by using Equation (40) and then Equation (41) to get the covariance matrix Σ and then its diagonal matrix D . To scale the output value we only calculate the p = e ( 1 2 ( X i μ ) T Σ 1 ( X i μ ) ) part of Equation (43). The classification logic for a given threshold V i is shown as Equation (47).
H = 1 , if p > V i 0 , if p V i
The last classifier used was an ANN, where features obtained from the previous subsection were fed into the input layer of the Artificial Neural Network. We then constructed one hidden layer with six cells and one bias cell and one output layer with two output nodes.
The comparison of the performance among different features and methods can be found in the Table 2.
We obtained the maximum recognition rate for KNN and GMM classifiers by using a specific threshold value. For the ANN, we evaluated the recognition rate for all three types of features. Our results showed that the proposed speech wave function probability features achieved the highest classification rate when using the KNN classifier. For the GMM classifier, the pitch and partials amplitude features outperformed the other two features. Finally, for the ANN model, the pitch and partials amplitude features provided the highest classification rate.

7. Conclusions

In this study, we aimed to evaluate the feasibility of using a knowledge-based intelligence program to differentiate pronunciation, specifically focusing on one of the most challenging aspects of pronunciation remediation—the rhotic /r/. Our approach involved using signal analysis and classification methods. We detected speech signals containing /r/ by implementing short-time processing, extracting relevant features, and utilizing classifiers for pronunciation diagnosis. Our experiment demonstrated that the proposed knowledge-based intelligence program is a promising solution to this problem. Without the loss of generality, this approach can be extended to other rhotics.
While algorithms commonly used for speech recognition have been extensively researched, they may not be suitable for differentiating correct and incorrect rhotic pronunciation based on detailed acoustic features. Therefore, we proposed two innovative features to address this issue.
The first proposed feature focused on dynamic changes in the fundamental frequency and its harmonics, called a pitch and particle gradient feature. To capture subtle differences among correct and various incorrect pronunciations, we borrowed the concept of Schrodinger’s equation from quantum physics and developed a wave function-based probability feature from speech signals. We collected data from 31 human participants and found that these two features performed well in our experiments.
The pitch and particle gradient feature outperformed the commonly used Mel-frequency cepstral coefficients (MFCCs) feature when the Gaussian Mixture Model (GMM) and Artificial Neural Network (ANN) classifiers were employed. In addition, the wave function-based amplitude probability feature achieved the best performance when using the ANN classifier.
In future work, we plan to explore the potential of our proposed features and classifiers in other aspects of speech recognition and pronunciation assessment. Additionally, we intend to investigate the generalization of our model to larger datasets and diverse populations.
Finally, we aim to investigate the use of additional signal processing techniques and advanced machine learning models such as deep learning to further improve the performance of our proposed features and classifiers.

Author Contributions

Conceptualization, L.L. and W.L.; Methodology, L.L., W.L. and S.M.; Software, L.L., W.L. and M.Z.; Validation, L.L., W.L. and S.M.; Formal analysis, L.L. and W.L.; Investigation, L.L., W.L. and S.M.; Resources, L.L. and W.L.; Data curation, L.L. and W.L.; Writing—original draft, L.L., W.L., S.M. and M.Z.; Writing—review and editing, L.L. and W.L.; Supervision, L.L. and W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Thompson, I. Foreign accents revisited: The English pronunciation of Russian immigrants. Lang. Learn. 1991, 41, 177–204. [Google Scholar] [CrossRef]
  2. Derwing, T.M.; Thomson, R.I.; Munro, M.J. English pronunciation and fluency development in Mandarin and Slavic speakers. System 2006, 34, 183–193. [Google Scholar] [CrossRef]
  3. Murphy, J.M. Intelligible, comprehensible, non-native models in ESL/EFL pronunciation teaching. System 2014, 42, 258–269. [Google Scholar] [CrossRef]
  4. Pronounce r Sound Pronuncian: American English Pronunciation. Pronuncian: American English Pronunciation. 2020. Available online: https://pronuncian.com/pronounce-r-sound (accessed on 16 March 2020).
  5. Wren, Y.; Pagnamenta, E.; Peters, T.J.; Emond, A.; Northstone, K.; Miller, L.L.; Roulstone, S. Educational Outcomes Associated with Persistent Speech Disorder. Int. J. Lang. Commun. Disord. 2021, 56, 299–312. [Google Scholar] [CrossRef] [PubMed]
  6. Hitchcock, E.R.; Harel, D.; Byun, T.M. Social, emotional, and academic impact of residual speech errors in school-aged children: A survey study. Semin. Speech Lang. 2015, 36, 283–294. [Google Scholar] [CrossRef] [PubMed]
  7. Silverman, F.H.; Paulus, P.G. Peer reactions to teen- agers who substitute /w/ for /r/. Lang. Speech Hear. Serv. Sch. 1989, 20, 219–221. [Google Scholar] [CrossRef]
  8. May-Mederake, B. Early intervention and assessment of speech and language development in young children with cochlear implants. Int. J. Pediatr. Otorhinolaryngol. 2012, 76, 939–946. [Google Scholar] [CrossRef] [PubMed]
  9. Shriberg, L.D.; Tomblin, J.B.; McSweeney, J.L. Prevalence of Speech Delay in 6-Year-Old Children and Comorbidity with Language Impairment. J. Speech Lang. Hear. Res. 1999, 42, 1461–1481. [Google Scholar] [CrossRef] [PubMed]
  10. Bleile, K.M. Speech Evaluation. In Speech Sound Disorders: For Class and Clinic; Plural Publishing: San Diego, CA, USA, 2020; pp. 149–164. [Google Scholar]
  11. Owens, F.J. Signal Processing of Speech; Macmillan Press Ltd.: London, UK, 1993. [Google Scholar]
  12. Kondoz, A.M. Digital Speech; John Wiley and Sons Ltd.: West Sussex, UK, 2004. [Google Scholar]
  13. Deng, L. Speech Processing: A Dynamic and Optimization-Oriented Approach; O’Shaughnessy, D., Ed.; Marcel Dekker: New York, NY, USA, 2003; pp. 41–48. [Google Scholar]
  14. Al-Sarayreh, K.T.; Al-Qutaish, R.E.; Al-Kasasbeh, B.M. Using the Sound Recognition Techniques to Reduce the Electricity Consumptions in Highways. J. Am. Sci. 2009, 5, 1–12. [Google Scholar]
  15. Yuan, Y.J.; Zhao, P.H.; Zhou, Q. Research of speaker recognition based on combination of LPCC and MFCC. In Proceedings of the 2010 IEEE International Conference on Intelligent Computing and Intelligent Systems, Xiamen, China, 29–31 October 2010; Volume 3, pp. 765–767. [Google Scholar]
  16. Rabiner, L.; Juang, B.-H. Fundamentals of Speech Recognition; Prentice-Hall: Englewood Cliffs, NJ, USA, 1993. [Google Scholar]
  17. Young, S.; Odell, J.; Ollason, D.; Valtchev, V.; Woodland, P. The HTK Book; Cambridge University: Cambridge, UK, 1997; Version 2.1, Chapter 5.4. [Google Scholar]
  18. Weisstein, E.W. Leakage. From MathWorld-A Wolfram Web Resource. Available online: http://mathworld.wolfram.com/Leakage.html (accessed on 11 March 2020).
  19. Mathworks.com. Fast Fourier Transform—MATLAB FFT. 2020. Available online: https://www.mathworks.com/help/matlab/ref/fft.html (accessed on 11 March 2020).
  20. Nakagawa, S.; Wang, L.; Ohtsuka, S. Speaker Identification and Verification by Combining MFCC and Phase Information. IEEE Trans. Audio Speech Lang. Process. 2012, 20, 1085–1095. [Google Scholar] [CrossRef]
  21. Gaurav, G.; Deiv, D.S.; Sharma, G.K.; Bhattacharya, M. Development of Application Specific Continuous Speech Recognition System in Hindi. J. Signal Inf. Process. 2012, 3, 394–401. [Google Scholar] [CrossRef]
  22. Mathwork.com. Discrete Cosine Transform—MATLAB DCT. 2020. Available online: http://www.mathworks.com/help/signal/ref/dct.html?s_tid=srchtitle (accessed on 12 August 2020).
  23. Sumithra, M.G.; Devika, A.K. A study on feature extraction techniques for text independent speaker identification. In Proceedings of the 2012 International Conference on Computer Communication and Informatics, Coimbatore, India, 10–12 January 2012; pp. 1–5. [Google Scholar]
  24. Mathworks.com. Hamming Window—MATLAB Hamming. 2020. Available online: https://www.mathworks.com/help/matlab/ref/fft.html (accessed on 11 March 2020).
  25. Griffiths, D. Introduction to Quantum Mechanics; Pearson: London, UK, 2014. [Google Scholar]
  26. Du, S.; Li, J. Parallel Processing of Improved KNN Text Classification Algorithm Based on Hadoop. In Proceedings of the 2019 7th International Conference on Information, Communication and Networks (ICICN), Macao, China, 24–26 April 2019; pp. 167–170. [Google Scholar]
  27. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006; Volume 4. [Google Scholar]
  28. Ham, F.M.; Kostanic, I. Principles of Neurocomputing for Science and Engineering; McGraw-Hill, Inc.: New York, NY, USA, 2001. [Google Scholar]
  29. Kudoh, E.; Karino, K. Location Estimation Applying Machine Learning Using Multiple Items of Sensed Information in Indoor Environments. In Proceedings of the 2019 Eleventh International Conference on Ubiquitous and Future Networks (ICUFN), Zagreb, Croatia, 2–5 July 2019; pp. 366–369. [Google Scholar] [CrossRef]
Figure 1. MFCC diagram.
Figure 1. MFCC diagram.
Electronics 12 02055 g001
Figure 2. Frequency to Mel-scale frequency curve.
Figure 2. Frequency to Mel-scale frequency curve.
Electronics 12 02055 g002
Figure 3. Audio in time domain.
Figure 3. Audio in time domain.
Electronics 12 02055 g003
Figure 4. Audio spatial distribution.
Figure 4. Audio spatial distribution.
Electronics 12 02055 g004
Figure 5. Standard speaker “/r/”.
Figure 5. Standard speaker “/r/”.
Electronics 12 02055 g005
Figure 6. Pre-emphasized standard speaker“/r/”.
Figure 6. Pre-emphasized standard speaker“/r/”.
Electronics 12 02055 g006
Figure 7. Frame wave.
Figure 7. Frame wave.
Electronics 12 02055 g007
Figure 8. Hamming window.
Figure 8. Hamming window.
Electronics 12 02055 g008
Figure 9. Frame wave in hamming window.
Figure 9. Frame wave in hamming window.
Electronics 12 02055 g009
Figure 10. Fast Fourier transform.
Figure 10. Fast Fourier transform.
Electronics 12 02055 g010
Figure 11. Amplitude changing trend in 225 Hz.
Figure 11. Amplitude changing trend in 225 Hz.
Electronics 12 02055 g011
Figure 12. Derivation in 225 Hz.
Figure 12. Derivation in 225 Hz.
Electronics 12 02055 g012
Figure 13. Probability of each frame.
Figure 13. Probability of each frame.
Electronics 12 02055 g013
Table 1. Sample of word list.
Table 1. Sample of word list.
r and Adjacent VowelWords
/ru/rude
/ru/room
/ru/ruse
/ru/root
/ri/reach
/ri/reef
/ri/reason
/ri/read
Table 2. Classification results for different features and methods.
Table 2. Classification results for different features and methods.
KNN
CR (%)
KNN
Threshold
GMM
CR (%)
GMM
Threshold
ANN
CR (%)
MFCC82.720.905381.770.448487.217
PPAG79.150.615682.380.214896.97
SWPA84.721.119680.770.456293.911
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, L.; Li, W.; Morris, S.; Zhuang, M. Knowledge-Based Features for Speech Analysis and Classification: Pronunciation Diagnoses. Electronics 2023, 12, 2055. https://doi.org/10.3390/electronics12092055

AMA Style

Liu L, Li W, Morris S, Zhuang M. Knowledge-Based Features for Speech Analysis and Classification: Pronunciation Diagnoses. Electronics. 2023; 12(9):2055. https://doi.org/10.3390/electronics12092055

Chicago/Turabian Style

Liu, Lichuan, Wei Li, Sherrill Morris, and Mutian Zhuang. 2023. "Knowledge-Based Features for Speech Analysis and Classification: Pronunciation Diagnoses" Electronics 12, no. 9: 2055. https://doi.org/10.3390/electronics12092055

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop