Next Article in Journal
Multicultural Identity Integration versus Compartmentalization as Predictors of Subjective Well-Being for Third Culture Kids: The Mediational Role of Self-Concept Consistency and Self-Efficacy
Previous Article in Journal
Factors Influencing Gaseous Emissions in Constructed Wetlands: A Meta-Analysis and Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Empirical Mode Decomposition and Hilbert Spectrum for Abnormality Detection in Normal and Abnormal Walking Transitions

1
School of Computing, Telkom University, Bandung 40257, Indonesia
2
School of Electrical Engineering, Telkom University, Bandung 40257, Indonesia
3
School of Applied Science, Telkom University, Bandung 40257, Indonesia
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2023, 20(5), 3879; https://doi.org/10.3390/ijerph20053879
Submission received: 24 December 2022 / Revised: 14 February 2023 / Accepted: 20 February 2023 / Published: 22 February 2023

Abstract

:
Sensor-based human activity recognition (HAR) is a method for observing a person’s activity in an environment. With this method, it is possible to monitor remotely. HAR can analyze a person’s gait, whether normal or abnormal. Some of its applications may use several sensors mounted on the body, but this method tends to be complex and inconvenient. One alternative to wearable sensors is using video. One of the most commonly used HAR platforms is PoseNET. PoseNET is a sophisticated platform that can detect the skeleton and joints of the body, which are then known as joints. However, a method is still needed to process the raw data from PoseNET to detect subject activity. Therefore, this research proposes a way to detect abnormalities in gait using empirical mode decomposition and the Hilbert spectrum and transforming keys-joints, and skeletons from vision-based pose detection into the angular displacement of walking gait patterns (signals). Joint change information is extracted using the Hilbert Huang Transform to study how the subject behaves in the turning position. Furthermore, it is determined whether the transition goes from normal to abnormal subjects by calculating the energy in the time-frequency domain signal. The test results show that during the transition period, the energy of the gait signal tends to be higher than during the walking period.

1. Introduction

Human Activity Recognition (HAR) aims to recognize a series of activities carried out by someone in an environment or area, either closed or open. HAR seeks to understand people’s daily activities by examining insights collected from individuals and their surroundings [1]. Currently, techniques are collected from various IoT sensors embedded in smartphones, wearables, and settings [2]. Human activity recognition (HAR) techniques play an essential role in monitoring daily activities, especially the activities of the elderly, activities investigation, health care, and sports [3]. Another study used smartphones with various embedded sensors for activity recognition [4]. Motion sensors such as accelerometers and gyroscopes are widely used as inertial sensors to identify multiple human physical conditions. In recent research, much work has been conducted regarding introducing human activity [1].
Human activity is the sum of human actions, where action can be defined as the essential and most minor elements of something achieved by humans. Human activity recognition is not a simple process because someone can perform an action periodically. Two activities may have almost similar properties in terms of the similarity of the signal captured from the camera. Therefore, how an activity is carried out may differ from person to person [1]. In their paper, Dang et al. comprehensively discuss sensor-based and sensor-based HAR computer vision [1]. Computer vision-based HAR relies on visual sensing technology, such as cameras or closed-circuit television (CCTV), to record human activity [5]. It does not require sensors, however, depending on the quality of the image obtained. Unlike computer vision, sensor-based HAR will mount a number of sensors on parts of the body so that data that represents human activity can be obtained. Data collected is a series of time series data and parameters. The effect will be analyzed using a statistical and probabilistic approach. Most of the introduction of human activities still do not consider the transition of activities because the duration is shorter than the activity or basic movement in general [6]. Activity transition is a limited event determined by the start time of the lock from one activity to another. On the other hand, in many practical scenarios, such as fitness monitoring systems, determining activity transitions based on kinematic patterns is very important because it is carried out quickly [7]. In terms of human activity recognition and transient activity perception systems, the classification will change slightly, and the absence of a defined activity may result in the wrong type. The previous activity transition approach has been carried out by Ortiz et. al. [6], but using the accelerometer sensor instead of computer vision. The weakness of HAR using an accelerometer sensor is that the sensor is attached to the patient’s body [1]. This form of the sensor is usually embedded in a smartphone [7]. One solution to the barriers to using wearable sensors is video-based HAR [8]. One of the HAR platforms is PoseNET, which can detect the skeleton of the body and body joints known as joints [9]. Creating PoseNET provides a platform for detecting frames and joints, but it is still necessary to process the data to see activity subjects such as gait and so on [10].
This paper proposes empirical mode decomposition with Hilbert spectrum techniques for detecting anomalies in gait pattern. The methods involve transforming vision-based pose estimation of key joints into angular displacement to obtain walking gait patterns (signals). Using the Hilbert Huang Transform to extract walking gait signals can identify information about the activity’s transition, in this case the turning position of the subsequent step. We observe that the energy from the time-frequency domain of gait signals tends to be greater during the transition between two walking activities than during the single walking activity. Therefore, turning information can be used to estimate whether a person has a normal gait pattern or one with an abnormal gait pattern. The proposed method may be used to monitor elderly individuals or people with gait irregularities or diseases such as Alzheimer’s or Parkinson’s.
The remainder of this paper is presented in four sections. Section 2 contains explanations of previous studies closely related to the proposed study. The materials and methods used in this study are presented in Section 3. Section 4 discusses the study’s results, including the energy features for each case and the classification results. A discussion of the results is also presented in this section. The last section contains conclusions, limitations, implications, and challenges for future studies.

2. Related Works

Sensor-based human activity recognition is the earliest method used by people. The advantages of HAR-based sensors, especially wearable sensors, are [11]:
  • Wearable sensors can overcome static environmental conditions due to static camera placement.
  • With many sensors used, the accuracy obtained will be higher.
  • Each sensor is intended to obtain specific signals from certain people.
  • Wearable sensors better protect user privacy compared with cameras that can record the user’s face and body.
  • There needs to be a supervisor to observe the video recording.
  • Video signal processing requires more computational challenges.
The weakness of sensor-based HAR is the difficulty in sensor placement, which can cause discomfort to the patient; the number of sensors is not certain depending on the needs, power supply, data transmission, and signal processing [2]. One of the most widely used sensors is the accelerometer and gyroscope [12]. These two sensors are often used for falling detection [13] or detecting respiration activity [14]. Apart from movement, wearable sensors that are often used for HAR, such as body temperature sensors, the Global Positioning System (GPS), heart rate monitors, and others, are intended to monitor the user’s health condition. The data acquired by the sensor is then analysed using signal processing methods and classified.
The study by Nurwulan and Jiang uses accelerometer data for recognition and then analyzes the effect of window selection on HAR [15]. Meanwhile, other researchers explored the impact of smartphone position transitions on acquiring data from sensors [16]. Deep learning is becoming the classification method of choice in today’s HAR. Bhattacharya et al. developed the Ensemble Deep Learning Model for HAR classification based on smartphone sensors [12]. However, some researchers still use traditional machine learning, such as KNN, which has been improved [17].
In addition to the weaknesses previously described, video-based HAR has several advantages. It does not interfere with movement, only one type of signal is taken, and it does not require multiple installations [18]. Generally, three types of vision sensors are often used: deep sensors, visible spectrum sensors, and infrared sensors [19]. The deep sensor commonly used is the Microsoft Kinect, which can extract the skeleton from images [20]. RGB cameras usually require more intensive image processing methods such as PoseNet [9,21]. Meanwhile, for infrared, you can use infrared break beam sensors, ultrasonic sensors, and passive infrared sensors [19]. PoseNet plays an important role in developing vision-based HAR [9]. PoseNet uses deep learning to generate 3D spatial coordinates of joint human skeleton poses. Examples of vision-based HAR applications include the detection of falls in the elderly [22] or the detection of the subject’s pose [23].

3. Materials and Methods

The proposed system block diagram is shown in Figure 1. In this study, the Openpose pose framework was used to extract the knee joint from the subject. Carnegie Mellon University (CMU) developed OpenPose, a supervised convolutional neural network based on Caffe, for real-time multi-person 2D pose estimation [24]. It can estimate the posture of human body movements, facial expressions, and finger movements. It has an excellent recognition effect and a fast recognition speed, making it suitable for single- and multiple-user settings. After obtaining the knee joint, an analysis was carried out using empirical mode decomposition (EMD) and the Hilbert Huang Transform (HHT) to detect gait patterns and turning events. Details of each stage are described in the sections below.
Figure 1 shows the three main stages in this study, including preprocessing, signal processing, and training and testing. At the preprocessing stage, the video dataset is taken for pose estimation, in this case using PoseNet. The output from the pose estimation engine is the position of the key joints and the artificial skeleton in each video frame, which is overlaid with the original video. However, for the next process, only the coordinates of the key joints and the artificial skeleton are used. In the second stage, signal processing is carried out by extracting joints and converting them into joint coordinates and joint kinematic angles (flexion and extension). The result of this extraction is in the form of gait activity signals. Furthermore, gait activity signals are extracted using the Hilbert Huang Transform, with the initial stage being EMD processing to obtain the intrinsic mode function (IMF) signal components [8]. In the case of this study, the 6th IMF signal is used as a feature for classification. The entire IMF is also processed by the Hilbert Transform algorithm to obtain the energy spectrum from the gait activity signal. The energy spectrum value, along with the subsequent IMF, is used as a feature for the K-Nearest Neighbor (KNN) classifier. The results of the classification of the training dataset and testing dataset are grouped into Normal Walking activity, Parkinson’s disease activity, and other disordered walking activity classes.

3.1. Dataset

The data used in this study are video recordings where the subject performs walking and turning activities. Videos are taken from several YouTube channels, one of which is https://www.youtube.com/c/MissionGait, accessed on 1 August 2022, with different subjects and varied backgrounds. The video used in this study is in mp4 format with a frame rate of 30 fps. Figure 2 shows an example of the video analyzed in this study. A total of 14 videos were processed in this study, including normal and abnormal walking.
In Figure 2 and Figure 3, you can see an example of a video taken from the Mission Gate dataset. In the picture, the left is the original video, and the right is the overlay video between the original video and the artificial skeleton and joints resulting from pose estimation at the preprocessing stage (see the method in Figure 1). From a total of 15 datasets used, the pose estimation engine succeeded in overlaying the original video with artificial skeletons and joints. Only the joints related to the lower limb are extracted for gait activities, namely the knee and hip joints, and the ankle joint is extracted so that an artificial skeleton can be constructed that connects the joints so that the angles between the joints can be calculated.
The video dataset used for this experiment can be seen in Table 1. Mission Gait provides a series of videos demonstrating real-life individuals with gait disorders. Thus, this video dataset can serve as a dataset for research, therapist practice, and telemedicine applications for early diagnosis. In this study, 15 datasets represented normal walking activities and some gait disorder activities, as seen in Table 1. The video dataset used for this experiment came from Mission Gait, with screenshots as shown in Table 1. Mission Gait provides a series of videos demonstrating real individuals with gait disorders. Thus, this video dataset can serve as a research dataset, therapist practice, and telemedicine for initial diagnosis. In this study, 13 datasets represented normal walking activities and some gait disorder activities, as seen in Table 1.
Some disorder terminology is taken from WebMD [12]. The several decoded categories of the video dataset, according to those on the mission gait site are as in Table 2.
In general, how the gait moves will provide early clues as to the underlying cause of the gait problem. This can help the doctor or therapist diagnose the problem and plan therapy. Each type of gait disorder has variations, and no two people will have the same symptoms. Thus, the dataset will produce gait patterns with the general characteristics extracted in this study using signal processing methods.

3.2. Empirical Mode Decomposition (EMD)

Huang et al. introduced Empirical Mode Decomposition (EMD) in 1998 as a new and effective tool for analyzing non-linear and non-stationary signals [25,26]. A complicated and multiscale signal can be adaptively decomposed into the sum of a finite number of zero-mean oscillating components known as Intrinsic Mode Functions using this method (IMF).
EMD’s specific steps are as follows:
Step 1: Determine the maximum and minimum signal × values (t).
Step 2: To obtain the highest possible value emax(t), use the cubic spline interpolation function to fit the upper envelope; use the cubic spline interpolation function to fit the lower envelope emin(t) for the minimum value.
Step 3: Calculate m(t):
m ( t ) = e max ( t ) e min ( t ) 2
Step 4: Calculate modal function c(t):
c ( t ) = x ( t ) m ( t )
Step 5: If c(t) satisfies IMF condition, then c(t) is an IMF component, and the original signal becomes xn+1(t), then:
x n + 1 ( t ) = x ( t ) c ( t )
If c(t) does not satisfy IMF condition, go back to step 3.
Step 6: The residual component r(t) is reserved and the decomposition is complete when the result signal has less than two extremum locations. The original signal is divided into n IMFs, with r(t) being the residual component.
x ( t ) = i = 1 N c i ( t ) + r ( t )

3.3. Hilbert Spectrum

Hilbert spectral analysis is a signal analysis technique that uses the Hilbert transform to calculate the instantaneous frequency of signals [27].
ω = d θ d t
After applying the Hilbert transform to each signal, we can express the data as in Equation (6).
X ( t ) = j = 1 n a j ( t ) e x p ( i ω j ( t ) d t )
As a function of time, this equation gives the amplitude and frequency of each component. It also allows us to represent the amplitude and instantaneous frequency as time functions in a three-dimensional plot, with the amplitude contoured on the frequency-time plane. The Hilbert amplitude spectrum, or simply Hilbert spectrum, is the frequency-time distribution of the amplitude. The Hilbert spectral analysis method is an essential component of the Hilbert-Huang transform [28].

4. Results and Discussion

4.1. Detection Confidence

Body pose estimation, in this case using machine learning-based pose estimation, works based on a specific confidence value, whether a landmark (joint) is detected or no activity is taking place. This argument will set the threshold value of the confidence level and range from [0.0, 1.0], i.e., the minimum confidence level is 0, and the maximum confidence level is one. By default, the value is 0.5. Thus, the pose estimation engine will determine whether the joint is detected based on the given confidence level. In Figure 4, it can be seen the evolution of the confidence value (in percent) during the pose estimation process.
The video used in this example is from the Case Study 10 Dataset, with the direction of activity running from right to left of the video. In the 4th second, the confidence value goes to the lowest point due to occlusion, or the object is covered so that the pose estimator detects joints with a low level of visibility (low detection confidence). Overall, the detection confidence of all dataset files can be mapped using the probability density function for both the left and right knee, as seen in Figure 4. The lowest mean value of detection confidence is around 90%, and the highest mean is about 98%. The closer the detection confidence value is to 1, the more accurately the detected joints are detected, so when it is used to define poses per video frame, the specified poses are also more accurate. Of course, this will affect the tracking results of an angle formed from various artificial joints and seconds.
Figure 5 displays the probability density function of detection confidence for all datasets used. It can be seen that the right knee produces a higher probability density than the left knee. This will be a consideration in selecting features for the classification of abnormalities.

4.2. Gait Signal Based on Video Dataset

Figure 6 shows the gait signal in the form of angular values taken from the right and left knees of a person walking normally. The graph shows the regular pattern of angular values throughout the observation time. Meanwhile, Table 3 shows the gait signal in several cases, represented by Case Study 10 for normal and Case Studies 6 and 5 for subjects with Parkinson’s and diabetic neuropathy. In case study 6, subjects with Parkinson’s produced a significant difference in turn times (4th to 14th seconds).

4.3. Feature Extraction Using Hilbert Huang Transform

HHT is a combination of two methodologies [29], namely, empirical mode decomposition (EMD) and Hilbert transform (HT). In the first step, the input signal will be decomposed into different components using EMD, which are called intrinsic mode functions (IMFs) [30]. In the second step, the Hilbert spectrum is obtained by conducting HT over IMF. HHT can adaptively decompose IMF and also has better time-frequency resolution when compared with the wavelet transform (WT) and short-time Fourier transform (STFT) [31]. Figure 7 shows IMF-1 through IMF-6 resulting from the gait signal decomposition case study 10. Meanwhile, Figure 8 shows the original gait signal and IMF-5, which represent two different activities. When the subject performs the turning activity, the signal value from IMF-5 drops significantly.
Figure 9 shows the Hilbert energy spectrum of the right and left knee from case studies 10 and 6. The energy is plotted over the time span, representing the two different activities. When the subject performs a turning activity, it produces a smaller energy value than walking activity. However, in subjects with Parkinson’s, low energy occurs with a longer duration, indicating that the subject has difficulty turning as depicted in Figure 10.
Based on the features formed, we created a class definition as display in Table 4.

4.4. Classification Result

Table 5 shows the accuracy of the three gait class classifications using the random forest classifier (RF). The parameters used to measure its performance are the True Positive Rate (TPR) and False Negative Rate (FNR). Mathematically, the TPR and FNR are expressed as in Equations (7) and (8) [32].
T P R = T P F N + T P
F N R = F N F N + T P = 1 T P R
From Table 4, it can be seen that for NW, most of the TPR values > FPR, meaning that more NW are recognized correctly as NW compared with those incorrectly recognized. Meanwhile, the OW tends to have a TPR value < FPR, which means that OW is more likely to be misrecognized. Meanwhile, due to the small amount of data for PD, it can be seen that for data 6, the TPR reaches 100% while the other data is TPR < FNR. In general, the results of the proposed method have yet to yield sufficiently good results; this is due to the results only showing when a transition occurs while running. These results have yet to be further analyzed to determine the difference between the transition signals under certain disease conditions and under normal conditions. From this initial observation, it was found that one of the differences that can be used is the duration of the transition process. In this study, the dataset’s limitations are also one reason the accuracy tends to be low. In addition, this study used only the left and right knee joints as signal sources. Adding the number of joints analyzed is also a potential area for further research. Despite the low accuracy, the proposed method has the advantage of being easy in the data acquisition process because it does not require a sensor attached to the subject’s body [21]. This method has the potential to be applied to in-home rehabilitation for the elderly.

5. Conclusions

This study proposes a method for detecting abnormalities in walking transitions using the EMD and Hilbert spectra. The input data is in the form of a walking video of the subject extracted using PostNet. In this study, only the left knee was analyzed because it was considered to represent the subject’s movement. A person with an abnormal gait has a longer turning duration than normal. In the case of people with Parkinson’s, it produces a turn time about 10 s, while in normal people it takes 2 s. From the experiment, it was found that the location of the walking transition can be seen using the Hilbert spectrum. Hilbert’s energy spectrum shows that when the subject performs the turning activity, it produces a smaller energy value than the walking activity. However, in subjects with Parkinson’s, low energy occurs with a longer duration, indicating that the subject has difficulty turning. Furthermore, the classification of cases of normal walking (NW), other disorder walking (OW), and Parkinson’s Disease (PD) was carried out. The classification results show that NW produces a higher TPR > FPR, which means that more NW are classified correctly. Whereas OW tends to have a TPR < FPR value, which means OW is more likely to be recognized incorrectly. Meanwhile, due to the small amount of PD data, it can be seen that for case study 6, the TPR reached 100% while the other data was TPR < FNR. A future study needs further analysis to determine the difference between transition signals under certain disease conditions and under normal conditions. In this study, the limited dataset is also one of the causes of the low accuracy. Exploration to extract the time of walking transitions is also the focus of the following research. The method proposed in this study has the potential to be applied to home rehabilitation for the elderly.

Author Contributions

Conceptualization, B.E. and A.R.; Methodology, A.R.; Software, B.E.; Validation, S.H. Formal Analysis, B.E. and S.H.; Resources, B.E.; Data Curation, A.R.; Writing—Original Draft Preparation, A.R. and S.H.; Writing—Review & Editing, A.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by the Ministry of Education, Culture, Research and Technology of the Republic of Indonesia grant number 126/SP2H/RT-MONO/LL4/2022).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Minh Dang, L.; Min, K.; Wang, H.; Jalil Piran, M.; Hee Lee, C.; Moon, H. Sensor-Based and Vision-Based Human Activity Recognition: A Comprehensive Survey. Pattern Recognit. 2020, 108, 107561. [Google Scholar] [CrossRef]
  2. Ahad, M.A.R.; das Antar, A.; Ahmed, M. Sensor-Based Human Activity Recognition: Challenges Ahead; Springer: Berlin/Heidelberg, Germany, 2021; pp. 175–189. [Google Scholar]
  3. Song, L.; Yu, G.; Yuan, J.; Liu, Z. Human Pose Estimation and Its Application to Action Recognition: A Survey. J. Vis. Commun. Image Represent 2021, 76, 103055. [Google Scholar] [CrossRef]
  4. Ye, Z.; Li, Y.; Zhao, Q.; Liu, X. A Falling Detection System with Wireless Sensor for the Elderly People Based on Ergonomics. Int. J. Smart Home 2014, 8, 187–196. [Google Scholar] [CrossRef]
  5. Lun, R.Z. Human Activity Tracking and Recognition Using Kinect Sensor. Ph.D. Thesis, Cleveland State University, Cleveland, OH, USA, 2018. [Google Scholar]
  6. Reyes-Ortiz, J.-L.; Oneto, L.; Samà, A.; Parra, X.; Anguita, D. Transition-Aware Human Activity Recognition Using Smartphones. Neurocomputing 2016, 171, 754–767. [Google Scholar] [CrossRef] [Green Version]
  7. Kwon, Y.; Kang, K.; Bae, C. Unsupervised Learning for Human Activity Recognition Using Smartphone Sensors. Expert Syst. Appl. 2014, 41, 6067–6074. [Google Scholar] [CrossRef]
  8. Munea, T.L.; Jembre, Y.Z.; Weldegebriel, H.T.; Chen, L.; Huang, C.; Yang, C. The Progress of Human Pose Estimation: A Survey and Taxonomy of Models Applied in 2D Human Pose Estimation. IEEE Access 2020, 8, 133330–133348. [Google Scholar] [CrossRef]
  9. Kendall, A.; Grimes, M.; Cipolla, R. PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; Volume 2015, pp. 2938–2946. [Google Scholar]
  10. Divya, R.; Peter, J.D. Smart Healthcare System-a Brain-like Computing Approach for Analyzing the Performance of Detectron2 and PoseNet Models for Anomalous Action Detection in Aged People with Movement Impairments. Complex Intell. Syst. 2022, 8, 3021–3040. [Google Scholar] [CrossRef]
  11. Serpush, F.; Menhaj, M.B.; Masoumi, B.; Karasfi, B. Wearable Sensor-Based Human Activity Recognition in the Smart Healthcare System. Comput. Intell. Neurosci. 2022, 2022, 1391906. [Google Scholar] [CrossRef]
  12. Bhattacharya, D.; Sharma, D.; Kim, W.; Ijaz, M.F.; Singh, P.K. Ensem-HAR: An Ensemble Deep Learning Model for Smartphone Sensor-Based Human Activity Recognition for Measurement of Elderly Health Monitoring. Biosensors 2022, 12, 393. [Google Scholar] [CrossRef]
  13. Ren, L.; Peng, Y. Research of Fall Detection and Fall Prevention Technologies: A Systematic Review. IEEE Access 2019, 7, 77702–77722. [Google Scholar] [CrossRef]
  14. Erfianto, B.; Rizal, A. IMU-Based Respiratory Signal Processing Using Cascade Complementary Filter Method. J. Sens. 2022, 2022, 7987159. [Google Scholar] [CrossRef]
  15. Retno, N.; Jiang, B.C. Window Selection Impact in Human Activity Recognition Window Selection Impact in Human Activity Recognition. Int. J. Innov. Technol. Interdiscip. Sci. 2020, 3, 381–394. [Google Scholar] [CrossRef]
  16. Amroun, H.; Ouarti, N.; Temkit, M.; Ammi, M. Impact of the Positions Transition of a Smartphone on Human Activity Recognition. In Proceedings of the 2017 IEEE International Conference on Internet of Things, IEEE Green Computing and Communications, IEEE Cyber, Physical and Social Computing, IEEE Smart Data, iThings-GreenCom-CPSCom-SmartData, Exeter, UK, 21–23 June 2017; Institute of Electrical and Electronics Engineers Inc.: Piscatevi, NJ, USA, 2018; Volume 2018, pp. 937–942. [Google Scholar]
  17. Saha, J.; Chowdhury, C.; Ghosh, D.; Bandyopadhyay, S. A Detailed Human Activity Transition Recognition Framework for Grossly Labeled Data from Smartphone Accelerometer. Multimed. Tools Appl. 2021, 80, 9895–9916. [Google Scholar] [CrossRef]
  18. Zhang, S.; Wei, Z.; Nie, J.; Huang, L.; Wang, S.; Li, Z. A Review on Human Activity Recognition Using Vision-Based Method. J. Healthc. Eng. 2017, 2017, 3090343. [Google Scholar] [CrossRef] [Green Version]
  19. Fu, B.; Damer, N.; Kirchbuchner, F.; Kuijper, A. Sensing Technology for Human Activity Recognition: A Comprehensive Survey. IEEE Access 2020, 8, 83791–83820. [Google Scholar] [CrossRef]
  20. Febriana, N.; Rizal, A.; Susanto, E. Sleep Monitoring System Based on Body Posture Movement Using Microsoft Kinect Sensor. In Proceedings of the 3rd Biomedical Engineering’s Recent Progress in Biomaterials, Drugs Development, and Medical Devices, Jakarta, Indonesia, 6–8 August 2018; p. 020012. [Google Scholar]
  21. Chua, J.; Ong, L.-Y.; Leow, M.-C. Telehealth Using PoseNet-Based System for In-Home Rehabilitation. Future Internet 2021, 13, 173. [Google Scholar] [CrossRef]
  22. Siddiq, M.I.; Rizal, A.; Erfianto, B.; Hadiyoso, S. Falling Estimation Based on PoseNet Using Camera with Difference Absolute Standard Deviation Value and Average Amplitude Change on Key-Joint. Appl. Mech. Mater. 2023, 913, 133–142. [Google Scholar]
  23. Guo, F.; He, Y.; Guan, L. RGB-D Camera Pose Estimation Using Deep Neural Network. In Proceedings of the 2017 IEEE Global Conference on Signal and Information Processing (GlobalSIP), Montreal, QC, Canada, 14–16 November 2017; pp. 408–412. [Google Scholar]
  24. Cao, Z.; Simon, T.; Wei, S.-E.; Sheikh, Y. Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  25. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.-C.; Tung, C.C.; Liu, H.H. The Empirical Mode Decomposition and the Hilbert Spectrum for Nonlinear and Non-Stationary Time Series Analysis. Proc. R. Soc. Lond. A 1998, 454, 903–995. [Google Scholar] [CrossRef]
  26. Shuvo, S.B.; Ali, S.N.; Swapnil, S.I.; Hasan, T.; Bhuiyan, M.I.H. A Lightweight CNN Model for Detecting Respiratory Diseases from Lung Auscultation Sounds Using EMD-CWT-Based Hybrid Scalogram. IEEE J. Biomed. Health Inform. 2021, 25, 2595–2603. [Google Scholar] [CrossRef]
  27. Villalobos-Castaldi, F.M.; Ruiz-Pinales, J.; Valverde, N.C.K.; Flores, M. Time-Frequency Analysis of Spontaneous Pupillary Oscillation Signals Using the Hilbert-Huang Transform. Biomed. Signal Process. Control 2016, 30, 106–116. [Google Scholar] [CrossRef]
  28. Caseiro, P.; Fonseca-Pinto, R.; Andrade, A. Screening of Obstructive Sleep Apnea Using Hilbert-Huang Decomposition of Oronasal Airway Pressure Recordings. Med. Eng. Phys. 2010, 32, 561–568. [Google Scholar] [CrossRef]
  29. Lozano, M.; Fiz, J.A.; Jané, R. Performance Evaluation of the Hilbert-Huang Transform for Respiratory Sound Analysis and Its Application to Continuous Adventitious Sound Characterization. Signal Process. 2016, 120, 99–116. [Google Scholar] [CrossRef] [Green Version]
  30. Rizal, A.; Hidayat, R.; Nugroho, H.A. Lung Sound Classification Using Empirical Mode Decomposition and the Hjorth Descriptor. Am. J. Appl. Sci. 2017, 14, 166–173. [Google Scholar] [CrossRef] [Green Version]
  31. Hernández, D.E.; Trujillo, L.; Z-Flores, E.; Villanueva, O.M.; Romo-Fewell, O. Detecting Epilepsy in EEG Signals Using Time, Frequency and Time-Frequency Domain Features. In Studies in Systems, Decision and Control; Springer: Cham, Switzerland, 2018. [Google Scholar]
  32. Chen, X.; Cheng, Z.; Wang, S.; Lu, G.; Xv, G.; Liu, Q.; Zhu, X. Atrial Fibrillation Detection Based on Multi-Feature Extraction and Convolutional Neural Network for Processing ECG Signals. Comput. Methods Programs Biomed. 2021, 202, 106009. [Google Scholar] [CrossRef]
Figure 1. Diagram block of proposed method.
Figure 1. Diagram block of proposed method.
Ijerph 20 03879 g001
Figure 2. Walking activity Case Study 5. (Left): Original Video. (Right): Overlaying Pose Estimation results into original video.
Figure 2. Walking activity Case Study 5. (Left): Original Video. (Right): Overlaying Pose Estimation results into original video.
Ijerph 20 03879 g002
Figure 3. Video showing someone walking.
Figure 3. Video showing someone walking.
Ijerph 20 03879 g003
Figure 4. Detection confidence during walking and turning activities.
Figure 4. Detection confidence during walking and turning activities.
Ijerph 20 03879 g004
Figure 5. Probability density function of detection confidence for all datasets used.
Figure 5. Probability density function of detection confidence for all datasets used.
Ijerph 20 03879 g005
Figure 6. Example of Gait Signal obtained from pose estimation.
Figure 6. Example of Gait Signal obtained from pose estimation.
Ijerph 20 03879 g006
Figure 7. Example of IMF Signals decomposed from Gait Signals using IMF 6.
Figure 7. Example of IMF Signals decomposed from Gait Signals using IMF 6.
Ijerph 20 03879 g007
Figure 8. Example of IMF 5 Signal with Original Gait Signals.
Figure 8. Example of IMF 5 Signal with Original Gait Signals.
Ijerph 20 03879 g008
Figure 9. Example of Hilbert spectrum from normal gait signal from case study 10. (a) HT Energy Spectrum. (b) Energy extracted from HHT Energy Spectrum.
Figure 9. Example of Hilbert spectrum from normal gait signal from case study 10. (a) HT Energy Spectrum. (b) Energy extracted from HHT Energy Spectrum.
Ijerph 20 03879 g009
Figure 10. Example of Hilbert spectrum from Parkinson Disease Walking Activity from case study 6. (a) HT Energy Spectrum. (b) Energy extracted from HHT Energy Spectrum.
Figure 10. Example of Hilbert spectrum from Parkinson Disease Walking Activity from case study 6. (a) HT Energy Spectrum. (b) Energy extracted from HHT Energy Spectrum.
Ijerph 20 03879 g010
Table 1. Video Dataset.
Table 1. Video Dataset.
Case Study
1234567
Ijerph 20 03879 i001Ijerph 20 03879 i002Ijerph 20 03879 i003Ijerph 20 03879 i004Ijerph 20 03879 i005Ijerph 20 03879 i006Ijerph 20 03879 i007
NLLDKAPD2DNPD 1N
Case Study
10111213172123
Ijerph 20 03879 i008Ijerph 20 03879 i009Ijerph 20 03879 i010Ijerph 20 03879 i011Ijerph 20 03879 i012Ijerph 20 03879 i013Ijerph 20 03879 i014
NNMSHPCHPD3BI
Table 2. Video Dataset Category.
Table 2. Video Dataset Category.
DatasetVideo Duration (Second)Initial
Direction
Status Gait
Case Study 113Right to LeftN: Normal, normal walking gait taken from author
Case Study 216Left to RightLLD: Leg Length Discrepancy
Case Study 325Right to LeftKA: Knee Amputee is amputation from knee below
Case Study 413Left to RightPD2: Parkinson Disease (Mild)
Case Study 520Left to RightDN: Diabetic Neuropathy is a motor nerve disorder caused by diabetes.
Case Study 617Left to RightPD1: Parkinson Disease (Severe)
Case Study 718Right to LeftNormal (Video from mission gait)
Case Study 1020Right to LeftNormal (Video from mission gait)
Case Study 1118Right to LeftNormal (Video from mission gait)
Case Study 1220Right to LeftMS: Multiple Sclerosis is a movement disorder due to communication disorders between the brain and the motor system of the body
Case Study 1321Left to RightHP: Hemiplegic, motoric disorder as common long-term consequence of stroke
Case Study 1729Left to RightCH: Chronic Hemiparetic
Case Study 2122Left to RightPD3: Parkinson Disease (Moderate)
Case Study 2311Left to RightBI: Brain Injury
Case Study 2414Left to RightDS: Dystonia is movement disorder that causes the muscles to contract involuntarily
Table 3. Gait value in normal and abnormal cases.
Table 3. Gait value in normal and abnormal cases.
Gait Signals
Case Study 10Ijerph 20 03879 i015
Case Study 6Ijerph 20 03879 i016
Case Study 5Ijerph 20 03879 i017
Table 4. Class definition of each gait signals. NW: Normal Walking, OW: Other disorder walking, PD: Parkinson Disease.
Table 4. Class definition of each gait signals. NW: Normal Walking, OW: Other disorder walking, PD: Parkinson Disease.
DatasetClass
Case Study 1NW
Case Study 2OW
Case Study 3OW
Case Study 4PD
Case Study 5OW
Case Study 6PD
Case Study 7NW
Case Study 10NW
Case Study 11NW
Case Study 12OW
Case Study 13OW
Case Study 17OW
Case Study 21PD
Case Study 23OW
Case Study 24OW
Table 5. Classification Results. NW: Normal Walking. OW: Other Disorder Walking. PD: Parkinson Disease Walking. TPR: True Positive Rate. FNR: False Negative Rate.
Table 5. Classification Results. NW: Normal Walking. OW: Other Disorder Walking. PD: Parkinson Disease Walking. TPR: True Positive Rate. FNR: False Negative Rate.
NWOWPD
TPRFNRTPRFNRTPRFNR
Case Study 1 [NW]0.6180.3820000
Case Study 2 [OW]001000
Case Study 3 [OW]000.3110.68800
Case Study 4 [PD]00000.140.86
Case Study 5 [OW]000.460.5400
Case Study 6 [PD]000010
Case Study 7 [NW]0.7590.2410000
Case Study 10 [NW]100000
Case Study 11 [NW]0.2560.7440000
Case Study 12 [OW]000.3280.67200
Case Study 13 [OW]000.4810.51900
Case Study 17 [OW]000.3440.65600
Case Study 21 [PD]00000.2660.733
Case Study 23 [OW]000.4520.54800
Case Study 24 [OW]000.3790.62100
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Erfianto, B.; Rizal, A.; Hadiyoso, S. Empirical Mode Decomposition and Hilbert Spectrum for Abnormality Detection in Normal and Abnormal Walking Transitions. Int. J. Environ. Res. Public Health 2023, 20, 3879. https://doi.org/10.3390/ijerph20053879

AMA Style

Erfianto B, Rizal A, Hadiyoso S. Empirical Mode Decomposition and Hilbert Spectrum for Abnormality Detection in Normal and Abnormal Walking Transitions. International Journal of Environmental Research and Public Health. 2023; 20(5):3879. https://doi.org/10.3390/ijerph20053879

Chicago/Turabian Style

Erfianto, Bayu, Achmad Rizal, and Sugondo Hadiyoso. 2023. "Empirical Mode Decomposition and Hilbert Spectrum for Abnormality Detection in Normal and Abnormal Walking Transitions" International Journal of Environmental Research and Public Health 20, no. 5: 3879. https://doi.org/10.3390/ijerph20053879

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop