Next Article in Journal
Investigating Visual Perception Impairments through Serious Games and Eye Tracking to Anticipate Handwriting Difficulties
Previous Article in Journal
Denoising of BOTDR Dynamic Strain Measurement Using Convolutional Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Explainable Spatial-Temporal Graphical Convolutional Network to Score Freezing of Gait in Parkinsonian Patients

1
Department of Biomedical Informatics, School of Medicine, Emory University, Atlanta, GA 30322, USA
2
Department of Epidemiology, Rollins School of Public Health, Emory University, Atlanta, GA 30322, USA
3
Jean and Paul Amos Parkinson’s Disease and Movement Disorders Program, Department of Neurology, School of Medicine, Emory University, Atlanta, GA 30322, USA
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(4), 1766; https://doi.org/10.3390/s23041766
Submission received: 13 January 2023 / Revised: 27 January 2023 / Accepted: 1 February 2023 / Published: 4 February 2023
(This article belongs to the Topic Human Movement Analysis)

Abstract

:
Freezing of gait (FOG) is a poorly understood heterogeneous gait disorder seen in patients with parkinsonism which contributes to significant morbidity and social isolation. FOG is currently measured with scales that are typically performed by movement disorders specialists (i.e., MDS-UPDRS), or through patient completed questionnaires (N-FOG-Q) both of which are inadequate in addressing the heterogeneous nature of the disorder and are unsuitable for use in clinical trials The purpose of this study was to devise a method to measure FOG objectively, hence improving our ability to identify it and accurately evaluate new therapies. A major innovation of our study is that it is the first study of its kind that uses the largest sample size (>30 h, N = 57) in order to apply explainable, multi-task deep learning models for quantifying FOG over the course of the medication cycle and at varying levels of parkinsonism severity. We trained interpretable deep learning models with multi-task learning to simultaneously score FOG (cross-validated F1 score 97.6%), identify medication state (OFF vs. ON levodopa; cross-validated F1 score 96.8%), and measure total PD severity (MDS-UPDRS-III score prediction error ≤ 2.7 points) using kinematic data of a well-characterized sample of N = 57 patients during levodopa challenge tests. The proposed model was able to explain how kinematic movements are associated with each FOG severity level that were highly consistent with the features, in which movement disorders specialists are trained to identify as characteristics of freezing. Overall, we demonstrate that deep learning models’ capability to capture complex movement patterns in kinematic data can automatically and objectively score FOG with high accuracy. These models have the potential to discover novel kinematic biomarkers for FOG that can be used for hypothesis generation and potentially as clinical trial outcome measures.

1. Introduction

Parkinson’s Disease (PD) is a slowly progressive neurodegenerative disorder that predominantly affects dopamine-producing neurons in the brain, and individuals with PD exceed more than 10 million people worldwide [1,2]. One of the most disabling features of PD and one of the greatest unmet needs is freezing of gait (FOG), which unfortunately is not always clearly treatable medically and/or surgically. FOG is described as brief arrests of stepping when initiating gait, turning, or walking straight ahead [3,4,5]. When a person freezes, they feel like their feet are “glued” to the floor. FOG is a frequent cause of falls and serious injuries, and represents a significant public health burden (∼86% of patients fall each year) [6,7,8].
One critical factor limiting our ability to treat FOG is that clinicians measure it relatively coarsely, primarily with expert rater observations as part of the Movement Disorder Society-Unified Parkinson’s Disease Rating Scale Part III (MDS-UPDRS-III) scale [9]. This scale requires specially trained raters who have typically completed movement disorders training. In addition, despite being resource-intensive, FOG is only quantified with a single item on an ordinal scale from 0 to 4, which may be too insensitive to detect small beneficial effects. The most established self-reporting scale used in research settings, the N-FOG-Q is acknowledged to be insufficiently sensitive for clinical trial use [10]. Previous work have shown that FOG may be associated with non-dopaminergic system changes [3,11,12], which suggests the potential for new treatments beyond dopaminergic medications like carbidopa-levodopa [4]. However, developing a novel drug that is effective in treating FOG requires accurately quantifying FOG to increase the precision for clinical trials.
Multiple studies have proposed methods to phenotype and rate FOG from kinematic data during walking. For example, those include capturing impaired gait patterns from lower back motion [13], describing gait complexity as a topological nonlinear dynamics system [14], or exploring combinations of sensor locations (shank, thigh, waist), axes (orthogonal, mediolateral, and antero-posterior), window lengths, and features (statistical, frequency, and time-series) to find the best setting that captures FOG characteristics.
Much of the prior work is characterized by a few substantial limitations [15], including (1) a small number of body-worn sensor locations, (2) small sample sizes with mostly early-stage PD patients lacking of severe FOG cases, (3) little consensus on proposed methods across studies, and (4) a relative paucity of studies conducted in the ON- and OFF- medication states, which is necessary to develop technology that will work over the entire medication cycle.
More importantly, most prior studies rely on hand-crafted features for identifying FOG, which may neglect important latent features within the data. For example, relative power in a “freeze band” of accelerometry or other signals [16,17,18], peak detection or similar methods applied to body segment motion [19,20], cycle-to-cycle variation in gait parameters [21], or a combination of the above were used in a support vector machine or other shallow machine learning models [22]. Due to the variability and complexity of FOG behavior, it is unlikely that manually designed spectral features will capture all the characteristics of FOG phenotypes. The popular “freeze band” analysis cannot capture pure akinetic freezing, which does not present with tremulousness.
Here, we use a deep learning approach to capture complex patterns in kinematic data and automatically score FOG, as well as identifying medication state and measure total MDS-UPDRS-III score during a rigorous levodopa challenge paradigm [3]. The proposed explainable (or interpretable) deep learning model automatically identifies the most relevant body joint or limb part and motion segment that is useful while performing multi-task predictions. According to the proposed model, fine-grained movement differences are associated with varying FOG scores, showing the heterogeneity of FOG, in agreement with movement disorder specialists’ observations. Our findings could be used to create new clinical trials to discover new kinematic biomarkers or to develop fine-grained and objective severity scales for individualized FOG treatment. We analyzed over 30 h of 3D motion capture data of 57 patients with varying PD disease duration and FOG severity, including 5 patients with primary progressive FOG, a distinct condition in which FOG presents without parkinsonian features [5]. This dataset is among the largest samples seen in the FOG literature (in which the average sample size was recently estimated as 18 ± 15 [15]). To our knowledge, this work is the first application of interpretable deep learning to solve such a multi-task problem in PD.

2. Materials Furthermore, Methods

We trained an interpretable deep learning model on whole-body 3D kinematic data taken from behavioral motor tasks in N = 57 patients with and without FOG. Clinical, imaging, and cerebrospinal fluid analysis results from patients in this sample have been reported previously [3,23].

2.1. Behavioral Testing

2.1.1. Study Participants

Although this was an observational study for which registration was not required, it was registered through clinicaltrials.gov (NCT02387281). Participants were recruited from the Emory Movement Disorders Clinic and provided written informed consent according to procedures approved by Emory University IRB. The inclusion criteria included: Age ≥ 18 years; PD diagnosis according to United Kingdom Brain Bank criteria [24]; Hoehn & Yahr stage I-IV in the OFF state; ability to sign a consent document and willing to participate in all aspects of the study. Participants with FOG were additionally required to have FOG noted in medical history and confirmed visually by examiner. Exclusion criteria included: vascular parkinsonism and drug-induced parkinsonism as well as the presence of cerebrovascular disease or extensive white matter disease; prior treatment with medications that cause parkinsonism; neurological or orthopedic disorders interfering with gait; dementia or other medical problems precluding completion of the study protocol. Demographic and clinical characteristics of study participants are presented in Table 1.

2.1.2. Levodopa Challenge Paradigm

Each participant was assessed twice using an identical testing protocol: first, in the practically defined “OFF” state > 12 h after the last intake of all antiparkinsonian medications, and second, after a levodopa equivalent dose of ∼150% of the typical morning dose sufficient to elicit a full “ON” state. Additional details of the levodopa testing procedure have been presented previously [3]. In each state, they were assessed with the MDS-UPDRS-III motor exam [9] and with timed-up-and-go (TUG) tests in the motion capture laboratory [25] in normal and cognitive dual-task conditions [26], with three replicates each. Patients were instructed to turn left on all TUG tests, consistent with our clinical testing paradigm. Performance was scored in person, and scores were confirmed from video if necessary.

2.1.3. Motion Capture

TUG tests were recorded using 3D optical motion capture (Motion Analysis Corporation, Santa Rosa, CA, USA). The motion capture facility is located in our clinical center and measures 5.8 m × 9.0 m with a capture area of 3.0 m × 4.6 m, and is equipped with 14 Osprey cameras with a resolution of 640 × 480 running at 120 Hz. During the testing session, patients wore tight-fitting clothes and were instrumented with reflective adhesive markers as recommended by the motion capture system manufacturer, configured as a superset of the Helen-Hayes kinematic marker set [27], incorporating additional markers on the hands. An example of the kinematic marker data is shown in Figure 1. Prior to analysis, all kinematic data were projected to a hip-centered coordinate system and normalized to zero mean and unit standard deviation.

2.2. Modeling

2.2.1. Model Overview

Applications for deep learning models range from industry [28] to healthcare [29], enabling breakthroughs in complex data analysis. Those also include human activity or motion analysis using a kinematic time-series dataset [30,31]. While high recognition accuracy is important, understanding the reasoning behind the model’s decisions is also essential. To understand these black-box models, various explainable deep learning techniques are introduced to detect complex patterns in time-series data to explain the model behavior in a human-interpretable manner [32]. This work uses a kind of interpretable deep learning methods, namely, attention-based approach [33,34], to quantify FOG in parkinsonism.
Our proposed model is an attention-based adaptive graphical convolutional network (AGCN, [34]) with adaptive trimming [35]. The overall model architecture is shown in Figure 2. We process the 3D motion capture data following a common deep learning-based human activity recognition paradigm [30,36]. Motion capture data from each testing sequence is comprised of three channels (x, anterior/posterior; y, lateral; z, vertical) for each of 60 kinematic markers, for a total of 180 independent channels. The data from each sequence is segmented into analysis windows of 4 s, N × C × T , where N = 60 , C = 3 , and T = 480 for 120 Hz signals, with 1 s intervals. A 4 s analysis window is chosen to capture a sufficient duration of FOG episodes while patients are walking [37].
Each 4 s analysis window is labeled with medication state (OFF/ON), FOG score (0, 1, 2, 3, or 4, from MDS-UPDRS-III item 3.11), and MDS-UPDRS-III total score, excluding the FOG score. The proposed model is trained to predict the labels for each analysis window based on the 3D kinematic data.
For extracting kinematic features from each 4 s window, the proposed model considers two aspects: (1) the core motion segment, which corresponds to the most relevant section of time within each window, and (2) the most relevant kinematic marker (joint) and edge between markers (bone) for the given multiple prediction tasks. The model uses adaptive trimming (AT) to identify the core motion segment within each 4 s analysis window and trims the given input signal for further analysis [35]. The trimmed core motion segment is processed to automatically identify the most relevant joint and limb parts for making predictions by using the AGCN model [34]. The AGCN model extracts feature representation by treating a given core motion sequence of 60 markers as a graphical model representing a human skeleton, where each node is marker position (joint) and the edge is connectivity between markers (bone) of ongoing kinematic sequence. The AGCN automatically learns the most important joint and bone motions across all samples (domain-dependent attention weights) and specific to given samples (input-specific attention weights) for predicting medication state, FOG score, and MDS-UPDRS-III total score, excluding FOG score.

2.2.2. Trimming Core Motion Segment

Adaptive Trimming (AT) enables the model to identify core motion segments and to flexibly trim the signal that is most useful for specific prediction tasks of interest. From a previous study, AT was very effective at detecting gym exercise classification task [35]. In this work, the AT is fully trained with a given kinematics dataset to predict the start and end time of the core motion segment from a 4 s analysis window, X R N × C × T .
c = s i g m o i d ( F c e n t e r ( F a t ( X ) ) )
w = exp ( F w i d t h ( F a t ( X ) ) )
F a t is a four-layer convolutional network for extracting feature to predict core motion locations. F c e n t e r and F w i d t h are two-layer fully connected models to predict center location, 0 < c < 1 and width of core motion segment, 0 < w < 1 , which are further processed to derive start, s, and end, e, indices of given window, where 0 < s , e < T .
s = T × s i g m o i d ( c w 2 )
e = T × s i g m o i d ( c + w 2 )
X C = X [ s : e ] = F c r o p ( X , s , e )
= F s a m p l e r ( F g r i d _ g e n ( X ) , s , e )
The cropping operation adapts grid generator, F g r i d _ g e n , and sampler, F s a m p l e r that is used in spatial transformer network (STN) [38] to learn differentiable geometric manipulator function for cropping 2D images for the most salient object in the scene for image recognition. For AT, F g r i d _ g e n generates 1D temporal grid with detected start, s, and end, e, indices of core motion signals and the temporal segment, X [ s : e ] R N × C × T is sampled with F s a m p l e r , where T = e s + 1 . This cropping operation resembles an interpolation process, which makes the whole AT model differential that can be trained with gradient back-propagation operation.

2.2.3. Adaptive Graph Convolution

The trimmed core motion segment is represented as temporal graphical sequence, G = ( V , E ) , where the node set, V = { v t i t = 1 , , T , i = 1 } , includes markers (joints) in a skeleton sequence. The edge set is composed of two subsets, in which the first edge subset is the intra-skeleton connectivity (limbs) E S = { v t i v t j ( i , j ) H } , where H is the set of connected joints defined by motion capture system, and second edge subset is the inter-frame edges, which connect the same joints in consecutive frames E F = { v t i v ( t + 1 ) i } .
Given a temporal-spatial graph representation of motion segment, we first encode spatial dimension by using an AGCN [34] with K v kernel size, which is defined as follows,
f o u t = k K v W k f i n ( A k + B k + C k )
where, f i n R C i n × T × N and f o u t R C o u t × T × N are input and output feature map and W k R C o u t × C i n × 1 × 1 is weight vector of the 1 × 1 convolution operation. A k = Λ 1 2 ( A ¯ k ) Λ k 1 2 is a normalized N × N adjacency matrix of defined skeleton structure from our motion capture system, where A ¯ k is a binary N × N adjacency matrix indicating the connectivity between the joints and Λ k i i = j ( A ¯ k i j ) + α is the normalized diagonal matrix using α = 0.001 to avoid empty rows.
The attention maps of each node (joint) and edge (limb) are encoded in B k and C k , which are learned fully data-driven manner. B k R N × N is an attention graph that encodes the underlining node and limb importance considering the entire samples of the task domain. B k is fixed once the parameters are trained and used for the inference. C k is an input-dependent attention graph to determine the strength of the connection between any two nodes in a given input graph sequence. Specifically, we applied embedded Gaussian Affinity [39] to calculate self-similarity between two nodes, v i and v j in a given input feature map, f i n .
C k i j = f ( v i , v j ) = e θ k ( v i ) T ϕ k ( v j ) j = 1 N e θ k ( v i ) T ( ϕ k ( v j ) )
Compared to A k and B k , C k can flexibly attend to more important joint and limb motions according to changing inputs at inference time. Combining A k (predefined skeletal connectivity), B k (domain-specific connectivity), and C k (input-specific connectivity) helps the model to fully adjust the graphical structure of the input sample to only focus on the motion signals that are useful for jointly predicting medication state, FOG score, and MDS-UPDRS-III total score excluding FOG score. Additionally, we did not restrict the learned B k and C k to be left and right body symmetric to take into account the potential for asymmetric symptoms [40,41].
To further encode temporal dimension, K t × 1 temporal convolution is applied to spatial feature, f o u t , extracted from the above mentioned attention-based graph convolution model, thereby, deriving spatial-temporal graphical representation, f o u t S T = c o n v K t × 1 ( f o u t ) . In this study, we use a four-layer temporal-spatial graphical convolutional network (TGCN) with 64 feature maps to encode core motion in the given 4 s analysis window. Temporal Average Pooling (TAP) [42] is applied to the output of the last layer to summarize the feature across the temporal axis.

2.2.4. Multi-Task Prediction

The feature representation from the last TGCN layer is used to simultaneously predict medication state, FOG score, and MDS-UPDRS-III total score excluding FOG score. (i) Medication state is a binary classification problem, either OFF or ON state. The feature representation is processed with two-layer fully connected model and a two-way softmax classifier, which is trained with binary cross-entropy loss. (ii) FOG score has 5 levels, from 0 (absent) to 4 (severe) FOG. For FOG score prediction, the feature representation is processed with two-layer fully connected model and five-way softmax classifier, which is trained with multi-class cross-entropy loss. (iii) MDS-UPDRS-III total score excluding FOG score is a positive integer ranging between 0 and 120. Before the model training, we apply Z-score normalization to marginalize the impact of outliers to bias the model prediction behaviors. To additionally consider the non-Gaussian distribution of the MDS-UPDRS-III total score excluding FOG score, we processed the feature representation with Gaussian Mixture Model (GMM) regression model [43].
p ( y x ) = i = 1 m α i ( x ) N ( μ i ( x ) , σ i 2 ( x ) )
where x R D is feature representation from the last TGCN layer, α = s o f t m a x ( f α ( x ) ) is mixing coefficients for Gaussian distributions and μ = f μ ( x ) and σ = e x p ( f σ ( x ) ) are mean and standard deviation of each Gaussian distribution. For projection functions, f α , f μ , f σ , two-layer fully connected models were used. In our experiment, the naive regression with a two-layer fully connected model and mean square error having a single Gaussian distribution assumption did not converge when training.

3. Experiment Setting

3.1. Model Hyperparameter, Training and Evaluation

For choosing model hyperparameters, we have followed the hyperparameter suggested in Ordóñez and Roggen [30], which was sufficient to predict fine-grained and complex human activities using movement data. (i) AT: Temporal kernel size and feature map were 3 × 1 and 64, respectively, for all four layers of the temporal convolutional model, F a t . Max pooling with × 1 2 was used at each output layer for aggregating temporal dimension. For predicting center location and width size of core motion segment, F c e n t e r and F w i d t h , two-layer fully connected layer model was used with 128 units and ReLU activation function [44]. (ii) AGCN: Four layers of the temporal graphical convolutional model were used, and kernel sizes of K v = 3 and K t = 5 , respectively, for graphical and temporal convolution. Across all layers and convolutions, we used 64 feature maps, ReLU activation function, and max pooling with × 1 2 to aggregate along the temporal dimension. (iii) Multi-task Prediction: For two-layer fully connected models to predict medication state, FOG score, and GMM regression parameters, we used 256 and 128 units with ReLU activation function.
For the training model, we used a learning rate fixed at 1 × 10 3 with Adam optimizer and used a batch size of 16. Model training was stopped when no decrease in loss is observed from the validation set, which model is also used for evaluating the test set.
For evaluating the proposed method, we used 10-fold cross-validation. At each fold, 50%, 20%, and 30% of the dataset was used for the training, validation, and testing sets, respectively. We avoided placing adjacent analysis windows in different folds to avoid pairwise similarity biasing the cross-validation results [45].
To understand the potential use case for real-time analysis, we also compare the inference time between each baseline model, especially deep learning models. Based on standard benchmarking practices, we report an average of 1000 runs for a batch size of 16 using a NVIDIA QUADRO RTX 6000 with 24 GB RAM memory size.

3.2. Performance Metrics

For performance metrics, we used binary F1 score and mean F1 score for medication state and FOG score prediction, respectively, which is widely used for evaluating prediction performance in the presence of label imbalance. As shown in Table 2, most participants had FOG scores 2 for both OFF and ON medication states. The mean F1 score is an average of per-class F1 score, which is the harmonic mean of precision and recall of each class.
P r e c i s i o n c = T P c T P c + F P c
R e c a l l c = T P c T P c + F N c
F 1 s c o r e c = 2 × P r e c i s i o n c × R e c a l l c P r e c i s i o n c + R e c a l l c
M e a n F 1 s c o r e = 1 C c C F 1 s c o r e c
where C is the number of classes and C = 5 for FOG Score classification. For a class c, T P c is a true positive that represents the total of successfully classified class windows, F P c is a false positive that represents the total misclassified class windows, and F N c is a false negative that represents the total misclassified non-class windows.
For evaluating the regression performance for MDS-UPDRS-III total score excluding FOG score, we used root mean square error (RMSE).

3.3. Comparison with Baseline Models

We compared the proposed model to: (i) shallow models with hand-crafted features, and (ii) deep learning models including convolutional networks and graphical convolutional networks. We compared classifier performance across models using 95% Wilson score confidence intervals [46] for Medication State and FOG Score and using standard normal approximation based 95 % confidence intervals for MDS-UPDRS-III. For an imbalanced dataset, such as ours, Wilson score is widely used to calculate confidence intervals, as discussed in Appendix A.
Shallow Baseline Models. The first baseline models we considered were shallow models, such as Random Forest (RF) and Support Vector Machine (SVM) with radial basis function (RBF) kernel, with FOG-related hand-crafted features. Following previous work [13,47,48], we extracted various time, frequency, and distribution features, including freezing index [49], variance, sample entropy [50], central frequency, dominant frequency, and wavelet mean [51] features from the acceleration signals at multiple on-body locations. We used second-order Savitsky-Golay differentiation to derive acceleration traces from joint marker kinematics.
To investigate whether the lateralization of parkinsonian symptoms would impact model performance, We iterated RF and SVM models using markers from the left side of the body only (RF-L, SVM-L) and using markers from both sides (RF-LR, SVM-LR). We focused on lower body parts and independently trained RF and SVM for each task separately, following previous work [13,47].
Deep Baseline Models. We also compared the proposed model to several deep learning models for processing human skeleton time-series, including Temporal convolutional network (TCN) [52], Graphical convolutional network (GCN) [53], GCN with attention model (AGCN) [34], and AGCN with Adaptive Trimming (AT+AGCN). We used identical hyperparameters for model architecture and training wherever possible in order to make the fairest possible comparisons between deep learning models. All models were 4-layer with 64 feature maps and × 1 2 max pooling. For deep learning models, and for the classification and regression, we used two-layer fully connected layer with 256 and 128 units with ReLU activation functions.

3.4. Comparison with Single-Task Prediction

Since the proposed model is constrained to learn features relevant to three simultaneous prediction tasks, we reasoned that the identified features might be sub-optimal for single task prediction, leading to decreased performance. Therefore, we re-trained the deep learning models (with the exception of AT+AGCN+GMM) on the FOG score prediction task only and assessed changes in performance. We did not include the AT+AGCN+GMM model in this analysis as without the MDS-UPDRS-III prediction task it is identical to the AT+AGCN model.

3.5. Model Interpretability

We considered it critical to assess the clinical relevance of features identified by the model as relevant to medication state, FOG score, or total MDS-UPDRS-III score. These included individual kinematic markers (often referred to as “joints” in the computer vision literature) and segments (“limbs”) with high attention scores, and kinematic marker trajectories with high relevance to particular labels.
To derive overall model attention to individual segments or limbs, we aggregated attention maps across all samples in the dataset by averaging the learned attention maps and graphical structure over all M samples:
E A = 1 M × K i M k K v ( A k + B k + C k i ) R N × N
where A k , B k , and C k ( x i ) are the normalized N × N adjacency matrix of predefined skeleton structure, the domain-wise N × N attention map, and the input-dependent N × N attention map at each kernel, respectively. The attention weights for joints and segments are then defined as the diagonal components E A j j and off-diagonal components E A i j , j i of E A , respectively.
To identify individual kinematic marker trajectories and core motion segments with high relevance to particular labels, we visualized individual analysis windows and core motion segments that the model predicted with high confidence, as measured by the entropy of the class prediction distribution. We visualized these data and discussed the interpretation with clinician experts within our project team, within the movement disorders group at our center, and at a regional forum in the Atlanta area hosted by the study sponsor in order to assess whether the identified features were consistent with the features that movement disorders specialists are trained to identify as characteristics of freezing.

3.6. Model Performance Furthermore, Potential Bias

After evaluating the proposed model against other candidate models, we assessed the potential for bias in model performance associated with participant demographics. After computing individual F1 score for each participant, we compared model performance across age and sex with linear models. Linear models used FOG study group (PD-FOG, PD-NoFOG, PP-FOG), dichotomized age, and sex as predictors of individual F1 score. Statistical significance was assessed with Wald tests at p = 0.05.

4. Results

4.1. Overall Model Performance

Here, we report the overall prediction performance of the proposed model (AT+AGCN+GMM in Table 3), compared with the performance of baseline models for predicting medication state, FOG score, and MDS-UPDRS-III total score excluding FOG item. In general, the proposed model’s performance was very high for both classification and regression tasks: Medication State, 97.6% cross-validated F1 score; FOG Score, 96.8% cross-validated F1 score; and MDS-UPDRS-III, 2.7 point RMSE, which is within the minimal clinically-important difference [54] for the instrument. In particular, the addition of the GMM regression component—which learns non-Gaussian distributions flexibly—to the second-best performing model architecture (AT+AGCN) significantly improved MDS-UPDRS-III performance. Performance of all models is summarized numerically in Table 3.
The prediction performance of the proposed model on MDS-UPDRS-III score excluding FOG item is shown in Figure 3. The overall RMSE was 2.7 ± 0.4 points. As expected, overall, ON medication sessions have lower and OFF medication sessions have higher MDS-UPDRS-III total scores, as indicated by the higher prevalence of red points to the left of the plot and the higher prevalence of blue points to the right of the plot. We noted that the model tended to overestimate lower scores and under-estimate higher scores, as indicated by datapoints in the upper left and lower right.

4.2. Comparison to Baseline Models

For comparison to the previous state-of-the-art models in FOG analysis, we started analysis with shallow models (RF and SVM) using only lower body parts. We tested the use of left only and both left and right lower body parts (RF-LR and SVM-LR). Using both sides of body significantly improved performance on all three tasks; increasing F1 score by 12% and 33% for Medication State and FOG Score, respectively, and decreasing MDS-UPDRS-III RMSE by 10%. Among the shallow models (RF-LR and SVM-LR), RF significantly outperformed SVM, increasing F1 score by 7% and 13% for Medication State and FOG Score, respectively, and decreasing MDS-UPDRS-III RMSE by 9%, presumably due to its ability to learn non-linear decision boundaries.
Deep learning models also substantially outperformed shallow ML models, providing evidence that learning FOG representations that capture complex patterns may be more effective than using existing hand-crafted FOG features. The same conclusion was reached in a previous study using wait-worn inertial measurement units (IMUs) and standard deep learning models for analyzing FOG gait [55]. Deep learning models also outperform handcrafted features with various machine learning applications in human activity recognition [30,31]. Compared to the best performing shallow model (RF-LR), the TCN model, which mainly captures temporal patterns of each joint movement sequence, improved F1 score by 33% and 24% for Medication State and FOG Score, respectively, and decreased MDS-UPDRS-III RMSE by 43%.
Among the deep learning models, we also noted significant performance improvements in F1 scores among graph-based models vs. the more traditional TCN, as graph-based models can additionally capture positional relations between joints with a graphical data structure defined as a human skeleton. The simplest graph-based model significantly outperformed the TCN on all three tasks (4%, 9%, and 11% improvements on medication state, FOG score, and MDS-UPDRS-III, respectively). Further significant improvements were noted with the addition of attention mechanisms which enable the model to adaptively concentrate its representation powers for the most relevant joint depending on the given input (4%, 2%, and 8%). The additions of adaptive trimming and the Gaussian mixture model prediction did not significantly improve F1 scores, but significantly reduced MDS-UPDRS-III RMSE (4% and 23%, respectively). We speculate that the flexibility of the GMM model stabilized the gradient backpropagated from the regression branch to help find a more effective feature representation for all tasks. Runtime increased with model complexity, as expected. Nevertheless, the proposed model, which is the most complex, showed a sub-second inference time, enough for real-time analysis.

4.3. Comparison to Single-Task Prediction

All four deep learning models tested showed significantly improved performance on FOG score prediction when trained on the multi-task problem (medication state, FOG score, and MDS-UPDRS-III) rather the single task problem (FOG Score only). When the models were trained on the single task problem, the TCN, GCN, AGCN, AT+AGCN demonstrated F1 scores of 0.825 ± 0.016 , 0.892 ± 0.033 , 0.903 ± 0.047 , and 0.925 ± 0.012 , respectively, a 3.8% decrease in performance on average compared to the multi-task problem. We speculate that additional information provided to the models by predicting medication states and MDS-UPDRS-III total scores helped to learn representations that are more targeted and personalized to discriminate detailed differences in FOG phenotypes in varying PD conditions, which eventually helped improve overall FOG score classification performance.

4.4. Model Interpretability

4.4.1. Most Relevant Joints and Limbs

We visualized markers and segments with the top ten largest attention weights to assess which body parts were most salient to the prediction task (Figure 4). Attention weights were concentrated on the head, chest, waist, hands, and (particularly left) legs. We suggest that the attention paid to markers on essentially all body segments reflects the fact that FOG is a full-body phenomenon, and suggests that the model may be attending to en-bloc turns [56]—which tend to be maintained across medication states [57]—or other elements of impaired intersegmental coordination. We noted that in particular, the model attended closely to segments on the left foot, which had been suggested previously by a clinical expert on our team as relevant to FOG in this testing condition, which requires left turns. Interestingly, the model also attended to the fingers and elbows. Although these body parts are not typically attended to during clinical FOG examination, patients with FOG can also freeze during upper limb movements [58], leaving open the possibility that the model was attending to hand movements characteristic of freezing.

4.4.2. Most Relevant Motion Segments

We also visualized patterns of left heel movement that were predicted as relevant to particular medication states and FOG scores with high confidence. Figure 5 shows the detected core motion segments of left heel movements from the adaptive trimming model during the timed-up-and-go trials of patients with ON and OFF medication states and different FOG scores. In general, the adaptive trimming model automatically captured approximately a single step cycle within each 4 s analysis window with movement patterns especially related to FOG.
The identified kinematic associated with detected core segments were highly consistent with the features that movement disorders specialists are trained to identify as characteristics of freezing. This analysis shows that the model focuses on periods with regular gait activity for epochs corresponding to FOG scores of 0 and 1, and periods of interrupted gait activity or pure akinesia for epochs with higher scores. For the samples with a FOG score of 0 (first column), the model considered normal stepping gait and used a walking cycle motion for making predictions. For the samples with a FOG score of 1 (second column), the models detected decreasing step length from the motion automatically. As the FOG score became higher, the model tended to detect more FOG-related gait motions. For the samples with a FOG score of 2 (third column), the model detected onset of gait signal related to festination (tendency to speed up in parallel with a loss of normal amplitude of repetitive movements) [59]. For the samples with FOG score of 3 and 4 (fourth and fifth columns), the model detected freezing gait, akinesia, and trembling signals as core motion signal that is relevant for predicting FOG scores.

4.5. Classifier Performance Furthermore, Potential Bias

After computing individual F1 score for each participant, we compared model performance across study and demographic groups to assess potential bias. Linear models found no significant differences in F1 score across study groups or sex, but found significantly decreased performance (reduction in F1 score of 17%, p < 0.01) among older participants (age ≥ 69 years) compared to younger participants.

5. Discussion and Conclusions

In our experiment, we designed a deep neural network model to simultaneously predict levodopa medication state (ON/OFF), FOG score (0–4), and MDS-UPDRS-III total score (less FOG score) from full-body kinematics data of 57 patients, including 5 patients with atypical parkinsonism, assessed with TUG tests in the off and on medication state. As compared to formal clinical assessments by a movement disorders specialist, our AGCN model classified levodopa medication state and FOG score with 96.4% and 96.2% F1 scores, respectively, and regressed MDS-UPDRS-III total score with root mean square error (RMSE) of 2.7 points.
To the best of our knowledge, this is the first work that applies an interpretable deep learning model with full-body kinematics for classifying FOG. This model detects time segments having characteristic movements of FOG during walking, e.g., small shuffling steps, akinesia, and tremulousness. Additional findings demonstrated that FOG is not limited to the lower extremity, and also significantly involves movements in the upper body, further supporting that FOG requires phenotyping using whole-body kinematics. Findings from our analysis may lead to novel hypotheses to define more granular FOG phenotypes, or potentially to technologies that enable continuous monitoring of FOG severity in order to test new therapies with improved precision.
Overall, while the current study uses 3D kinematic data, we believe that the underlying approach will generalize to motion estimates obtained through pose estimation or body-worn sensors, enabling future applications in clinical and home settings with 2D video. The patterns of body motion recorded here result from fundamental principles of physics and biomechanics, which are likely to hold regardless of the method used to measure motion. For example, the laws of motion and principles of energy conservation apply regardless of whether motion is measured using 3D kinematic data, pose estimation, or body-worn sensors. This is likely why it is feasible to estimate virtual IMU signals from video data [35]. The proposed model is expected to be also useful for discovering novel phenotypes for other types of movement disorders, such as, Cerebral Palsy [60], Ataxia [61], or Essential Tremor [62], which conventionally use hand-crafted features for the analysis.
The study has three main limitations. First, we did not attempt to identify freezing of gait (FOG) at the millisecond level, which would be necessary for use in assistive technology. Second, we did not attempt to measure FOG severity as a continuous outcome, which could increase precision in clinical trials. Finally, the study sample was predominantly white and had fewer females than would be representative of the Parkinson’s disease (PD) population [63], so the generalizability of the results to the entire PD population may be limited.
One primary contribution of this work is the application of deep learning to the problem of scoring FOG, which has primarily been examined with hand-crafted and engineered features such as spectral power in a prespecified “freeze band” [16] calculated from a prespecified set of body segments. Indeed, despite the typical notion that FOG is an interruption of walking—leg movements—our results indicate that scoring FOG with high accuracy may require attention to body parts across the body, including the hands and head. We believe that adopting a data-driven approach with explainable deep learning models represents an important way forward in modeling kinematics from walking and turning motions of parkinsonian patients.
We hope that using deep learning to discover data-driven kinematic features will lead to the development of a more fine-grained and objective FOG severity scales, which could provide valuable information to clinicians and researchers, help to improve diagnosis, treatment, and overall management of FOG (cf. [10]).

Author Contributions

Conceptualization, H.K., S.A.F. and J.L.M.; methodology, H.K.; software, H.K.; validation, H.K., I.G. and J.L.M.; formal analysis, H.K. and J.L.M.; investigation, H.K. and J.L.M.; resources, S.A.F., J.L.M., G.D.C. and C.D.E.; data curation, D.B. and J.L.M.; writing—original draft preparation, H.K.; writing—review and editing, J.L.M., G.D.C., I.G., C.D.E., S.A.F.; visualization, H.K.; supervision, G.D.C., J.L.M. and S.A.F.; project administration, J.L.M. and S.A.F.; funding acquisition, S.A.F. and J.L.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by The McCamish Center for Parkinson’s Disease Innovation, the by Curtis Family Fund, Sartain Lanier Family Foundation (S.A.F.), and NIH K25HD086276 (J.L.M.).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of Emory University (IRB Number 2688 approved on 2 June 2021 and IRB Number 73518 approved on 10 December 2014).

Informed Consent Statement

Written informed consent was obtained from all subjects involved in the study according to procedures approved by the Emory University Institutional Review Board.

Data Availability Statement

The deidentified raw data and code for supporting the evidences in this work will be made available by the corresponding author upon reasonable request.

Acknowledgments

The computational needs for this research are supported in part by Oracle Cloud credits and related resources provided by the Oracle for Research program.

Conflicts of Interest

J.L.M. performs paid consulting work for Biocircuit technologies. None of these interests are directly related to the outcomes of this study. S.A.F. has the following competing interests: Honoraria: Lundbeck, Teva, Sunovion, Biogen, Acadia, Neuroderm, Acorda, CereSpire. Grants: Ipsen, Medtronics, Boston Scientific, Teva, US World Meds, Sunovion Therapeutics, Vaccinex, Voyager, Jazz Pharmaceuticals, Lilly, CHDI Foundation, Michael J. Fox Foundation, NIH, Royalties: Demos, Blackwell Futura for textbooks, Uptodate, Other Bracket Global LLC, CNS Ratings LLC. None of these interests are directly related to the outcomes of this study. The other authors declare no other interests. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A. Wilson Score Interval

Binomial proportion confidence interval calculates the outcome of series of Bernoulli trials to estimate the confidence interval for the probability of success. Wilson score interval is a asymmetric approximation of binomial confidence interval, which tackles two problems that rises when using naive symmetric normal approximated confidence interval [64], which are overshoot and zero width intervals [46]. Moreover, wilson score interval is robust with small samples and skewed observations as in our dataset (Table 2), which is common in human behavior analysis problems [33,35]. Wilson score interval can be calculated as follows:
p 1 1 + z 2 n p ^ + z 2 2 n ± z 1 + z 2 n p ^ ( 1 p ^ ) n + z 2 4 n 2
where p ^ is the success probability and n is the number of experiments. For 95% confidence interval, z = 1.96 .

References

  1. Pringsheim, T.; Jette, N.; Frolkis, A.; Steeves, T. The prevalence of Parkinson’s disease: A systematic review and meta-analysis. Mov. Disord. 2014, 29, 1583–1590. [Google Scholar] [CrossRef] [PubMed]
  2. Dorsey, E.; Sherer, T.; Okun, M.S.; Bloem, B.R. The emerging evidence of the Parkinson pandemic. J. Parkinson’s Dis. 2018, 8, S3–S8. [Google Scholar] [CrossRef] [PubMed]
  3. Lucas McKay, J.; Goldstein, F.C.; Sommerfeld, B.; Bernhard, D.; Perez Parra, S.; Factor, S.A. Freezing of Gait can persist after an acute levodopa challenge in Parkinson’s disease. NPJ Parkinson’s Dis. 2019, 5, 25. [Google Scholar] [CrossRef]
  4. Nonnekes, J.; Snijders, A.H.; Nutt, J.G.; Deuschl, G.; Giladi, N.; Bloem, B.R. Freezing of gait: A practical approach to management. Lancet Neurol. 2015, 14, 768–778. [Google Scholar] [CrossRef] [PubMed]
  5. Factor, S.A.; Jennings, D.L.; Molho, E.S.; Marek, K.L. The natural history of the syndrome of primary progressive freezing gait. Archiv. Neurol. 2002, 59, 1778–1783. [Google Scholar] [CrossRef] [PubMed]
  6. Haddad, Y.K.; Bergen, G.; Florence, C. Estimating the economic burden related to older adult falls by state. J. Public Health Manag. Pract. JPHMP 2019, 25, E17. [Google Scholar] [CrossRef] [PubMed]
  7. Florence, C.S.; Bergen, G.; Atherly, A.; Burns, E.; Stevens, J.; Drake, C. Medical costs of fatal and nonfatal falls in older adults. J. Am. Geriatr. Soc. 2018, 66, 693–698. [Google Scholar] [CrossRef]
  8. Pelicioni, P.H.; Menant, J.C.; Latt, M.D.; Lord, S.R. Falls in Parkinson’s disease subtypes: Risk factors, locations and circumstances. Int. J. Environ. Res. Public Health 2019, 16, 2216. [Google Scholar] [CrossRef]
  9. Goetz, C.G.; Tilley, B.C.; Shaftman, S.R.; Stebbins, G.T.; Fahn, S.; Martinez-Martin, P.; Poewe, W.; Sampaio, C.; Stern, M.B.; Dodel, R.; et al. Movement Disorder Society-sponsored revision of the Unified Parkinson’s Disease Rating Scale (MDS-UPDRS): Scale presentation and clinimetric testing results. Mov. Disord. Off. J. Mov. Disord. Soc. 2008, 23, 2129–2170. [Google Scholar] [CrossRef]
  10. Hulzinga, F.; Nieuwboer, A.; Dijkstra, B.W.; Mancini, M.; Strouwen, C.; Bloem, B.R.; Ginis, P. The New Freezing of Gait Questionnaire: Unsuitable as an outcome in clinical trials? Mov. Disord. Clin. Pract. 2020, 7, 199–205. [Google Scholar] [CrossRef]
  11. Hatcher-Martin, J.; McKay, J.; Sommerfeld, B.; Howell, J.; Goldstein, F.; Hu, W.; Factor, S. Cerebrospinal fluid Aβ42 and fractalkine are associated with Parkinson’s disease with freezing of gait. medRxiv 2020, 1–12. [Google Scholar] [CrossRef]
  12. Forsaa, E.; Larsen, J.; Wentzel-Larsen, T.; Alves, G. A 12-year population-based study of freezing of gait in Parkinson’s disease. Park. Relat. Disord. 2015, 21, 254–258. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Yan, W.; Yao, Y.; Bint Ahmed, J.; Tan, Y.; Gu, D. Prediction of freezing of gait in patients with Parkinson’s disease by identifying impaired gait patterns. IEEE Trans. Neural Syst. Rehab. Eng. 2020, 28, 591–600. [Google Scholar] [CrossRef]
  14. Yan, Y.; Liu, Y.; Li, C.; Wang, J.; Ma, L.; Xiong, J.; Zhao, X.; Wang, L. Topological Descriptors of Gait Nonlinear Dynamics toward Freezing-of-Gait Episodes Recognition in Parkinson’s Disease. IEEE Sens. J. 2022, 22, 4294–4304. [Google Scholar] [CrossRef]
  15. Silva de Lima, A.; Evers, L.J.; Hahn, T.; Bataille, L.; Hamilton, J.; Little, M.; Okuma, Y.; Bloem, B.; Faber, M. Freezing of gait and fall detection in Parkinson’s disease using wearable sensors: A systematic review. J. Neurol. 2017, 264, 1642–1654. [Google Scholar]
  16. Yungher, D.; Morris, T.; Dilda, V.; Shine, J.; Naismith, S.; Lewis, S.J.; Moore, S.T. Temporal characteristics of high-frequency lower-limb oscillation during freezing of gait in Parkinson’s disease. Parkinson’s Dis. 2014, 2014, 606427. [Google Scholar] [CrossRef]
  17. Rodríguez-Martín, D.; Samà, A.; Pérez-López, C.; Català, A.; Moreno Arostegui, J.M.; Cabestany, J.; Bayés, À.; Alcaine, S.; Mestre, B.; Prats, A.; et al. Home detection of freezing of gait using support vector machines through a single waist-worn triaxial accelerometer. PLoS ONE 2017, 12, e0171764. [Google Scholar] [CrossRef]
  18. Tahafchi, P.; Molina, R.; Roper, J.A.; Sowalsky, K.; Hass, C.J.; Gunduz, A.; Okun, M.S.; Judy, J.W. Freezing-of-Gait detection using temporal, spatial, and physiological features with a support-vector-machine classifier. In Proceedings of the 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Jeju Island, Republic of Korea, 11–15 July 2017; pp. 2867–2870. [Google Scholar]
  19. Reches, T.; Dagan, M.; Herman, T.; Gazit, E.; Gouskova, N.A.; Giladi, N.; Manor, B.; Hausdorff, J.M. Using wearable sensors and machine learning to automatically detect freezing of gait during a FOG-provoking test. Sensors 2020, 20, 4474. [Google Scholar]
  20. Ferrari, A.; Ginis, P.; Hardegger, M.; Casamassima, F.; Rocchi, L.; Chiari, L. A mobile Kalman-filter based solution for the real-time estimation of spatio-temporal gait parameters. IEEE Trans. Neural Syst. Rehab. Eng. 2015, 24, 764–773. [Google Scholar] [CrossRef]
  21. Diep, C.; O’Day, J.; Kehnemouyi, Y.; Burnett, G.; Bronte-Stewart, H. Gait Parameters Measured from Wearable Sensors Reliably Detect Freezing of Gait in a Stepping in Place Task. Sensors 2021, 21, 2661. [Google Scholar] [CrossRef]
  22. Mancini, M.; Shah, V.V.; Stuart, S.; Curtze, C.; Horak, F.B.; Safarpour, D.; Nutt, J.G. Measuring freezing of gait during daily-life: An open-source, wearable sensors approach. J. NeuroEng. Rehab. 2021, 18, 1. [Google Scholar] [CrossRef]
  23. Hatcher-Martin, J.M.; McKay, J.L.; Pybus, A.F.; Sommerfeld, B.; Howell, J.C.; Goldstein, F.C.; Wood, L.; Hu, W.T.; Factor, S.A. Cerebrospinal fluid biomarkers in Parkinson’s disease with freezing of gait: An exploratory analysis. NPJ Parkinsons Dis. 2021, 7, 105. [Google Scholar] [CrossRef]
  24. Hughes, A.J.; Daniel, S.E.; Kilford, L.; Lees, A.J. Accuracy of clinical diagnosis of idiopathic Parkinson’s disease: A clinico-pathological study of 100 cases. J. Neurol. Neurosurg. Psychiatr. 1992, 55, 181–184. [Google Scholar] [CrossRef]
  25. Nocera, J.R.; Stegemöller, E.L.; Malaty, I.A.; Okun, M.S.; Marsiske, M.; Hass, C.J.; National Parkinson Foundation Quality Improvement Initiative Investigators. Using the Timed Up & Go test in a clinical setting to predict falling in Parkinson’s disease. Archiv. Phys. Med. Rehab. 2013, 94, 1300–1305. [Google Scholar]
  26. Shumway-Cook, A.; Baldwin, M.; Polissar, N.L.; Gruber, W. Predicting the probability for falls in community-dwelling older adults. Phys. Ther. 1997, 77, 812–819. [Google Scholar] [CrossRef]
  27. Kadaba, M.P.; Ramakrishnan, H.; Wootten, M. Measurement of lower extremity kinematics during level walking. J. Orthop. Res. 1990, 8, 383–392. [Google Scholar] [CrossRef]
  28. Yu, Y.; Samali, B.; Rashidi, M.; Mohammadi, M.; Nguyen, T.N.; Zhang, G. Vision-based concrete crack detection using a hybrid framework considering noise effect. J. Build. Eng. 2022, 61, 105246. [Google Scholar] [CrossRef]
  29. Soffer, S.; Ben-Cohen, A.; Shimon, O.; Amitai, M.M.; Greenspan, H.; Klang, E. Convolutional neural networks for radiologic images: A radiologist’s guide. Radiology 2019, 290, 590–606. [Google Scholar] [CrossRef]
  30. Ordóñez, F.J.; Roggen, D. Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef]
  31. Du, Y.; Wang, W.; Wang, L. Hierarchical recurrent neural network for skeleton based action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1110–1118. [Google Scholar]
  32. Zhang, Y.; Tiňo, P.; Leonardis, A.; Tang, K. A survey on neural network interpretability. IEEE Trans. Emerg. Top. Comput. Intell. 2021, 5, 726–742. [Google Scholar] [CrossRef]
  33. Murahari, V.S.; Plötz, T. On attention models for human activity recognition. In Proceedings of the 2018 ACM International Symposium on Wearable Computers, Singapore, 8–12 October 2018; pp. 100–103. [Google Scholar]
  34. Shi, L.; Zhang, Y.; Cheng, J.; Lu, H. Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12026–12035. [Google Scholar]
  35. Kwon, H.; Abowd, G.; Plötz, T. Complex Deep Neural Networks from Large Scale Virtual IMU Data for Effective Human Activity Recognition Using Wearables. Sensors 2021, 21, 8337. [Google Scholar] [CrossRef]
  36. Hammerla, N.Y.; Halloran, S.; Plötz, T. Deep, Convolutional, and Recurrent Models for Human Activity Recognition Using Wearables. In Proceedings of the IJCAI, New York, NY, USA, 9–15 July 2016. [Google Scholar]
  37. Bachlin, M.; Plotnik, M.; Roggen, D.; Maidan, I.; Hausdorff, J.M.; Giladi, N.; Troster, G. Wearable assistant for Parkinson’s disease patients with the freezing of gait symptom. IEEE Trans. Inf. Technol. Biomed. 2009, 14, 436–446. [Google Scholar] [CrossRef]
  38. Jaderberg, M.; Simonyan, K.; Zisserman, A.; Kavukcuoglu, K. Spatial transformer networks. Adv. Neural Inf. Process. Syst. 2015, 28. [Google Scholar] [CrossRef]
  39. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  40. Toth, C.; Rajput, M.; Rajput, A.H. Anomalies of asymmetry of clinical signs in parkinsonism. Mov. Disord. Off. J. Mov. Disord. Soc. 2004, 19, 151–157. [Google Scholar] [CrossRef]
  41. Djaldetti, R.; Ziv, I.; Melamed, E. The mystery of motor asymmetry in Parkinson’s disease. Lancet Neurol. 2006, 5, 796–802. [Google Scholar] [CrossRef]
  42. Lin, M.; Chen, Q.; Yan, S. Network in network. arXiv 2013, arXiv:1312.4400. [Google Scholar]
  43. Bishop, C.M. Mixture Density Networks; Aston University: Birmingham, UK, 1994. [Google Scholar]
  44. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the Icml, Haifa, Israel, 21–24 June 2010. [Google Scholar]
  45. Hammerla, N.Y.; Plötz, T. Let us (not) stick together: Pairwise similarity biases cross-validation in activity recognition. In Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Umeda, Osaka, Japan, 7–11 September 2015; pp. 1041–1051. [Google Scholar]
  46. Wilson, E.B. Probable inference, the law of succession, and statistical inference. J. Am. Stat. Assoc. 1927, 22, 209–212. [Google Scholar] [CrossRef]
  47. Naghavi, N.; Wade, E. Prediction of freezing of gait in Parkinson’s disease using statistical inference and lower–limb acceleration data. IEEE Trans. Neural Syst. Rehab. Eng. 2019, 27, 947–955. [Google Scholar] [CrossRef]
  48. Palmerini, L.; Rocchi, L.; Mazilu, S.; Gazit, E.; Hausdorff, J.M.; Chiari, L. Identification of characteristic motor patterns preceding freezing of gait in Parkinson’s disease using wearable sensors. Front. Neurol. 2017, 8, 394. [Google Scholar] [CrossRef]
  49. Moore, S.T.; MacDougall, H.G.; Ondo, W.G. Ambulatory monitoring of freezing of gait in Parkinson’s disease. J. Neurosci. Methods 2008, 167, 340–348. [Google Scholar] [CrossRef]
  50. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol.-Heart Circul. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [PubMed]
  51. Baby, M.S.; Saji, A.; Kumar, C.S. Parkinsons disease classification using wavelet transform based feature extraction of gait data. In Proceedings of the 2017 International Conference on Circuit, Power and Computing Technologies (ICCPCT), Kollam, India, 20–21 April 2017; pp. 1–6. [Google Scholar]
  52. Du, Y.; Fu, Y.; Wang, L. Skeleton based action recognition with convolutional neural network. In Proceedings of the 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), Kuala Lumpur, Malaysia, 3–6 November 2015; pp. 579–583. [Google Scholar]
  53. Yan, S.; Xiong, Y.; Lin, D. Spatial temporal graph convolutional networks for skeleton-based action recognition. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  54. Horváth, K.; Aschermann, Z.; Ács, P.; Deli, G.; Janszky, J.; Komoly, S.; Balázs, É.; Takács, K.; Karádi, K.; Kovács, N. Minimal clinically important difference on the Motor Examination part of MDS-UPDRS. Park. Relat. Disord. 2015, 21, 1421–1426. [Google Scholar] [CrossRef] [PubMed]
  55. Camps, J.; Sama, A.; Martin, M.; Rodriguez-Martin, D.; Perez-Lopez, C.; Arostegui, J.M.M.; Cabestany, J.; Català, A.; Alcaine, S.; Mestre, B.; et al. Deep learning for freezing of gait detection in Parkinson’s disease patients in their homes using a waist-worn inertial measurement unit. Knowl.-Based Syst. 2018, 139, 119–131. [Google Scholar] [CrossRef]
  56. Khobkhun, F.; Hollands, M.; Tretriluxana, J.; Srivanitchapoom, P.; Richards, J.; Ajjimaporn, A. Benefits of task-specific movement program on en bloc turning in Parkinson’s disease: A randomized controlled trial. Physiother. Res. Int. 2022, 27, e1963. [Google Scholar] [CrossRef] [PubMed]
  57. Hong, M.; Earhart, G.M. Effects of medication on turning deficits in individuals with Parkinson’s disease. J. Neurol. Phys. Ther. 2010, 34, 11–16. [Google Scholar] [CrossRef]
  58. Heremans, E.; Nackaerts, E.; Vervoort, G.; Vercruysse, S.; Broeder, S.; Strouwen, C.; Swinnen, S.P.; Nieuwboer, A. Amplitude manipulation evokes upper limb freezing during handwriting in patients with Parkinson’s disease with freezing of gait. PLoS ONE 2015, 10, e0142874. [Google Scholar] [CrossRef]
  59. Imai, H. Festination and freezing. Rinsho Shinkeigaku = Clin. Neurol. 1993, 33, 1307–1309. [Google Scholar]
  60. Dobson, F.; Morris, M.E.; Baker, R.; Graham, H.K. Gait classification in children with cerebral palsy: A systematic review. Gait Post. 2007, 25, 140–152. [Google Scholar] [CrossRef]
  61. Nunes, A.S.; Kozhemiako, N.; Stephen, C.D.; Schmahmann, J.D.; Khan, S.; Gupta, A.S. Automatic Classification and Severity Estimation of Ataxia From Finger Tapping Videos. Front. Neurol. 2022, 12, 2587. [Google Scholar] [CrossRef]
  62. Cheng, F.; Duan, Y.; Jiang, H.; Zeng, Y.; Chen, X.; Qin, L.; Zhao, L.; Yi, F.; Tang, Y.; Liu, C. Identifying and distinguishing of essential tremor and Parkinson’s disease with grouped stability analysis based on searchlight-based MVPA. BioMedical Eng. OnLine 2022, 21, 81. [Google Scholar] [CrossRef]
  63. GBD 2016 Parkinson’s Disease Collaborators. Global, regional, and national burden of Parkinson’s disease, 1990–2016: A systematic analysis for the Global Burden of Disease Study 2016. Lancet Neurol. 2018, 17, 939–953. [Google Scholar] [CrossRef] [PubMed]
  64. Wallis, S. Binomial confidence intervals and contingency tests: Mathematical fundamentals and the evaluation of alternative methods. J. Quant. Linguist. 2013, 20, 178–208. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Motion capture recording during timed-up-and-go testing. Left: clinical motion capture laboratory. Right: example of kinematic marker data. Participants were instructed to rise from the stool, walk to the taped box, and return three times during each test.
Figure 1. Motion capture recording during timed-up-and-go testing. Left: clinical motion capture laboratory. Right: example of kinematic marker data. Participants were instructed to rise from the stool, walk to the taped box, and return three times during each test.
Sensors 23 01766 g001
Figure 2. Overall model architecture. The recorded motion capture sequence is segmented into 4 s analysis windows, which are first processed with the Adaptive Trimming (AT) model. AT model, which uses a 4-layer temporal convolutional network (TCN), predicts the start and end index of the core motion segment that is most relevant for the prediction task. The core motion segment is processed with a 4-layer adaptive temporal-spatial graphical convolutional network (AGCN), which automatically learns the attention map for the most relevant joint and limb motion for the prediction task. The feature representation from the final layer of AGCN is processed with temporal average pooling (TAP) to summarize temporal information, which is then used to predict medication state, FOG score, and MDS-UPDRS-III total score (excluding FOG score) at the same time. Specifically for regressing the MDS-UPDRS-III total score (excluding FOG score), Gaussian Mixture Model (GMM)-based regressor is used to take account of the non-Gaussian distribution of the target values.
Figure 2. Overall model architecture. The recorded motion capture sequence is segmented into 4 s analysis windows, which are first processed with the Adaptive Trimming (AT) model. AT model, which uses a 4-layer temporal convolutional network (TCN), predicts the start and end index of the core motion segment that is most relevant for the prediction task. The core motion segment is processed with a 4-layer adaptive temporal-spatial graphical convolutional network (AGCN), which automatically learns the attention map for the most relevant joint and limb motion for the prediction task. The feature representation from the final layer of AGCN is processed with temporal average pooling (TAP) to summarize temporal information, which is then used to predict medication state, FOG score, and MDS-UPDRS-III total score (excluding FOG score) at the same time. Specifically for regressing the MDS-UPDRS-III total score (excluding FOG score), Gaussian Mixture Model (GMM)-based regressor is used to take account of the non-Gaussian distribution of the target values.
Sensors 23 01766 g002
Figure 3. Scatter plot comparing clinician-rated versus model-estimated MDS-UPDRS-III total score, excluding the FOG item (3.11). Unity line is shown for reference. Each dot represents a single 4 s analysis window. Colors are used to represent the FOG item scores corresponding to each analysis window, with darker colors indicating more severe FOG in the OFF (blue) and ON (red) medication states.
Figure 3. Scatter plot comparing clinician-rated versus model-estimated MDS-UPDRS-III total score, excluding the FOG item (3.11). Unity line is shown for reference. Each dot represents a single 4 s analysis window. Colors are used to represent the FOG item scores corresponding to each analysis window, with darker colors indicating more severe FOG in the OFF (blue) and ON (red) medication states.
Sensors 23 01766 g003
Figure 4. Kinematic markers (referred to as “joints” in pose estimation literature, red) and segments (referred to as “limbs” in pose estimation literature, blue) with the top 10 attention weights in the prediction tasks.
Figure 4. Kinematic markers (referred to as “joints” in pose estimation literature, red) and segments (referred to as “limbs” in pose estimation literature, blue) with the top 10 attention weights in the prediction tasks.
Sensors 23 01766 g004
Figure 5. Examples of detected core motion segments of left heel motions from the adaptive trimming model for different medication states and FOG scores. The horizontal and vertical axes of plots are in seconds and millimeters, respectively, with the vertical axis indicating the height of the left heel marker above the laboratory floor. All plots depict four seconds of recorded movement. Core motion segments detected by the adaptive trimming model are depicted in blue.
Figure 5. Examples of detected core motion segments of left heel motions from the adaptive trimming model for different medication states and FOG scores. The horizontal and vertical axes of plots are in seconds and millimeters, respectively, with the vertical axis indicating the height of the left heel marker above the laboratory floor. All plots depict four seconds of recorded movement. Core motion segments detected by the adaptive trimming model are depicted in blue.
Sensors 23 01766 g005
Table 1. Clinical and demographic features of study participants.
Table 1. Clinical and demographic features of study participants.
PD-FOGPD-NoFOGPP-FOG
N35175
Age, y 69 ± 7 67 ± 12 66 ± 6
Sex, M/F30/511/62/3
Disease duration, y 10.5 ± 6.7 6.0 ± 3.6 6.0 ± 3.3
LED, mg 1429 ± 673 833 ± 303 1258 ± 640
MDS-UPDRS-III (OFF) 34.0 ± 10.6 30.8 ± 13.2 39.4 ± 7.8
MDS-UPDRS-III (ON) 20.7 ± 8.7 18.4 ± 14.5 31.6 ± 9.0
NFOG-Q 20.1 ± 4.9 0.0 ± 0.0 17.8 ± 7.5
Table 2. Summary of timed-up-and-go testing sessions stratified by medication state and FOG score.
Table 2. Summary of timed-up-and-go testing sessions stratified by medication state and FOG score.
MedicationFOG Score
State01234
OFF2115987
ON3811731
Table 3. Prediction performance of the proposed model (AT+AGCN+GMM) and comparison to baseline models. We also compare the run time of each deep learning model used in the experiment. Performance metrics are presented as mean ± 95% confidence interval. Deep learning models are indicated by italics. Abbreviations are described in text. a Total score with FOG item (3.11) subtracted. p < 0.05, improvement in RF vs. SVM. p < 0.05, improvement in -LR vs. -L. * p < 0.05, improvement in deep learning models vs. preceding row. § p < 0.05, improvement in multi-task vs. single task prediction.
Table 3. Prediction performance of the proposed model (AT+AGCN+GMM) and comparison to baseline models. We also compare the run time of each deep learning model used in the experiment. Performance metrics are presented as mean ± 95% confidence interval. Deep learning models are indicated by italics. Abbreviations are described in text. a Total score with FOG item (3.11) subtracted. p < 0.05, improvement in RF vs. SVM. p < 0.05, improvement in -LR vs. -L. * p < 0.05, improvement in deep learning models vs. preceding row. § p < 0.05, improvement in multi-task vs. single task prediction.
Medication StateFOG ScoreMDS-UPDRS-III a Inference Time
Model(F1)(F1)(RMSE) (Seconds)
SVM-L 0.540 ± 0.016 0.429 ± 0.026 9.346 ± 0.138 -
RF-L 0.594 ± 0.012 0.553 ± 0.038 9.189 ± 0.301 -
SVM-LR 0.616 ± 0.017 0.608 ± 0.031 8.714 ± 0.101 -
RF-LR 0.657 ± 0.019 , 0.684 ± 0.040 , 7.918 ± 0.427 , -
TCN [52] 0.875 ± 0.017  * 0.851 ± 0.020  * , § 4.551 ± 0.276  * 0.0055
GCN [53] 0.913 ± 0.015  * 0.929 ± 0.021  * , § 4.023 ± 0.373  * 0.0433
AGCN [34] 0.949 ± 0.010  * 0.948 ± 0.018  * , § 3.703 ± 0.300  * 0.0435
AT+AGCN 0.955 ± 0.021 0.955 ± 0.026 § 3.555 ± 0.394  * 0.0469
AT+AGCN+GMM 0.975 ± 0.018 0.967 ± 0.022 2.753 ± 0.440 * 0.0471
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kwon, H.; Clifford, G.D.; Genias, I.; Bernhard, D.; Esper, C.D.; Factor, S.A.; McKay, J.L. An Explainable Spatial-Temporal Graphical Convolutional Network to Score Freezing of Gait in Parkinsonian Patients. Sensors 2023, 23, 1766. https://doi.org/10.3390/s23041766

AMA Style

Kwon H, Clifford GD, Genias I, Bernhard D, Esper CD, Factor SA, McKay JL. An Explainable Spatial-Temporal Graphical Convolutional Network to Score Freezing of Gait in Parkinsonian Patients. Sensors. 2023; 23(4):1766. https://doi.org/10.3390/s23041766

Chicago/Turabian Style

Kwon, Hyeokhyen, Gari D. Clifford, Imari Genias, Doug Bernhard, Christine D. Esper, Stewart A. Factor, and J. Lucas McKay. 2023. "An Explainable Spatial-Temporal Graphical Convolutional Network to Score Freezing of Gait in Parkinsonian Patients" Sensors 23, no. 4: 1766. https://doi.org/10.3390/s23041766

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop