Next Article in Journal
Stereo Vision for Plant Detection in Dense Scenes
Previous Article in Journal
Estimating Compressive and Shear Forces at L5-S1: Exploring the Effects of Load Weight, Asymmetry, and Height Using Optical and Inertial Motion Capture Systems
Previous Article in Special Issue
Virtual Scenarios of Earthquake Early Warning to Disaster Management in Smart Cities Based on Auxiliary Classifier Generative Adversarial Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

From Movements to Metrics: Evaluating Explainable AI Methods in Skeleton-Based Human Activity Recognition

1
Department of Neuromedicine and Movement Science, Faculty of Medicine and Health Sciences, Norwegian University of Science and Technology, 7034 Trondheim, Norway
2
Department of Computer Science, Faculty of Information Technology and Electrical Engineering, Norwegian University of Science and Technology, 7034 Trondheim, Norway
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(6), 1940; https://doi.org/10.3390/s24061940
Submission received: 17 January 2024 / Revised: 7 March 2024 / Accepted: 8 March 2024 / Published: 18 March 2024

Abstract

:
The advancement of deep learning in human activity recognition (HAR) using 3D skeleton data is critical for applications in healthcare, security, sports, and human–computer interaction. This paper tackles a well-known gap in the field, which is the lack of testing in the applicability and reliability of XAI evaluation metrics in the skeleton-based HAR domain. We have tested established XAI metrics, namely faithfulness and stability on Class Activation Mapping (CAM) and Gradient-weighted Class Activation Mapping (Grad-CAM) to address this problem. This study introduces a perturbation method that produces variations within the error tolerance of motion sensor tracking, ensuring the resultant skeletal data points remain within the plausible output range of human movement as captured by the tracking device. We used the NTU RGB+D 60 dataset and the EfficientGCN architecture for HAR model training and testing. The evaluation involved systematically perturbing the 3D skeleton data by applying controlled displacements at different magnitudes to assess the impact on XAI metric performance across multiple action classes. Our findings reveal that faithfulness may not consistently serve as a reliable metric across all classes for the EfficientGCN model, indicating its limited applicability in certain contexts. In contrast, stability proves to be a more robust metric, showing dependability across different perturbation magnitudes. Additionally, CAM and Grad-CAM yielded almost identical explanations, leading to closely similar metric outcomes. This suggests a need for the exploration of additional metrics and the application of more diverse XAI methods to broaden the understanding and effectiveness of XAI in skeleton-based HAR.

1. Introduction

Analyzing human movement through 3D skeleton data has promising nontrivial applications in high-stake sectors such as healthcare and rehabilitation [1], security and surveillance [2], sports and athletics [3], and human–computer interaction [4]. Because of this, integrating deep learning in skeleton data analysis requires an understanding of the model’s decision-making processes. One particular application is compliance with the EU’s proposed AI Act, which emphasizes that transparency and human oversight should be embedded in high-risk applications such as AI-assisted medical diagnostics [5]. A basic form of deep learning technique applied to human movement analysis using skeleton data is human activity recognition (HAR). State-of-the-art HAR models are continually being developed and improved. It started with the introduction of Spatial–Temporal Graph Convolutional Network (ST-GCN) architectures in 2018 [6], as they were the first to use graph convolution for HAR. ST-GCN then became the baseline for dozens of emerging skeleton-based HAR models that seek to improve this original implementation.
Recent advancements in HAR model architectures have been significant, but strides in their explainability remain limited. Class Activation Mapping (CAM) was used in EfficientGCN [7] and ST-GCN [8] to highlight the body points significant for specific actions. In [9], Gradient-weighted Class Activation Mapping (Grad-CAM) was implemented in ST-GCN. There is a growing trend towards using explainable AI (XAI) methods, extending from CNNs to ST-GCNs, yet XAI metrics to assess their reliability in this domain have yet to be tested. There is also a lack of comparative analysis between these methods, which leaves a gap in understanding their relative performance in HAR applications. Additionally, research is limited regarding metrics that assess XAI methods in the context of data perturbation.
While the paper in [9] evaluated the faithfulness and contrastivity of Grad-CAM, it did not offer insights into its performance relative to other XAI methods. Moreover, their choice of using masking/occlusion to check for changes in prediction output raises concerns. Masking can potentially distort the standard human skeletal structure that GCN-based models are trained on. Movements of the human body are governed by biomechanical principles, and perturbations that do not respect these principles can potentially result in a misleading understanding of the model’s faithfulness. Recognizing the growing relevance of skeleton-based HAR models in critical areas, our paper aims to test established metrics that assess their corresponding feature attribution techniques. Additionally, this study pioneers the evaluation of explainability metrics within the context of perturbations that fall within the error tolerance of depth-sensing technology, such as that of the Kinect device, effectively simulating realistic variations in skeletal data. In [10,11], the perturbation of skeleton data was employed, which they claim maintained normal human kinematics during perturbation. However, the objective was for an adversarial attack; thus, the perturbed skeleton joints were neither controlled nor deliberately targeted. In essence, our work also addresses the gap in evaluation metrics by leveraging targeted perturbations that align with the Kinect device’s error tolerance, providing a practical approach to simulate realistic skeletal data variations.
Alongside the pursuit of improved explainability, assessing the stability of model decisions and their explanations is important. As human skeletal data can exhibit subtle variances due to minor changes in posture, movement, or data-capturing techniques, the decisions from the model and the explanations from the XAI methods should remain consistent and trustworthy. That is, dramatic shifts in explanations due to minor input changes cast doubt on model reliability. Moreover, the imprecise estimation of joint center positions in 3D skeletal data analysis underscores the need to evaluate decision and explanation robustness using perturbations that remain within the realistic operational bounds of the capturing device’s tracking capabilities. To address this, we draw from metrics established for other data types. In this work, we focus on the two primary metrics highlighted in [12]: faithfulness, which gauges how closely an explanation mirrors the model’s internal logic, and stability, which pertains to the consistency of a model’s explanations across similar inputs.
This paper’s key contributions are as follows:
  • Testing established metrics and assessing their applicability for evaluating XAI methods in skeleton-based HAR.
  • Introducing a controlled approach to perturb 3D skeleton data for XAI evaluation that accommodates realistic variations within the inherent inaccuracies of the tracking device.
  • Assessing the impact of perturbation magnitude variations on metrics.
  • Comparing the performance of CAM, Grad-CAM, and a random feature attribution method for HAR.

2. Materials

To provide the framework for this research, the dataset used, the neural network architecture trained and tested, and the XAI metrics implemented are briefly summarized below.

2.1. NTU RGB+D 60 Dataset and EfficientGCN

The NTU RGB+D 60 dataset [13] contains 60 action classes with over 56 thousand 3D skeleton data, each composed of sequential frames captured from 40 different subjects using the Kinect v2 camera with depth sensor. For evaluation purposes, the dataset is further divided into cross-subject and cross-view subgroups—the former is composed of different human subjects performing the same actions, while the latter uses different camera angle views.
The EfficientGCN architecture [14] is a result of extending the concept of EfficientNet [15] for CNNs to ST-GCNs to reduce the computing resource demand for HAR. There are a total of 24 different EfficientGCN network configurations with different scaling coefficients that the user can choose and test. In this paper, we use the B4 configuration, which has achieved the highest accuracy at 92.1% on the cross-subject dataset and 96.1% on cross-view, compared with the 81.5% and 88.3% of the baseline ST-GCN, respectively.

2.2. Evaluation Metrics

2.2.1. Faithfulness

In XAI, fidelity or faithfulness, as described in [16,17,18], measures how well an explanation shows what truly influences a model’s decisions, focusing on the importance of different features. This concept checks if the explanation accurately matches the actual effect of these features on the model’s predictions, providing a crucial way to judge if explanations are trustworthy. Both the impact of important and unimportant features, as identified by the XAI method, can be quantitatively assessed for their accuracy in reflecting the model’s decision-making process, with the mathematical framework for this assessment provided in [12]. The Prediction Gap on Important feature perturbation (PGI) measures how much prediction changes when top-k features are perturbed. Conversely, the Prediction Gap on Unimportant feature perturbation (PGU) measures the change in prediction when unimportant features are perturbed. Let X represent the original input data with its associated explanation e X and f ( · ) represent the output probability. Then, X signifies the perturbed variant of X and e X the revised explanation after perturbation.
P G I ( X ,   f ,   e X ,   k ) = E X perturb ( X , e X , top - k ) [ | f ( X ) f ( X ) | ]
P G U ( X ,   f ,   e X ,   k ) = E X perturb ( X , e X , non   top - k ) [ | f ( X ) f ( X ) | ]

2.2.2. Stability

The concept of stability, also known as robustness, refers to the maximum amount of change in explanation (i.e., attribution scores) when the input data are slightly perturbed, as defined in [19,20]. The idea is that when original data are slightly perturbed, the resulting explanation should not drastically change, ensuring the model’s interpretations are stable and reliable across minor variations.. There are three submetrics that can be calculated, as enumerated in [12].
Relative Input Stability (RIS) measures the maximum change in attribution scores with respect to a corresponding perturbation in the input. Given that EfficientGCN has multiple input branches, it is essential to compute the RIS for each branch namely joint, velocity, and bone. From hereon, they are referred to as RISj, RISv, and RISb, respectively.
RIS X , X , e X , e X = max X e X e X e X p max X X X p , ϵ min , X s . t . X N X ; f ( X ) = f ( X )
Relative Output Stability (ROS) measures the maximum ratio of how much the explanation changes to how much the model’s prediction probability changes due to small perturbations in the input data.
ROS X , X , e X , e X = max X e X e X e X p max f ( X ) f X f ( X ) p , ϵ min X s . t . X N X ; f ( X ) = f ( X )
Relative Representation Stability (RRS) measures the maximum change in a model’s explanations relative to changes in the model’s internal representations brought about by small input perturbations. In this context, the internal representation denoted as L X typically refers to an intermediate layer’s output in a neural network, capturing the model’s understanding of the input data. In our experiment, we extract and use the logits from the layer preceding the softmax function for our computations.
RRS X , X , e X , e X = max X e X e X e X p max L X L X L X p , ϵ min X s . t . X N X ; f ( X ) = f ( X )

3. Methods

3.1. Skeleton Data Perturbation

In 3D space, skeleton joints can be perturbed using spherical coordinates by generating random angles θ and ϕ for perturbation direction, sourced from a Gaussian distribution. The magnitude of this perturbation is controlled by radius r. In a standard XAI metric test, we recommend adhering to the principle that X should be within the neighborhood of X to ensure that inputs remain representative of human kinematics and avoid skewing model predictions. This means constraining the magnitude of r, which in our pipeline is initially set to 2.5 cm. When it comes to body point tracking, the Kinect v2’s tracking error ranges from 1 to 7 cm compared with the gold-standard Vicon system [21], so a 2.5 cm perturbation ensures the perturbed data stay within Kinect’s typical accuracy tolerance. However, we also tested the metrics with increasing r (in cm: 2.5, 5, 10, 20, 40, and 80). While this contradicts our initial recommendation, varying the perturbation magnitude would allow us to (a) test the hypothesis that a small perturbation should result in meaningful changes in the prediction, which should be reflected in faithfulness results, and (b) observe its effects on the explanations, which should be reflected in the stability results.
The point P′(x′, y′, z′) in Figure 1 can be calculated using the equations below, which are used to convert a point from spherical to Cartesian (rectangular) coordinates. In these equations, r represents the distance from the two points, θ denotes the azimuthal angle, and ϕ is the polar angle. A fixed random seed was used to generate reproducible angles θ and ϕ . The variables d x , d y , and d z are computed once, and each joint is given its own unique set of these values. When added to the original coordinates across all video frames, a mildly perturbed 3D point is produced. This method ensures that a particular joint undergoes the same random adjustment across all frames, rather than different perturbations in each frame.
x = x + d x , d x = r sin ( ϕ ) cos ( θ ) y = y + d y , d y = r sin ( ϕ ) sin ( θ ) z = z + d z , d z = r cos ( ϕ )

3.2. Calculation and Evaluation of XAI Metrics

To calculate the metric values for a given action class, we employ the Area Under the Curve (AUC) to aggregate results into a single measure. Our process involves initializing variables and perturbing each data instance to n = 50 times at set perturbation magnitudes. For each data instance, we input the original skeleton data into the model, capturing the explanation as detailed in Equations (6) and (7) for CAM and Grad-CAM, respectively. The model output and other necessary variables are also obtained as shown in the EfficientGCN pipeline [14] in Figure 2. To ensure metrics comparability across XAI methods, we normalize attribution scores between 0 and 1, and then rank features from highest to lowest based on the average score across all video frames. From the definition of CAM [22], w in Equation (6) are the weights after Global Average Pooling (GAP) for the specific output c l a s s , and F n denotes the nth feature map. Similarly, α in Equation (7) for Grad-CAM [23] is calculated by averaging the gradients. Figure 3 shows sample CAM and Grad-CAM explanations in comparison with the random baseline. More samples can be found in Appendix B.
e X , CAM = n w n c l a s s F n
e X , Grad - CAM = n α n c l a s s F n
Next, we systematically perturb the top-k body points (k = 1 to 25) n times, calculating the PGI as per Equation (1), and similarly perturb the remaining points to compute the PGU using Equation (2). Stability metrics are derived from the original and new explanations using Equations (3)–(5). Finally, we calculate the AUC for each metric across all k values for each data instance. The mean and standard deviation of these AUCs across all instances in a class provide the overall metric values.
Since the metrics are unitless, a random method serves as the benchmark for the least desirable outcome by randomly assigning feature attribution scores. Higher PGI values are optimal, indicating that altering important skeletal nodes has a marked impact on the prediction. Lower PGU values are better, suggesting that perturbing the identified unimportant skeletal nodes does not cause a significant change in the model’s output prediction. Lastly, a stability (RIS, RRS, and ROS) closer to zero is indicative of a model’s robustness, signifying that minor perturbations to the input data do not lead to significant changes in the explanation. In order to thoroughly assess the applicability and consistency of the XAI evaluation metrics, we test them on both the most accurately classified class (class 26—‘jump up’) and the class with the highest misclassification rate (class 11—‘writing’) in the NTU dataset. Of the 276 samples in the class 26 test set, only 1 sample was misclassified by the EfficientGCN-B4 model, while only 174 were correctly classified out of 272 samples in class 11. With 4 parallel Tesla V100 GPUs in our hardware setup, the total test time to generate one table in the Appendix A section was about 165 h.
To assess the generalizability of our findings further, we also extended our evaluation to include the metrics of Class 32 (representing the action ‘checking time on watch’) and Class 9 (representing the action ‘clapping’). For comparison, the model accurately predicted 253 out of 276 test instances for Class 32, and 221 out of 273 test instances for Class 9. Detailed graphical representations of these results, along with their corresponding numerical values, are presented in Appendix A Table A3 and Table A4, respectively.

4. Results

To help us gauge the reliability of the XAI evaluation metrics, we tested them by slowly increasing the perturbation magnitude, as described in the Section 3. Figure 4 and Figure 5 show the line graphs for each metric test on class 11 and 26, respectively, comparing the different XAI methods. From hereon, we refer to class 26 as the strongest class, while class 11 is referred to as the weakest class. For the exact numerical values of the results with confidence intervals, please refer to Appendix A Table A1 and Table A2.

4.1. Faithfulness

The identical PGI and PGU values for CAM and Grad-CAM mean both methods have the exact same ranking of features, although the numerical attribution scores are not exactly the same. Unexpectedly, the random method appears to outperform both in PGI in the weakest class, except at r = 80 cm. Conversely, looking at the results for the strongest class in Figure 5 suggests equal PGI performance among all methods up to r 5 cm, beyond which the random method seems to surpass the others. In essence, in a class where the model has the best classification performance, the PGI test aligns with expected outcomes only at higher values of r, where data distortion is significant, which is no longer consistent with the rule that X N X . Moreover, where model classification is the weakest, PGI results consistently give unexpected outcomes. In class 32 (results shown in Figure A1), which has a classification accuracy of 91.67%, both CAM and Grad-CAM outperform the random method in terms of PGI, albeit with a small margin. Conversely, in class 9 (results shown in Figure A2), with a classification accuracy of 80.95%, CAM and Grad-CAM also outperform the random method across all perturbation magnitudes, except at r = 80 cm.
An analysis of the PGU results of the weakest class indicates that CAM and Grad-CAM outperform the random method, as expected. In the strongest class, however, conformity to expected outcomes occurs only when r 40 cm, with random exhibiting higher values, while in lower perturbation magnitudes, the results seem to suggest that the three methods exhibit either comparable performance or that the random method has marginally higher PGU values. Therefore, PGU tests only meet expected outcomes primarily when there is weak model performance or when input perturbation is significant during strong model performance. For classes 32 and 9, PGU scores of both CAM and Grad-CAM surpass the random method, as anticipated, with increasing margins as perturbation magnitude increases.
These irregularities in both PGI and PGU suggest that the hypothesis that faithfulness is anchored on—minor perturbations causing meaningful prediction shifts—is not upheld in all action classes of the NTU dataset for the EfficientGCN model. Using output predictions for gauging explanation fidelity proves unreliable in this context.

4.2. Stability

Stability assessments for CAM and Grad-CAM yield nearly identical values, diverging only in less significant decimal places. This implies that despite both methods giving different raw scores, they tend to converge upon normalization. Stability test results, contrary to faithfulness, demonstrate robustness against increased perturbation, consistently indicating the superiority of CAM and Grad-CAM over random in the four classes tested. Thus, stability testing affirms that input perturbations do not drastically alter explanations compared with the random baseline.
It can be observed in Figure 4f and Figure 5f that ROS results register very high numerical values, with the y-axis scaled to 1 × 10 7 . We inspected the individual terms in each of the ROS results and found that the cause for such high numbers is the extremely small denominator terms (typically less than 1). Since the denominator term of ROS is the difference between the original and perturbed predictions, it means that the change in the model’s output probability is very small, even when the perturbation magnitude is large. A small denominator, reflecting little changes in output probabilities even with substantial perturbations, corroborates the inefficacy of the PGI and PGU tests in our context. These tests, which are reliant on shifts in prediction probabilities, fail to yield meaningful results in response to input perturbations, further supporting our hypothesis regarding the model’s behavior under examination.

5. Discussion and Conclusions

Our research contributes to the understanding of explainable AI in the context of skeleton-based HAR, advancing the state-of-the-art by testing known metrics in this emerging domain and introducing a perturbation technique informed by the practical constraints of skeletal tracking technology. A key finding from our experiments is that faithfulness—a widely recognized XAI metric—may falter in certain models, such as the EfficientGCN-B4 tested in this study. This finding serves as a caution to XAI practitioners when using an XAI metric that measures the reliability of XAI methods indirectly through the change in the model’s prediction probability. In contrast, stability, which measures direct changes in explanations, emerged as a dependable metric. However, this leaves us with only a single metric, which offers a limited view of an XAI method’s efficacy, underscoring the need for developing or adapting additional testing approaches in this field. Our skeleton perturbation method, which simulates realistic variations within the operational tolerances of skeletal tracking technologies, offers a promising framework for validating upcoming XAI metrics.
This study also identifies other gaps in XAI for ST-GCN-based HAR, which is an opportunity for future research directions. The nearly identical explanations produced by Grad-CAM and CAM when applied to EfficientGCN highlight a need for more diverse XAI techniques, such as adaptations of model agnostic methods like LIME [24] and SHAP [25] for this specific domain. Additionally, comparative studies of XAI metrics across different existing HAR models can also be explored, which could be valuable as a guide for model selection where explainability is as important as accuracy.
Lastly, our comparative analysis between CAM and Grad-CAM revealing negligible stability differences suggests that neither method is superior; they are essentially equivalent. Yet, CAM’s use of static model weights obtained post training means it demands less computational load compared with Grad-CAM, which needs gradient computation per data instance. This highlights CAM’s suitability for large-scale data analysis. This consideration is especially pertinent for applications where computational efficiency is vital alongside accuracy and reliability.

Author Contributions

Conceptualization, all; methodology, all; software, K.N.P.; formal analysis, all; investigation, all; writing—original draft preparation, K.N.P.; writing—review and editing, all; visualization, K.N.P.; supervision, I.S. and E.A.F.I.; project administration, I.S. and E.A.F.I.; funding acquisition, I.S. and E.A.F.I. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by grant 327146 from the Research Council of Norway and the Horizon 2020 project PERSEUS grant 101034240 of the European Union.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in the paper is openly accessible and does not require any authorization or permission: NTU-RGB+D 60 (https://rose1.ntu.edu.sg/dataset/actionRecognition/ accessed on 24 October 2022).

Acknowledgments

We would like to thank the Norwegian Open AI Lab for the additional computational resources provided to perform the experiments.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Class 11 tabular data: ↑ indicates that higher values are better, while ↓ indicates that lower values are optimal.
Table A1. Class 11 tabular data: ↑ indicates that higher values are better, while ↓ indicates that lower values are optimal.
Methodr (cm)PGI ↑PGU ↓RISjRISvRISbROSRRS
CAM2.52.080 ± 0.3941.507 ± 0.268174.133 ± 16.9461288.370 ± 138.91214.424 ± 1.2021,702,553.834 ± 1,116,223.129285.874 ± 46.673
53.142 ± 0.5142.547 ± 0.394137.460 ± 11.303968.090 ± 86.85512.924 ± 0.9642,530,301.100 ± 2,226,508.985264.838 ± 33.044
104.109 ± 0.5693.807 ± 0.48991.225 ± 6.774654.886 ± 53.71110.877 ± 0.741971,050.496 ± 676,047.094222.611 ± 20.795
205.722 ± 0.6165.294 ± 0.54152.854 ± 3.473380.917 ± 30.1899.210 ± 0.556355,153.157 ± 272,318.174180.486 ± 11.691
4010.859 ± 0.8117.826 ± 0.54928.373 ± 2.139190.976 ± 18.5487.112 ± 0.416334,731.175 ± 221,923.029132.782 ± 8.488
8015.059 ± 1.02310.787 ± 0.67912.199 ± 1.40876.734 ± 10.9113.764 ± 0.330340,629.702 ± 291,218.22292.569 ± 8.165
Grad-CAM2.52.080 ± 0.3941.507 ± 0.268174.133 ± 16.9461288.371 ± 138.91214.424 ± 1.2021,702,554.201 ± 1,116,223.411285.874 ± 46.673
53.142 ± 0.5142.547 ± 0.394137.460 ± 11.303968.090 ± 86.85512.924 ± 0.9642,530,301.513 ± 2,226,509.277264.838 ± 33.044
104.109 ± 0.5693.807 ± 0.48991.225 ± 6.774654.886 ± 53.71110.877 ± 0.741971,050.337 ± 676,047.046222.611 ± 20.795
205.722 ± 0.6165.294 ± 0.54152.854 ± 3.473380.917 ± 30.1899.210 ± 0.556355,153.184 ± 272,318.220180.486 ± 11.691
4010.859 ± 0.8117.826 ± 0.54928.373 ± 2.139190.976 ± 18.5487.112 ± 0.416334,731.302 ± 221,923.109132.782 ± 8.488
8015.059 ± 1.02310.787 ± 0.67912.199 ± 1.40876.734 ± 10.9113.764 ± 0.330340,629.723 ± 291,218.25292.569 ± 8.165
Random2.52.190 ± 0.4101.831 ± 0.3061228.105 ± 43.4419627.812 ± 502.59195.762 ± 3.63020,728,354.210 ± 14,139,843.2943205.744 ± 333.826
53.559 ± 0.5773.165 ± 0.445598.011 ± 21.1234555.435 ± 238.30950.612 ± 1.84313,986,453.073 ± 9,801,817.5261617.393 ± 160.022
104.613 ± 0.6644.905 ± 0.542300.288 ± 11.3032300.006 ± 112.75931.848 ± 1.1203,835,957.191 ± 2,481,844.039940.518 ± 87.484
206.173 ± 0.6436.357 ± 0.596144.791 ± 5.9821069.703 ± 56.38223.135 ± 0.9761,563,644.666 ± 1,072,148.002580.579 ± 46.894
4012.232 ± 0.8209.155 ± 0.65467.294 ± 3.949386.528 ± 35.89514.813 ± 0.7681,187,301.059 ± 987,974.000350.415 ± 26.835
8014.989 ± 1.00813.253 ± 0.85227.610 ± 2.828115.425 ± 19.1286.621 ± 0.418978,432.004 ± 1,211,946.037266.365 ± 27.954
Table A2. Class 26 tabular data: ↑ indicates that higher values are better, while ↓ indicates that lower values are optimal.
Table A2. Class 26 tabular data: ↑ indicates that higher values are better, while ↓ indicates that lower values are optimal.
Methodr (cm)PGI ↑PGU ↓RISjRISvRISbROSRRS
CAM2.50.002 ± 0.0010.002 ± 0.00138.783 ± 1.685241.329 ± 10.0333.504 ± 0.1123,793,568.715 ± 599,734.044137.439 ± 4.602
50.003 ± 0.0010.003 ± 0.00135.563 ± 1.469226.272 ± 8.8213.447 ± 0.1034,244,155.156 ± 614,303.198130.751 ± 3.800
100.005 ± 0.0020.006 ± 0.00329.652 ± 1.165191.308 ± 7.2243.452 ± 0.1104,307,133.768 ± 667,017.130116.767 ± 3.276
200.010 ± 0.0050.010 ± 0.00422.343 ± 0.782140.125 ± 5.0273.698 ± 0.1083,926,629.242 ± 573,923.634114.482 ± 3.087
400.051 ± 0.0400.036 ± 0.01415.016 ± 0.48191.628 ± 3.5433.793 ± 0.1112,119,077.311 ± 355,383.073119.608 ± 2.883
802.281 ± 0.2990.938 ± 0.14011.679 ± 0.34745.246 ± 2.8273.324 ± 0.090339,790.687 ± 75,217.805110.699 ± 2.141
Grad-CAM2.50.002 ± 0.0010.002 ± 0.00138.783 ± 1.685241.329 ± 10.0333.504 ± 0.1123,793,569.048 ± 599,734.059137.439 ± 4.602
50.003 ± 0.0010.003 ± 0.00135.563 ± 1.469226.272 ± 8.8213.447 ± 0.1034,244,156.522 ± 614,303.356130.751 ± 3.800
100.005 ± 0.0020.006 ± 0.00329.652 ± 1.165191.308 ± 7.2243.452 ± 0.1104,307,133.975 ± 667,017.145116.767 ± 3.276
200.010 ± 0.0050.010 ± 0.00422.343 ± 0.782140.125 ± 5.0273.698 ± 0.1083,926,629.059 ± 573,923.586114.482 ± 3.087
400.051 ± 0.0400.036 ± 0.01415.016 ± 0.48191.628 ± 3.5433.793 ± 0.1112,119,077.265 ± 355,383.079119.608 ± 2.883
802.281 ± 0.2990.938 ± 0.14011.679 ± 0.34745.246 ± 2.8273.324 ± 0.090339,790.661 ± 75,217.791110.699 ± 2.141
Random2.50.002 ± 0.0010.002 ± 0.0011233.418 ± 22.3767891.926 ± 206.38697.067 ± 1.61693,264,563.501 ± 12,716,437.7995008.974 ± 133.687
50.003 ± 0.0020.003 ± 0.001619.974 ± 10.8673948.918 ± 101.99854.711 ± 0.72763,286,627.152 ± 9,322,027.6322557.063 ± 64.789
100.006 ± 0.0020.005 ± 0.002312.310 ± 5.3321958.503 ± 50.63835.273 ± 0.33342,392,862.818 ± 6,446,239.4341393.590 ± 32.140
200.012 ± 0.0050.010 ± 0.004156.615 ± 2.673966.698 ± 25.44526.557 ± 0.19329,527,933.778 ± 4,113,487.016864.778 ± 18.918
400.080 ± 0.0550.060 ± 0.02777.913 ± 1.346454.485 ± 13.90319.684 ± 0.2419,244,726.238 ± 1,352,133.114627.195 ± 13.200
805.474 ± 0.3341.149 ± 0.17635.110 ± 0.858149.132 ± 7.4048.684 ± 0.1452,129,144.307 ± 407,137.423419.232 ± 8.483
Table A3. Class 32 tabular data: ↑ indicates that higher values are better, while ↓ indicates that lower values are optimal.
Table A3. Class 32 tabular data: ↑ indicates that higher values are better, while ↓ indicates that lower values are optimal.
Methodr (cm)PGI ↑PGU ↓RISjRISvRISbROSRRS
CAM2.50.530 ± 0.1720.275 ± 0.113104.752 ± 12.666567.154 ± 63.4817.991 ± 1.0362,916,268.239 ± 1,006,355.218150.627 ± 22.877
51.923 ± 0.4660.469 ± 0.140100.686 ± 9.080547.544 ± 50.2588.872 ± 0.8312,155,297.199 ± 768,443.139136.450 ± 18.700
106.421 ± 0.7931.169 ± 0.22989.387 ± 6.276492.629 ± 39.84710.145 ± 0.6451,787,187.499 ± 738,937.270122.012 ± 13.391
2013.318 ± 0.7652.366 ± 0.34466.464 ± 5.372352.318 ± 34.57010.860 ± 0.728836,212.856 ± 204,359.147127.161 ± 9.837
4018.903 ± 0.6794.338 ± 0.43827.222 ± 2.617124.864 ± 14.6606.080 ± 0.462628,225.894 ± 157,229.139100.372 ± 6.547
8021.380 ± 0.63710.536 ± 0.56713.050 ± 1.49653.424 ± 8.6083.111 ± 0.243314,147.370 ± 133,974.31077.798 ± 5.265
Grad-CAM2.50.530 ± 0.1720.275 ± 0.113104.752 ± 12.666567.154 ± 63.4817.991 ± 1.0362,916,267.512 ± 1,006,355.183150.627 ± 22.877
51.923 ± 0.4660.469 ± 0.140100.686 ± 9.080547.544 ± 50.2588.872 ± 0.8312,155,296.259 ± 768,443.051136.450 ± 18.700
106.421 ± 0.7931.169 ± 0.22989.387 ± 6.276492.629 ± 39.84710.145 ± 0.6451,787,187.829 ± 738,937.896122.012 ± 13.391
2013.318 ± 0.7652.366 ± 0.34466.464 ± 5.372352.318 ± 34.57010.860 ± 0.728836,212.707 ± 204,359.110127.161 ± 9.837
4018.903 ± 0.6794.338 ± 0.43827.222 ± 2.617124.864 ± 14.6606.080 ± 0.462628,225.943 ± 157,229.158100.372 ± 6.547
8021.380 ± 0.63710.536 ± 0.56713.050 ± 1.49653.424 ± 8.6083.111 ± 0.243314,147.317 ± 133,974.30577.798 ± 5.265
Random2.50.516 ± 0.1690.402 ± 0.1401171.479 ± 26.5227546.971 ± 238.18192.114 ± 2.27463,030,515.690 ± 17,269,283.2382668.959 ± 142.801
51.318 ± 0.3330.928 ± 0.225580.844 ± 13.8553727.449 ± 119.08251.361 ± 1.15035,297,240.144 ± 10,033,888.5661360.739 ± 71.238
104.401 ± 0.6252.923 ± 0.430286.405 ± 6.9341781.205 ± 57.35632.552 ± 0.64011,237,060.760 ± 3,678,599.967688.543 ± 34.206
2011.434 ± 0.7446.970 ± 0.681139.891 ± 3.732819.171 ± 35.88823.906 ± 0.5092,949,464.762 ± 1,189,151.206383.922 ± 18.093
4017.962 ± 0.63112.742 ± 0.78966.049 ± 2.985381.788 ± 25.71816.294 ± 0.7181,293,045.700 ± 455,084.066260.787 ± 17.520
8020.176 ± 0.61019.286 ± 0.59222.288 ± 2.462147.523 ± 15.5266.688 ± 0.579978,000.924 ± 471,660.208191.267 ± 22.531
Table A4. Class 9 tabular data: ↑ indicates that higher values are better, while ↓ indicates that lower values are optimal.
Table A4. Class 9 tabular data: ↑ indicates that higher values are better, while ↓ indicates that lower values are optimal.
Methodr (cm)PGI ↑PGU ↓RISjRISvRISbROSRRS
CAM2.51.446 ± 0.2990.834 ± 0.156139.642 ± 9.088859.375 ± 59.02510.975 ± 0.7492,813,109.184 ± 1,254,700.809152.976 ± 8.557
52.985 ± 0.4991.811 ± 0.280112.662 ± 6.406693.715 ± 42.5279.969 ± 0.5451,667,239.107 ± 838,594.507133.983 ± 6.921
106.648 ± 0.7274.144 ± 0.45175.351 ± 4.038453.420 ± 27.1088.299 ± 0.409825,855.283 ± 303,557.804107.961 ± 5.679
2010.697 ± 0.7866.490 ± 0.49446.355 ± 2.759260.949 ± 18.1627.391 ± 0.375518,198.588 ± 209,846.95694.920 ± 5.303
4013.535 ± 0.7038.259 ± 0.50824.799 ± 1.502127.559 ± 9.3715.708 ± 0.299516,925.509 ± 228,941.57686.767 ± 4.868
8014.963 ± 0.6619.610 ± 0.48811.064 ± 1.07557.528 ± 6.9813.150 ± 0.150611,673.611 ± 310,288.37986.007 ± 4.309
Grad-CAM2.51.446 ± 0.2990.834 ± 0.156139.642 ± 9.088859.375 ± 59.02510.975 ± 0.7492,813,108.561 ± 1,254,700.783152.976 ± 8.557
52.985 ± 0.4991.811 ± 0.280112.662 ± 6.406693.715 ± 42.5279.969 ± 0.5451,667,238.830 ± 838,594.559133.983 ± 6.921
106.648 ± 0.7274.144 ± 0.45175.351 ± 4.038453.420 ± 27.1088.299 ± 0.409825,854.822 ± 303,557.678107.961 ± 5.679
2010.697 ± 0.7866.490 ± 0.49446.355 ± 2.759260.949 ± 18.1627.391 ± 0.375518,198.555 ± 209,846.91694.920 ± 5.303
4013.535 ± 0.7038.259 ± 0.50824.799 ± 1.502127.559 ± 9.3715.708 ± 0.299516,925.478 ± 228,941.53986.767 ± 4.868
8014.963 ± 0.6619.610 ± 0.48811.064 ± 1.07557.528 ± 6.9813.150 ± 0.150611,673.499 ± 310,288.28786.007 ± 4.309
Random2.51.279 ± 0.2651.340 ± 0.2511154.662 ± 32.5207879.336 ± 286.30893.075 ± 2.39648,749,279.991 ± 19,674,016.0582113.066 ± 106.944
52.638 ± 0.4263.309 ± 0.493559.807 ± 17.1763827.702 ± 145.16150.903 ± 1.29719,522,191.192 ± 8,735,689.8391025.301 ± 50.038
106.366 ± 0.6346.861 ± 0.695275.334 ± 9.4211830.920 ± 78.62232.519 ± 1.4057,999,255.135 ± 3,838,244.128565.638 ± 31.075
209.807 ± 0.7229.956 ± 0.719134.102 ± 5.852853.467 ± 42.65523.369 ± 1.1222,254,755.568 ± 1,106,476.034362.369 ± 26.496
4012.680 ± 0.70312.389 ± 0.69169.849 ± 3.386418.035 ± 24.77516.281 ± 0.5751,252,448.556 ± 804,210.707272.395 ± 19.120
8016.093 ± 0.71013.919 ± 0.64025.073 ± 1.905124.852 ± 12.6486.706 ± 0.273808,021.756 ± 771,280.581205.893 ± 14.800
Figure A1. Evaluation metric outcomes for ‘checking time on watch’ (Class 32), showing CAM (blue), Grad-CAM (orange), and the random (green) methods for (a) PGI, (b) PGU, (c) RISb, (d) RISj, (e) RISv, (f) ROS, and (g) RRS. Despite increasing perturbation magnitudes, CAM and Grad-CAM exhibit only marginally better performance compared with the random method in terms of PGI. This suggests that the expected correlation between increasing the perturbation magnitude of important features and significant changes in prediction output may not consistently apply to this particular case. Conversely, for PGU, CAM and Grad-CAM demonstrate more effective identification of unimportant features compared with the random method. As perturbation magnitude increases, the random method results in a significantly larger discrepancy between the prediction probabilities of the original and perturbed data.
Figure A1. Evaluation metric outcomes for ‘checking time on watch’ (Class 32), showing CAM (blue), Grad-CAM (orange), and the random (green) methods for (a) PGI, (b) PGU, (c) RISb, (d) RISj, (e) RISv, (f) ROS, and (g) RRS. Despite increasing perturbation magnitudes, CAM and Grad-CAM exhibit only marginally better performance compared with the random method in terms of PGI. This suggests that the expected correlation between increasing the perturbation magnitude of important features and significant changes in prediction output may not consistently apply to this particular case. Conversely, for PGU, CAM and Grad-CAM demonstrate more effective identification of unimportant features compared with the random method. As perturbation magnitude increases, the random method results in a significantly larger discrepancy between the prediction probabilities of the original and perturbed data.
Sensors 24 01940 g0a1
Figure A2. Evaluation metric outcomes for ‘clapping’ (Class 9), showing CAM (blue), Grad-CAM (orange), and the random (green) methods for (a) PGI, (b) PGU, (c) RISb, (d) RISj, (e) RISv, (f) ROS, and (g) RRS. Similar to class 32, the PGI results for class 9 show only a slight difference in performance between CAM and Grad-CAM versus the random method, even as perturbation magnitude increases. The PGU results still echo those in class 32, with CAM and Grad-CAM outperforming the random method in distinguishing unimportant features.
Figure A2. Evaluation metric outcomes for ‘clapping’ (Class 9), showing CAM (blue), Grad-CAM (orange), and the random (green) methods for (a) PGI, (b) PGU, (c) RISb, (d) RISj, (e) RISv, (f) ROS, and (g) RRS. Similar to class 32, the PGI results for class 9 show only a slight difference in performance between CAM and Grad-CAM versus the random method, even as perturbation magnitude increases. The PGU results still echo those in class 32, with CAM and Grad-CAM outperforming the random method in distinguishing unimportant features.
Sensors 24 01940 g0a2

Appendix B. Sample Data Instances and Their Corresponding CAM and Grad-CAM Scores (Normalized)

Table A5. Sample from class 9. k denotes body point number.
Table A5. Sample from class 9. k denotes body point number.
kCAMGrad-CAM
10.104999270.10499939
20.84571030.84571093
30.863372450.8633743
40.144989760.14498973
51.01.0
60.62175590.6217579
70.59377770.5937787
80.76765690.7676586
90.966938850.9669416
100.713242950.71324545
110.74239110.74239343
120.678572830.6785746
130.261586670.26158726
140.089999330.08999972
150.024059460.02405963
160.00.0
170.238065880.23806643
180.04834540.04834554
190.020317450.02031763
200.01408390.01408408
210.91717620.91717803
220.496377380.4963787
230.577730830.57773197
240.476859060.47686023
250.34912720.3491279
Table A6. Sample from class 11. k denotes body point number.
Table A6. Sample from class 11. k denotes body point number.
kCAMGrad-CAM
10.41922640.41922605
20.44854020.4485402
30.4845370.484536
40.38698530.38698477
50.400522950.40052223
60.257560670.2575602
70.246084750.24608456
80.146581040.14658089
90.61068670.610686
100.75916470.75916386
111.01.0
120.690970360.69096994
130.333058950.3330585
140.194020550.19402048
150.00.0
160.082376030.08237588
170.379312750.3793123
180.187308090.18730754
190.045331330.04533128
200.072445550.07244562
210.475199970.47519904
220.116827850.11682767
230.230020540.23002037
240.729682450.7296809
250.87520010.87519866
Table A7. Sample from class 26. k denotes body point number.
Table A7. Sample from class 26. k denotes body point number.
kCAMGrad-CAM
10.67755050.67755353
20.82871680.8287187
31.01.0
40.77158710.77158725
50.78182660.78182703
60.720559950.72056246
70.778308030.7783087
80.31765440.31765595
90.720879730.72088116
100.768083750.76808643
110.60134770.6013488
120.340844040.3408448
130.78554720.78554976
140.69869920.6986986
150.443961980.44396254
160.351806640.35180822
170.78082260.78082335
180.8398650.8398684
190.67823830.6782419
200.600778460.6007782
210.959547940.95955056
220.00.0
230.108865510.10886557
240.021422510.02142265
250.052662530.05266258
Table A8. Sample from class 32. k denotes body point number.
Table A8. Sample from class 32. k denotes body point number.
kCAMGrad-CAM
10.079231030.07923108
20.28505660.28505734
30.267728480.2677287
40.089147540.08914759
50.68605530.68605727
60.982236860.9822402
71.01.0
80.938313660.93831354
90.395569530.39556977
100.373555930.37355682
110.47071050.4707117
120.548542440.5485435
130.042359590.0423594
140.036659320.03665918
150.004186660.00418662
160.00637880.00637893
170.049259540.04925963
180.045810130.04581019
190.00.0
200.016380520.01638054
210.325322480.32532266
220.564110640.56411153
230.86762060.8676238
240.390548470.39054966
250.452427240.45242897

References

  1. Nguyen, T.N.; Huynh, H.H.; Meunier, J. Skeleton-based abnormal gait detection. Sensors 2016, 16, 1792. [Google Scholar] [CrossRef] [PubMed]
  2. Liu, C.; Fu, R.; Li, Y.; Gao, Y.; Shi, L.; Li, W. A self-attention augmented graph convolutional clustering networks for skeleton-based video anomaly behavior detection. Appl. Sci. 2021, 12, 4. [Google Scholar] [CrossRef]
  3. Guo, J.; Liu, H.; Li, X.; Xu, D.; Zhang, Y. An attention enhanced spatial–temporal graph convolutional LSTM network for action recognition in Karate. Appl. Sci. 2021, 11, 8641. [Google Scholar] [CrossRef]
  4. Usman, M.; Zhong, J. Skeleton-based motion prediction: A survey. Front. Comput. Neurosci. 2022, 16, 1051222. [Google Scholar] [CrossRef] [PubMed]
  5. Commission, E. Proposal for a Regulation of the European Parliament and of the Council Laying Down Harmonised Rules on Artificial Intelligence (Artificial Intelligence Act) and Amending Certain Union legIslative Acts; Proposal, European Commission: Brussels, Belgium, 2021. [Google Scholar]
  6. Yan, S.; Xiong, Y.; Lin, D. Spatial temporal graph convolutional networks for skeleton-based action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  7. Song, Y.F.; Zhang, Z.; Shan, C.; Wang, L. Stronger, faster and more explainable: A graph convolutional baseline for skeleton-based action recognition. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 1625–1633. [Google Scholar]
  8. Ghaleb, E.; Mertens, A.; Asteriadis, S.; Weiss, G. Skeleton-based explainable bodily expressed emotion recognition through graph convolutional networks. In Proceedings of the 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021), Jodhpur, India, 15–18 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–8. [Google Scholar]
  9. Das, P.; Ortega, A. Gradient-weighted class activation mapping for spatio temporal graph convolutional network. In Proceedings of the ICASSP 2022—2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 4043–4047. [Google Scholar]
  10. Wang, H.; He, F.; Peng, Z.; Shao, T.; Yang, Y.L.; Zhou, K.; Hogg, D. Understanding the robustness of skeleton-based action recognition under adversarial attack. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 14656–14665. [Google Scholar]
  11. Liu, J.; Akhtar, N.; Mian, A. Adversarial attack on skeleton-based human action recognition. IEEE Trans. Neural Netw. Learn. Syst. 2020, 33, 1609–1622. [Google Scholar] [CrossRef] [PubMed]
  12. Agarwal, C.; Krishna, S.; Saxena, E.; Pawelczyk, M.; Johnson, N.; Puri, I.; Zitnik, M.; Lakkaraju, H. Openxai: Towards a transparent evaluation of model explanations. Adv. Neural Inf. Process. Syst. 2022, 35, 15784–15799. [Google Scholar]
  13. Shahroudy, A.; Liu, J.; Ng, T.T.; Wang, G. Ntu rgb+ d: A large scale dataset for 3d human activity analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1010–1019. [Google Scholar]
  14. Song, Y.F.; Zhang, Z.; Shan, C.; Wang, L. Constructing stronger and faster baselines for skeleton-based action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1474–1488. [Google Scholar] [CrossRef] [PubMed]
  15. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  16. Alvarez Melis, D.; Jaakkola, T. Towards robust interpretability with self-explaining neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; Volume 31. [Google Scholar]
  17. Zhou, J.; Gandomi, A.H.; Chen, F.; Holzinger, A. Evaluating the quality of machine learning explanations: A survey on methods and metrics. Electronics 2021, 10, 593. [Google Scholar] [CrossRef]
  18. Markus, A.F.; Kors, J.A.; Rijnbeek, P.R. The role of explainability in creating trustworthy artificial intelligence for health care: A comprehensive survey of the terminology, design choices, and evaluation strategies. J. Biomed. Inform. 2021, 113, 103655. [Google Scholar] [CrossRef] [PubMed]
  19. Alvarez-Melis, D.; Jaakkola, T.S. On the robustness of interpretability methods. arXiv 2018, arXiv:1806.08049. [Google Scholar]
  20. Agarwal, C.; Johnson, N.; Pawelczyk, M.; Krishna, S.; Saxena, E.; Zitnik, M.; Lakkaraju, H. Rethinking stability for attribution-based explanations. arXiv 2022, arXiv:2203.06877. [Google Scholar]
  21. Otte, K.; Kayser, B.; Mansow-Model, S.; Verrel, J.; Paul, F.; Brandt, A.U.; Schmitz-Hübsch, T. Accuracy and reliability of the kinect version 2 for clinical measurement of motor function. PLoS ONE 2016, 11, e0166532. [Google Scholar] [CrossRef] [PubMed]
  22. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning deep features for discriminative localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar]
  23. Selvaraju, R.R.; Das, A.; Vedantam, R.; Cogswell, M.; Parikh, D.; Batra, D. Grad-CAM: Why did you say that? arXiv 2016, arXiv:1611.07450. [Google Scholar]
  24. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  25. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
Figure 1. Illustration of perturbing a point P(x, y, z) in 3D space to a new position P′(x′, y′, z′) using spherical coordinates. The perturbation magnitude is represented by r, with azimuthal angle θ and polar angle ϕ .
Figure 1. Illustration of perturbing a point P(x, y, z) in 3D space to a new position P′(x′, y′, z′) using spherical coordinates. The perturbation magnitude is represented by r, with azimuthal angle θ and polar angle ϕ .
Sensors 24 01940 g001
Figure 2. The EfficientGCN pipeline showing the variables for calculating faithfulness and stability. Perturbation is performed in the Data Preprocess stage.
Figure 2. The EfficientGCN pipeline showing the variables for calculating faithfulness and stability. Perturbation is performed in the Data Preprocess stage.
Sensors 24 01940 g002
Figure 3. (left) CAM, Grad-CAM, and baseline random attributions for a data instance in ‘standing up’ (class 8), averaged for all frames and normalized. The color gradient denotes the score intensity: blue indicates 0, and progressing to red indicates a score of 1; (right) the numerical values of the attribution scores, with k denoting the body point number.
Figure 3. (left) CAM, Grad-CAM, and baseline random attributions for a data instance in ‘standing up’ (class 8), averaged for all frames and normalized. The color gradient denotes the score intensity: blue indicates 0, and progressing to red indicates a score of 1; (right) the numerical values of the attribution scores, with k denoting the body point number.
Sensors 24 01940 g003
Figure 4. Evaluation metric outcomes for ‘Writing’ (Class 11, i.e., the weakest class), showing CAM (blue), Grad-CAM (orange), and the random (green) methods for (a) PGI, (b) PGU, (c) RISb, (d) RISj, (e) RISv, (f) ROS, and (g) RRS. The y-axis measures the metric values, while the x-axis shows the perturbation magnitude. CAM and Grad-CAM graphs overlap due to extremely similar metric outcomes.
Figure 4. Evaluation metric outcomes for ‘Writing’ (Class 11, i.e., the weakest class), showing CAM (blue), Grad-CAM (orange), and the random (green) methods for (a) PGI, (b) PGU, (c) RISb, (d) RISj, (e) RISv, (f) ROS, and (g) RRS. The y-axis measures the metric values, while the x-axis shows the perturbation magnitude. CAM and Grad-CAM graphs overlap due to extremely similar metric outcomes.
Sensors 24 01940 g004
Figure 5. Evaluation metric outcomes for ‘Jump Up’ (Class 26, i.e., the strongest class), showing CAM (blue), Grad-CAM (orange), and the random (green) methods for (a) PGI, (b) PGU, (c) RISb, (d) RISj, (e) RISv, (f) ROS, and (g) RRS. The y-axis measures the metric values, while the x-axis shows the perturbation magnitude. CAM and Grad-CAM graphs overlap due to extremely similar metric outcomes.
Figure 5. Evaluation metric outcomes for ‘Jump Up’ (Class 26, i.e., the strongest class), showing CAM (blue), Grad-CAM (orange), and the random (green) methods for (a) PGI, (b) PGU, (c) RISb, (d) RISj, (e) RISv, (f) ROS, and (g) RRS. The y-axis measures the metric values, while the x-axis shows the perturbation magnitude. CAM and Grad-CAM graphs overlap due to extremely similar metric outcomes.
Sensors 24 01940 g005
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pellano, K.N.; Strümke, I.; Ihlen, E.A.F. From Movements to Metrics: Evaluating Explainable AI Methods in Skeleton-Based Human Activity Recognition. Sensors 2024, 24, 1940. https://doi.org/10.3390/s24061940

AMA Style

Pellano KN, Strümke I, Ihlen EAF. From Movements to Metrics: Evaluating Explainable AI Methods in Skeleton-Based Human Activity Recognition. Sensors. 2024; 24(6):1940. https://doi.org/10.3390/s24061940

Chicago/Turabian Style

Pellano, Kimji N., Inga Strümke, and Espen A. F. Ihlen. 2024. "From Movements to Metrics: Evaluating Explainable AI Methods in Skeleton-Based Human Activity Recognition" Sensors 24, no. 6: 1940. https://doi.org/10.3390/s24061940

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop