Next Article in Journal
Minimizing Global Buffer Access in a Deep Learning Accelerator Using a Local Register File with a Rearranged Computational Sequence
Next Article in Special Issue
Implementation of Disassembler on Microcontroller Using Side-Channel Power Consumption Leakage
Previous Article in Journal
Performance Analysis of Optically Pumped 4He Magnetometers vs. Conventional SQUIDs: From Adult to Infant Head Models
Previous Article in Special Issue
Two-Stage Intelligent Model for Detecting Malicious DDoS Behavior
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Residual Network for Smartwatch-Based User Identification through Complex Hand Movements

by
Sakorn Mekruksavanich
1 and
Anuchit Jitpattanakul
2,3,*
1
Department of Computer Engineering, School of Information and Communication Technology, University of Phayao, Phayao 56000, Thailand
2
Department of Mathematics, Faculty of Applied Science, King Mongkut’s University of Technology North Bangkok, Bangkok 10800, Thailand
3
Intelligent and Nonlinear Dynamic Innovations Research Center, Science and Technology Research Institute, King Mongkut’s University of Technology North Bangkok, Bangkok 10800, Thailand
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(8), 3094; https://doi.org/10.3390/s22083094
Submission received: 22 February 2022 / Revised: 8 April 2022 / Accepted: 13 April 2022 / Published: 18 April 2022
(This article belongs to the Special Issue Security for Mobile Sensing Networks)

Abstract

:
Wearable technology has advanced significantly and is now used in various entertainment and business contexts. Authentication methods could be trustworthy, transparent, and non-intrusive to guarantee that users can engage in online communications without consequences. An authentication system on a security framework starts with a process for identifying the user to ensure that the user is permitted. Establishing and verifying an individual’s appearance usually requires a lot of effort. Recent years have seen an increase in the usage of activity-based user identification systems to identify individuals. Despite this, there has not been much research into how complex hand movements can be used to determine the identity of an individual. This research used a one-dimensional residual network with squeeze-and-excitation (SE) configurations called the 1D-ResNet-SE model to investigate hand movements and user identification. According to the findings, the SE modules have enhanced the one-dimensional residual network’s identification ability. As a deep learning model, the proposed methodology is capable of effectively identifying features from the input smartwatch sensor and could be utilized as an end-to-end model to clarify the modeling process. The 1D-ResNet-SE identification model is superior to the other models. Hand movement assessment based on deep learning is an effective technique to identify smartwatch users.

1. Introduction

Annually, the quantities of information produced by wearable devices linked to the Internet increase [1]. Smartphones play a significant part among these gadgets because of their increasing functionality and consumer acceptability. As a result, the safety and security of this equipment are the top priority throughout the design phase. Biometrics can be employed in several of the newest strategies for controlling illegal access to mobile devices. An individual’s observable characteristics and behaviors are examined and measured to recognize or identify that person.
User authentication is an excellent way to protect personal information. The design of the authentication mechanism must take into account the fact that the aim of authenticating is to validate the user’s information [2]. In preventing identity theft, many new solutions have been introduced in recent years. Identifying the user and providing a pleasant user experience are the primary goals of these detection techniques, but there are still several obstacles to overcome. Digital identity is increasingly built on usernames and passwords [3], making it vulnerable to theft, hacking, and fraud. Cryptographic algorithm-based digital signatures are another common choice [4]. A highly capable computer system is needed to produce digital signatures; therefore, devices with fewer resources have difficulties establishing this identification. Hardware-based PUF (Physical Unclonable Function) has recently emerged to identify individuals, and several authentication methods have been constructed on this basis [5]. There are certain drawbacks to PUF, however, including the need for additional equipment. A hardware-based identification solution is implemented via tokens and access cards [6].
A biometrics-based identification approach is the next advance in identifying and verifying individuals [7]. Due to obvious reasons, individuals are regarded more efficiently than the previously listed digital IDs. Because they are a part of ourselves, biometrics are easy to utilize. Compared to more conventional verification and identification methods, including credentials, PINs, and tokens, biometrics are almost impossible to lose or steal [8]. Secondly, since each person’s biometrics are distinctive, they are complicated to reproduce. It is also easy to verify the properties of biometric IDs [9]. Modern computer systems have several physiological biometrically based identifiers. Several products employ face detection to verify that the individual is who they indicate they are. When it comes to biometrics, the fingerprint is the most often utilized [10]. There are a number of common biometric signatures, including ECG/EEG characteristics [11], iris recognition [12], and palm vein variations [13]. All of these options need specialized technology to collect biometric data. This could be prohibitively costly, time consuming, and intrusive to the participant. A further drawback of these physiologically based approaches is that they are vulnerable to emulation. Fraudulent activity includes voice impersonation, iris-copying lenses, and concealment, and these are but a few examples.
Many emerging biometric identification alternatives are low cost, better suited than classic biometrics, or could be used in conjunction with more classical biometrics like multi-factor authentication to enhance security and usability. On the other hand, some biometric authentication methods demand human engagement, which might be difficult for the end-user. A few alternatives include entering the password, unlocking the smartphone via face recognition, or tapping the fingerprint reader. In continuous authentication, the user is required to authenticate many times [14]. This is more challenging for the user. Since biometric characteristics are collected indirectly when the user interacts with the device, movement sensor-based identification approaches such as wearable sensor-based gait recognition [15], contact gesture-based recognition, keystroke-based recognition, etc., could tackle this issue. Compared to standard vision-based movement identification, these methods are more private [16] and use less energy.
The advancement of science and technology has influenced the techniques of biometric identification. Fingerprinting, facial detection, retinal scanning, palm geometry, and voice recognition are some of the most well known techniques, but there are many more. Meanwhile, less invasive biometric variations are making their way into widespread utilization. Identifying persons based on the features of their activity is an example of this. There are both pros and cons to this approach, which indicate how individuals perform their daily routines. Since the person needs to carry the equipment (which will generally be a smartphone) or be captured to perform recognition by computer vision, its key benefits are that it enables automated, regular, and non-intrusive recognition. Biometric techniques that are less accurate than fingerprints might be seen as a drawback. Continuous and periodic identification could help with this. For gait-based person recognition systems, performance can be improved if the data samples are created by a broad range of user actions, ensuring that a valid classification is achieved in the least amount of time feasible.
In many homes and workplaces, smartphones have become an integral part of our daily lives and are routinely used to access cloud-based security apparatuses. Smartwatches offer an interesting environment for authentic identity verification through cloud-based solutions like Internet banking if a smartphone is easily stolen or cobbled. When using cloud-based or other data sources to connect mission-critical Internet services, it is vital to identify the genuine user who is doing so reliably. Automated and non-bypassable identification is required.
In the recent decade, learning techniques, including machine learning (ML), have been employed to achieve good outcomes with biometric-based user identification. Within controlled circumstances, machine learning techniques such as K-nearest neighbors, Support Vector Machine (SVM), Decision Tree (DT), and Random Forest (RF), among others, have been established to deliver satisfactory results [17,18,19]. The accuracy of these standard machine learning models is highly dependent on the strategy of human-manually extracted and selected features.
Nowadays, deep learning (DL) algorithms have succeeded in user identification studies. One of the most significant components of deep learning is its ability to automatically identify and classify features with increased accuracy, influencing user identification studies [20,21,22]. Deep neural networks can learn discriminative characteristics from raw data automatically, and they have revealed tremendous promise for evaluating diverse data and have a high capacity for generalization. Numerous primary and sophisticated deep learning models have been proposed to capitalize on deep learning approaches by compensating for the drawbacks of traditional machine learning while leveraging the multiple levels of characteristics available in various hierarchies. A hierarchy of layers is used in machine learning techniques to handle low- and high-level features, along with linear and nonlinear feature conversions at different levels, contributing to learning and optimizing features. To this end, deep learning models such as Recurrent Neural Networks (RNN), Convolutional Neural Networks (CNN), and Long Short-Term Memory (LSTM) are utilized to overcome the limitations of conventional machine learning algorithms that relied on manual feature selection, where an error in feature selection could have negative consequences for the applications at hand. As a result, deep learning networks have found practical applications in identification tasks and are often employed in activity recognition studies for feature extraction. One disadvantage of the deep learning approach, mainly when complex DL architectures are used, is the higher expense of processing the massive number of accessible datasets. Nevertheless, the cost is justified since an identification approach relies on the accuracy of the classification consequences of the deep learning model.
While enhancing the level of a CNN could extract more abstract features and improve effectiveness, it may also result in degraded performance [23]. To address this issue, He et al. [24] suggested a residual network (ResNet) for image identification, which has been integrated into the investigation of human behavior. For instance, Li et al. [25] used the 1D-ResNet model to extract spatial features from multidimensional inertial sensor inputs and bidirectional LSTM. This approach can achieve improved performance with fewer parameters. Moreover, Ronald et al. [26] established a link between wearable sensor vectors and individual motions using an improved deep learning model based on ResNet and Inception modules. The improved model suggested here shows exceptional arrangement in human activity recognition (HAR) implementations.
Using a squeeze-and-excitation (SE) module [27] is a technique for channel attention that could be contained in current CNNs to increase classification interpretation. The SE block operates as an embedding unit, improving the effectiveness of deep neural networks. The SE block is one of the popular improvements of the many CNN architectures because it is simply added without changing the shape of the existing model. For example, [28,29,30] concatenated the SE block into CNNs with varying convolutional layer levels. The findings indicate that CNNs with the SE block outperform simple CNNs with accuracy rates. Additionally, the efficiency of the SE block has been shown in the assessment of pork freshness using NIRS [31] and ECG-signal classification [32].
Inspired by the works mentioned earlier, this research combines the SE block with a 1D-ResNet to evaluate the SE block’s potential for user identification using sensor data from a smartwatch. The main difference between the proposed network and our previous works in [28,29,30] is that the proposed network uses a one-dimensional deep residual network with shortcut connections instead of a one-dimensional convolutional neural network. This work aims to investigate user identification using a smartwatch on the basis of complex hand movements. We use a residual network to extract more abstract spatial characteristics from CNN. Squeeze-and-excitation modules were also incorporated in the one-dimensional ResNet to improve recognition interpretation.
The following is an overview of the study’s most significant contributions:
  • This article aims to investigate the possibility of the 1D-ResNet-SE for sensor-based user identification by analyzing complex hand movement signals captured by smartwatch sensors. We compared standard CNN-based deep learning models to RNN-based LSTM networks for sensor-based user identification using smartwatch sensor data to examine the algorithm’s effectiveness.
  • We conducted comprehensive experiments using many smartwatch-based HAR datasets encompassing simple and complex hand movements to increase ecological validity. We observed the connection between hand motion patterns and the recognition of smartwatch owners using the 1D-ResNet-SE model. The SE blocks are combined with the residual network to increase the sensitivity to relevant features. Compared to CNN- and LSTM-based deep learning models, the model demonstrated here shower superior performance in user identification in complex hand movement situations.
The rest of this article is organized as follows: In the second section, we examine the latest research on sensor-based deep learning algorithms for user identification. Proposals for new techniques are introduced in Section 3. Experiments are described, and findings are shown in Section 4. Deep-learning algorithms used in the research to achieve effectiveness in identification are discussed in detail in Section 5. The last section, Section 6, focuses on the study’s limitations and suggests new opportunities for future research.

2. Related Works

Identifying users based on their activities has proved to be a challenging problem to solve. We have compiled a collection of resources connected to our study in this area.

2.1. Sensor-Based User Identification

Wearable sensors have been considered in recent years as part of sensor-based identification systems. For example, [33,34] proposed a mechanism for explicitly and continuously identifying the individual. However, there is no compelling evidence that the individual’s behavior has changed enough to warrant an apparent classification in most circumstances. Luca et al. [35] present a technique for explicitly determining the distance between pattern traces using the dynamic time warping mechanism. Most of the 22 unusual touch patterns shown by Sae-Bae et al. [36] include using all five fingers concurrently. They used k-nearest neighbors and support vector machines to categorize the 22 analytical characteristics from touch traces analyzed in the study [37].
As a result, each action is correlated with two essential characteristics: Time and space, according to the behavior-based model’s principle. For example, if you are looking to identify a user, you could look to activities such as those described by [38]. Multi-model continuous user identification was suggested by the works of [39]. Another distinctive architecture for ongoing user identification was proposed in [40] by leveraging historical smartphone records and positions.
To a certain degree, all the tasks mentioned above need additional details and a source of user identification. Casale et al. [41] provided a gait-based user identification over an inconspicuous biometric pattern to address these concerns. A four-layered structure made use of the geometric principle of a convex hull. For example, it only functions in specific locations which is a severe disadvantage. Wearable devices based on gait signals recorded using a three-dimensional accelerometer were employed in the studies of [8,42], where the accelerometer was simply attached to the individual’s waist at the rear. A three-fold technique for user identification based on data distribution statistics, correlation, and time-frequency characteristics was developed by [43]. At the same time, the people were deliberately requested to stroll at different velocities, such as slow, regular, or quick. The fundamental disadvantage of Mantyjarvi’s work is that only one person can walk at a time, and with relatively restricted variants.
There are many existing approaches that used gait-based systems which are summarized above and in Table 1. There is still scope for further improvement based on physical adjustments, carrying objects, orientation, placement, movement surface, psychosocial factors of a participant, stimulants, and other considerations. These limitations significantly hamper the gait-based system’s performance in real-world situations.

2.2. Deep Learning Approaches for User Identification

2.2.1. Convolutional Neural Network

Several research studies on time series classification (TSC) have focused on deep learning and obtained notable results in recent years. CNN has been a prominent deep learning technique in TSC because of its capability to extract the connection between local organizations in the form of array information. Yang et al. [46] reveal one of the first applications of CNN in TSC. According to the researchers, a higher-level description of raw sensor data can be derived using CNN’s deep architecture. Additionally, combining feature learning and classification in a single model makes the learned features more discriminative. According to Ronao and Cho [47], a deep CNN with 1D convolutional processes outperforms conventional pattern recognition algorithms for movement categorization employing smartphone sensors. Jiang and colleagues [48] sent the sensor data into a two-dimensional neural network instead of utilizing a one-dimensional convolution to capture both temporal and spatial characteristics from the action patterns for the classification test. The two-stage CNN model [49] increases the classification performance of actions with complicated structures and limited training data.
TSC has subsequently benefited from the use of numerous cutting-edge CNN architectures that have been introduced in the machine vision area. A TSC model based on U-Net [50] was presented in [51] to conduct sampling point-level forecasting, thereby overcoming the multi-class issue. Mahmud et al. [52] use a residual block-based CNN to extract features and categorize behaviors from 1D time-series sensor data. TSC’s compact deep convolutional neural network is constructed by Tang et al. [53] using the Lego filter [54].

2.2.2. Recurrent Neural Networks

Time series sensor data are commonly processed using recurrent neural networks (RNNs) because they store insights into the history of previous items in a sequence. Zeng et al. [55] presented a long short-term memory (LSTM) model based on continuous attention that emphasizes relevant sensor modalities and significant sections of the sensor data during TSC analysis. Barut et al. [56] constructed a multitask framework employing layered LSTM layers to classify and estimate activity intensity from raw sensor data. Rather than utilizing raw data, the bidirectional LSTM recurrent neural network in [57] is employed to feature data generated from principal component analysis (PCA) and discrete wavelet transform (DWT). Where there is a lack of label data, [58] recommends extracting features using spectrograms. Then the identification is carried out using an extended support vector machine (SVM). Fusing LSTM-RNN with handcrafted elements could improve the performance of a system, according to [59], where it was shown. Local feature-based LSTM networks suggested by Chen et al. [60] can encode temporal dependence and learn features from a high sampling rate of acceleration data.

2.2.3. Hybrid Neural Networks

In recent years, considerable research has demonstrated that substantial TSC effectiveness could well be achieved by combining hybrid models derived from several kinds of deep learning approaches. GRUs (gated recurrent units) have been introduced in [61] to uncover sequential temporal relationships in complex activity recognition by using an inception module-based CNN [62]. In a sleep-wake detection system, Chen et al. [63] employed a 1D- CNN-LSTM model to capture feature information from lengthy acceleration sequences, then combined an attention mechanism with the handcrafted characteristics of heart rate variability data. In [64], a recurrent convolutional attention model was presented to cope with the imbalance of the labeled data in a semi-supervised manner. Small segments of window data are supplied into an LSTM layer for motion identification after a CNN is applied to the data. For the first time, an LSTM-CNN model was suggested by Xia et al. [65] in which a two-layer LSTM is applied topically to the raw sensor data before actually employing 2D convolutional layers. Deep learning and traditional pattern recognition approaches are successful in the research; however, further examination exposes several gaps and flaws. Rather than evaluating the connection between neighboring windows, most research has solely looked at the data from specific windows to make predictions about behavioral aspects. In multi-class classification applications, including face recognition, this technique could deliver great accuracy, but it can lack the characteristic of long-term reliance on sensor data. A method named MFAP, developed by Chen et al. [59], addresses this weakness by considering both the past and present a priori data. To maintain the assumption of independence between the observed values and the preceding ones, we consider the activity sequence a first-order Markov chain. This strategy, unfortunately, necessitates an additional manual job on the result of the deep neural network’s Softmax layer.
There has also been some research involved to apply and evaluate the principles under real-life scenarios; however, most past research employs clean datasets. The data of each action are gathered, interpreted, and preserved independently without taking transitions between movements into consideration. In reality, activities must be performed sequentially, and some, such as lying down and jogging, cannot be performed side by side without a transition. A hierarchical hybrid approach, known as HiHAR, has been proposed to overcome these issues. The process can determine local and global temporal dependence in window sequences using the hierarchical design.

2.3. Simple and Complex Human Activities

Human activities could well be classified into two categories according to [66,67,68,69]: Simple human activities (SHA) and complex human activities (CHA). As Shoaib et al. [70] observed, simple human activities are repeated, common movements that could be primarily determined using an accelerometer, such as strolling, running, sitting, and standing. Another issue is that behaviors that are not repeatable, such as smoking, eating, delivering a speech, or sipping coffee, cannot be clearly detected at smaller segmentation windows, in contrast to repetitive activities such as walking, running, or cycling. Human behaviors that are complex are less repetitive than those that are simple. Complex activities often need the use of the hands, such as smoking, eating, and drinking. Additional sensors, like a gyroscope, could be utilized to determine if CHA is present. Due to the difficulty of characterizing such actions with a single accelerometer, this research categorized stair-related movements as CHA.
Alo et al. [66] distinguished two types of human activities: Simple and complex. Walking, running, sitting, standing, and jogging are simple human activities that are quick human behaviors. On the other hand, complex human activities, such as smoking, eating, taking medicine, cooking, and writing, are composed of longer-duration operations. Peng et al. [67] divided human activities into simple ones (e.g., walking, jogging, or sitting) based on repetitive movements or a single body position, which does not genuinely describe everyday activities. On the other hand, complex activities are more challenging and are composed of many straightforward operations. Complex actions, such as eating breakfast, office working, or shopping usually require an extended period of time and have broad meanings. These are more accurate components of people’s everyday lives. According to Liu et al. [68], human activity is complicated. A complex activity is a collection of chronologically and productively related atomic engagements. In contrast, an atomic movement is a single unit-level action that cannot be further decomposed under practical comprehension. Rather than doing a single atomic operation, individuals frequently perform several activities in various ways, both sequentially and simultaneously. Chen et al. [69] distinguished two types of human activities: Simple and complex. SHA could be considered as a single repeated motion that a single accelerometer could recognize. CHA would rarely occur in repeatable form similar to simple activities and will usually include many simultaneous or overlapping actions that can be observed only via multimodal sensor data.

2.4. Available Sensor-Based Activity Datasets

Many sensor-based activity datasets are accessible to the public and could develop deep learning models.
All 51 individuals in the WISDM-HARB dataset [17] were recorded while participating in 18 activities of daily life. Each participant wore a smartwatch on their dominant wrist while completing the tasks to ensure accuracy. The research goal was to identify which integration form of accelerometer and gyroscope sensors achieved the best performance on both smartphones and smartwatches.
Smartwatch and smartphone data loggers are included in the UT-Smoke dataset [71,72] to collect various sensor data simultaneously. For three months, the participants in this study smoked for a total of 17 hours while strolling, standing, sitting, or speaking with others. Eleven people volunteered to take part in these events. According to our knowledge, this is the most significant dataset compared to other research of this type.
Annotated data from complicated hand-based movements recorded by smartwatches are utilized as a baseline for complex hand movement studies in the two datasets above [73,74,75]. UT-Complex [70], PAMAP2 [76], and OPPORTUNITY [77] are further sensor-based activity datasets. On the other hand, this research did not include data from an annotated smartwatch sensor for sophisticated hand-based tasks.

3. Proposed Methodology

An emphasis is placed in this part on the methods used for the training of the advanced learning model and the identification of individuals through wearable sensors and smartwatches which incorporate built-in sensors. In Figure 1, the proposed methodology for the CHM-UserIden framework is shown, which comprises data acquisition, pre-processing, training the model, and user identification. Each stage is explained in further detail as follows.

3.1. Data Acquisition

The benchmark datasets utilized to evaluate this research were the focus of this section. The assessment employed two public datasets (UT-Smoke and WISDM-HARB datasets). The inertial data from smartwatch sensors were included in the UT-Smoke and WISDM-HARB datasets. As a group of individuals engaged in everyday tasks, such as dining, having a drink, smoking, and so on, the data in each dataset were gathered. Data from accelerometer, gyroscope, and magnetometer-equipped smartwatch sensors were used to produce these datasets.
To investigate user identification through a smartwatch, we classified human activities using the SC 2 representational taxonomy [68]. This division of human activities into simple and complex ones was based on their chronological interconnections.
  • A simple activity cannot be subdivided further at the atomic scale. For instance, walking, running, and ascending are all considered simple activities owing to their inability to be coupled with other activities.
  • A complex activity is a high-level activity formed via the sequencing or overlapping of atomic-level activities. For example, the representation ”smoking while strolling” incorporates the two atomic actions of ”strolling” and ”smoking”.
Characteristics of both activity-based datasets are described in Table 2.

3.1.1. UT-Smoke Dataset

The UT-Smoke dataset, which was previously provided in [71,72], is used in this study as a public complex hand-based activity dataset. Over three months, 11 volunteers (two female and nine male) aged 20–45 were tracked using a smartwatch application. The program records data from a smartwatch and a smartphone’s triaxial accelerometer and gyroscope, as well as a timestamp. 50 Hz is the sampling rate for all data. The activities included smoking while standing (SmokeSD), smoking while sitting (SmokeST), smoking while walking (SmokeW), smoking in a group chat (SmokeG), drinking while standing (DrinkSD), drinking while sitting (DrinkST), dining, standing, sitting, and walking (Walk). Identifying smoking and behaviors comparable to smoking are the primary goals of this dataset. Every individual takes part in every activity except for SmokeG and SmokeW, which are carried out by a combined total of eight and three individuals, respectively.

3.1.2. WISDM-HARB Dataset

Fifty-one people were recruited to participate in the WISDM-HARB dataset [17] and complete various everyday tasks, including easy and sophisticated studies using smartphones and smartphone sensors. The subjects performed these tasks for three minutes. The accelerometer and gyroscope sensors recorded data at 20 Hz. Individuals aged 19 to 48 volunteered to participate in the study, which collected sensor data.

3.2. Data Pre-Processing

Due to the participants’ lively motions throughout the data collection, raw sensor data contained measurement noise and other unanticipated noise. Signals with a lot of noise distort the data they convey. As a result, it was critical to limit the impact of noise on signal processing so that useful information could be retrieved from the signal [42,78]. Mean, low-pass, and Wavelet filtering are some of the most frequently used techniques for filtration. Using a 3rd order Butterworth filter, we de-noised all three dimensions of accelerometers, gyroscopes, and magnetometers using the 20 Hz cutoff frequency. At this pace, 99.9% of body movements are captured, making it ideal for the recording of motion [79].
It was necessary to alter the sensor data once it had been cleansed of unwanted noise. Each data point was transformed using a Min–Max normalization approach, which projects its values into the range [0, 1]. Having a way to balance the impacts of different dimensions might be beneficial for the learning processes. Normalized data from all sensors are split into equal-sized sections for model training using fixed-size sliding windows in the data segmentation stage of the process. To construct sensory data streams with a length, we employed a sliding window with a duration of 10 s in this study as suggested by [17]. The 10-s window is utilized for user identification because it is long enough to record crucial features of a person’s activities, such as numerous repeats of fundamental motions such as walking and stair ascending, and it enables faster biometric identification. Additionally, prior activity recognition investigations revealed that a 10-s window size surpasses others [80].

3.3. Data Generation

Data samples are separated into training and test data in this phase, while temporal windows from the signals are utilized to create a model, and test data are used to assess the learned model. Cross-validation is the standard approach for separating data into training and test sets [81]. Numerous strategies, such as k-fold cross-validation [7], could be used to separate the data for training and testing. This stage estimates the learning algorithm’s capacity to generalize to new data. This stage takes advantage of stratified ten-fold cross-validation inside the framework for smartwatch-based user identification. The entire dataset is partitioned into 10 equal folds or subsets for this validation approach. Nine of these folds are utilized for training and one for testing in each cycle. This procedure is performed 10 times, utilizing all data for both training and testing. Stratified data imply that each fold has about the same amount of data from each participant.

3.4. The Proposed 1D-ResNet-SE Model

3.4.1. Architecture of the Proposed Model

This work introduces the 1D-ResNet-SE identification model based on complex hand motions acquired by smartwatch sensors to achieve effective biometric user identification. The presented model can automatically extract characteristic features from input sensor data. A convolutional block and SE-ResNet blocks are shown in Figure 2 for capturing spatial features, accompanied by a global average pooling (GAP) layer, and flattened and connected layers directly for further processing.
The input sensors were processed using convolutional blocks and SE-ResNet blocks. An ELU layer, a convolutional layer, a batch normalization layer, and a max-pooling layer were all included in the convolution component. Each of the trainable convolutional kernels in the convolutional layer creates a feature map, which is then used in the convolutional layer. One-dimensional kernels are just like the input spectrum. Because of this, BN was used to stabilize and speed up the learning process. The model’s expression capability was improved with the help of ELU, a nonlinear function. Preserving key characteristics was achieved by using the MP layer to minimize map size. The following section goes into further detail about the SE-ResNet module. Flattened layers were utilized to turn the averages of each feature map into a 1D vector using the GAP. Using a Softmax function, the result of the fully linked layer was transformed into probabilistic reasoning. The cross entropy loss function, which is often used in classification applications, was applied to compute the network’s losses.

3.4.2. SE-ResNet Block

As the network layers increased, a degradation incident occurred: Accuracy rapidly reached saturation and ultimately declined [82]. Adding a bypass link to ResNet’s residual block could successfully solve the degradation issue [24]. Figure 3 depicts the architecture of a residual block. Convolutional layers, BN, ELU, and a bypass connection are all part of this algorithm. There are no differences between the residual block and the convolutional block except for the bypass link. A residual function F(x): = H(x) − x is defined as the sum of the foundation mappings H ( x ) and F ( x ) placed on top of each other. Since the initial mapping was transformed into F ( x ) + x , the initial mapping is no longer relevant. The residual learning is simpler to implement and avoids the degradation issue than simply fitting H ( x ) using stacked layers.

3.4.3. Squeeze-and-Excitation Module

By combining spatial and channel-specific data, convolutional neural networks extracted the features [83]. The SE module aims to improve the representative capacity of the model’s channel association. After the convolution procedure, many feature maps are obtained. Nonetheless, a few feature maps could well be overloaded with duplication of data. The SE block performs feature recalibration to improve the valuable traits while inhibiting the less valuable ones. Each feature map is squeezed as a first step, and a weight vector is generated. The feature weights are then redistributed using fully connected layers and a sigmoid activation function in the excitation procedure. A gradient descent technique is used to direct the redistribution. Weights are then used to adjust the weights of the features. To recalibrate the feature maps obtained from the stacked layers, the SE block was put behind BN in each residual block in this investigation. Figure 4 depicts the SE-ResNet component’s overall structure and functionality.

3.4.4. Activation Function

The activation function provides a nonlinear element in the model as an essential component. Nonlinear distributed data are challenging to adjust in a network lacking activation functions. Because of this, a network’s capacity to conform to its environment is greatly improved by the activation function. The activation functions that are utilized in this study are as follows.
  • Sigmoid function:
    σ ( x ) = 1 1 + e x
  • Rectified Linear Unit (ReLU):
    R e L U ( x ) = m a x ( x , 0 ) = x if x 0 0 if x < 0
  • Exponential Linear Unit (ELU)
    E L U ( x ) = x if x 0 α ( e x 1 ) if x < 0 , and α defaults to 1.0

3.5. Evaluation Metrics

User identification could be seen as a categorization with many classes. Accuracy, F1-score, and Equal Error Rate are commonly used performance indicators for evaluating and comparing identification systems. These performance indicators are determined using a confusion matrix to determine the model’s ability to detect objects.
Let consider a multiclass classification problem with set A containing the n different class labels C i ( i = 1 , 2 , 3 , , n ) denoted by { C 1 , C 2 , C 3 , , C n } . The confusion matrix for that problem is an n × n matrix presented in Figure 5. Each row of the matrix represents the instances of an actual class, while each column represents the instances of a predicted class. An element C i j of the confusion matrix at row i and column j provides the number of instances for which the actual class is i and the predicted class is j.
True positive (TP), false positive (FP), true negative (TN), and false negative (FN) are all aspects that could be derived from the confusion matrix and utilized to produce performance metrics. Consider the following mathematical formulae for calculating the label classes C i , T P ( C i ) , F P ( C i ) , F N ( C i ) , and T N ( C i ) .
T P ( C i ) = C i i
F P ( C i ) = l = 1 n C l i T P ( C i )
F N ( C i ) = l = 1 n C i l T P ( C i )
T N ( C i ) = l = 1 n k = 1 n C l k T P ( C i ) F P ( C i ) F N ( C i )
From Equations (4)–(7), we defined accuracy, precision, recall, and f1-score, for a multiclass confusion matrix in Table 3.
When a biometric-based user identification approach recognizes an invalid individual or fails to realize an actual individual, an error of accuracy happens. False Acceptance Rate (FAR) and False Rejection Rate (FRR) are the most usually exploited measures to determine issues. Equal Error Rate (EER) represents the rate at which FAR and FRR become equal, and hence, a lower EER rate indicates more accuracy.
The typical technique for assessing FAR and FRR for multiclass classifiers is converting the multiclass classification issue to multiple binary classifications. Each class has its own FAR and FRR error values. The EER can be calculated as, E E R = F A R + F R R 2 , where |FAR + FRR| is the smallest value.

4. Experimental Results

This section provides the results of all of the experiments we conducted to find the successful deep learning models for sensor-based user identification. The UT-Smoke and WISDM-HARB datasets were used as the two benchmark datasets for person identification utilizing smartwatch sensing data in the research. The accuracy, F1-score, and confusion matrix of the deep learning models were evaluated using these measures.

4.1. Software Configuration

Google Colab Pro+ [84] was utilized in this investigation. A graphics processor device called the Tesla V100-SXM2-16GB was used to accelerate the training of the deep learning models. There are several basic deep learning techniques in the Python library, including the 1D-ResNet-SE and Tensorflow backend (version 3.9.1) [85] and CUDA (8.0.6). The following Python libraries were the subject of these explorations:
  • When reading, manipulating, and interpreting sensor data, Numpy and Pandas were utilized for data management.
  • For plotting and displaying the outcomes of data discovery and model assessment, Matplotlib and Seaborn were utilized.
  • Scikit-learn (Sklearn) was used in experiments as a library for sampling and data generation.
  • Deep learning models were implemented and trained using TensorFlow, Keras, and TensorBoard.

4.2. Experimental Findings

The UT-Smoke and WISDM-HARB datasets were used to validate the developed approach against baseline deep learning methods. These deep learning approaches trained on smartwatch sensing data from the benchmark datasets are described in the following subsections, which give experimental findings. Summary hyperparameters of all models conducted in this study are described in Appendix A.

4.2.1. UT-Smoke

The UT-Smoke dataset was used to collect smartwatch sensor data from 11 participants. Smoking, Eating, Drinking, and Inactive are the four categories of physical activities listed in Table 4. Using classification performance indicators, the evaluated results of the deep learning models were measured (Accuracy and F1 measurements).
Several combinations of smartwatch sensor data and other deep learning techniques, such as CNN and the proposed 1D-ResNet-SE method, can be investigated on the UT-Smoke dataset. To see the categorization results for the DL models mentioned in Table 4, F1 of our recommended 1D-ResNet-SE model derives the score of 97.24% for smoking, 98.13% for eating, and 96.44% for drinking when employing accelerometer, gyroscope, and magnetometer data accordingly. The proposed 1D-ResNet-SE has a greater accuracy and F1 score than existing smartwatch sensor variations. Therefore, we could infer that the approach we propose can recognize smartwatch users quite effectively by employing complex hand movements.

4.2.2. WISDM-HARB

As a second dataset, we used the WISDM-HARB dataset. Smartwatch sensor readings from 44 persons performing 18 physical activities are included in this dataset. This dataset’s classification effectiveness is summarized in Table 5, Table 6 and Table 7.
With the WISDM-HARB dataset, we conducted extensive analysis utilizing two baseline DL models and proposed the 1D-ResNet-SE model. Three separate sensor configurations made use of data from the smartwatch. For ”Clapping” and ”Teeth”, the proposed approach obtained the highest F1 (>95 percent) utilizing both accelerometer and gyroscope data, as shown in Table 5, Table 6 and Table 7.

5. Research Discussion

This study aimed to present a deep learning-based framework for identifying users through complicated hand movements using a smartwatch. The proposed approach was evaluated against two distinct benchmark datasets comprising sensor data of various physical human activities acquired by smartwatch motion sensors (accelerometer, gyroscope, and magnetometer). The 1D-ResNet-SE model outperformed previous standard deep learning techniques for smartwatch-based user identification according to experimental outcomes. The 1D-ResNet-SE model uses shortcut connections to resolve the network’s vanishing gradient issue. The proposed model includes SE-ResNet blocks consisting of Conv1D layers, BN layers, ELU layers, squeeze-and-excitation (SE) modules, and a shortcut connection. By combining spatial and channel-specific data, the SE-ResNet block improves identification performance and hierarchically extracts features.

5.1. Impact of Squeeze-and-Excitation Modules

It was hypothesized that the squeeze-and-excitation (SE) module might enhance a deep learning model’s channel representational capability. This effort necessitates numerous feature maps and subsequent convolutional procedures. Repetitive information could exist in a few feature maps. The SE module performs feature recalibration to improve the significant attributes while inhibiting the less effective ones. Additional experiments were conducted to compare the introduced 1D-ResNet-SE model versus a modified model that took out the SE component to explore how the SE module affected the results.
To analyze the improvement, a statistical analysis was performed to find out whether there are significant performance differences of accuracy between the baseline 1D-ResNet model and the proposed 1D-ResNet-SE. As suggestion in [86], we perform the Wilcoxon test [87], which is non-parametric statistical test for pairwise comparing the significant difference. In the statistical test, we assume that the null hypothesis H 0 is as follows: ”There are no significant difference between the model performances”. When performing the non-parametric Wilcoxon test, the null hypothesis that all model performances were equal could be rejected with a significance level of α = 0.05. Hence, the result is statistically significant when p-value < 0.05.
Table 8 and Table 9 report the statistical analysis performed via the Wilcoxon test on the UT-Smoke and WISDM-HARB datasets, respectively. Based on the UT-Smoke dataset, the statistical test reveals that the SE module significantly improves accuracy of smartwatch-based user identification using the sensor data of smoking and drinking activities. For statistical analysis based on the WISDM-HARB, the analyzed results reveal similarly that the SE module can improve the user identification using the sensor data of typing, writing, clapping, eating sandwiches, and drinking activities with statistical significance.

5.2. Impact of Sensor Combinations

Each sensor’s usefulness to the smartwatch-based user identification is examined in this task. Using accelerometer and gyroscope data as independent inputs, we evaluated the efficiency of the suggested 1D-ResNet-SE model. Utilizing raw accelerometer data with the proposed model resulted in a superior F1 score compared to gyroscope data for all hand-based activities. To analyze the impact of sensor combinations, we utilized the Friedman aligned ranking test [88], which is a non-parametric statistical test for comparing the significant difference. In addition, we applied the Finner post-hoc test [89] with a significance level of α = 0.05 to examine whether the differences in the performance of the model accuracies were statistically significant.
Table 10 and Table 11 present the statistical analyses, performed with non-parametric comparisons that relate to the accuracy metrics of the 1D-ResNet-SE using different sensor data for user identification. The statistical results indicate that the accuracy performance of the 1D-ResNet-SE can be improved significantly by using both accelerometer and gyroscope for smoking, drinking, and eating activities.

5.3. Comparison with Previous Works

The recommended 1D-ResNet-SE model is compared to previously trained models on the same dataset (WISDM-HARB). Previous research [17] has revealed that using a machine learning technique called the Random Forest (RF) technique makes it possible to reach high-performance user identification using smartwatch sensors. Prior work presented the stratified 10-fold cross-validation approach, which we employed in our study. Table 12 outlines the comparative findings. The comparison results indicate that the proposed 1D-ResNet-SE model achieved better accuracy than the previous model for most of the activities.

6. Conclusions and Future Studies

Using complicated hand gestures and a smartwatch, this study proposes a heterogeneous framework for user identification. Two independent benchmark datasets comprising sensor data from smartwatch motion sensors acquired during diverse individual biological movements were used to evaluate the system (accelerometer, gyroscope, and magnetometer). Three deep learning models were used to classify each dataset’s sensor data, including the standard CNN and LSTM and our proposed 1D-ResNet-SE model.
Metrics such as accuracy and the F-measure were used to determine the experimental outcomes. Classifiers were compared to see how well they performed. Across both datasets, the proposed 1D-ResNet-SE classifier outperformed every other classifier by a wide margin. For user identification, the UT-Smoke dataset delivered high performances from complex hand movements such as eating, smoking, and drinking. We used all three smartwatch sensors (accelerometer, gyroscope, and magnetometer) to classify eating behavior in the UT-Smoke dataset to get a great outcome. Each DL classifier employed in this study performed well with accelerometer data when evaluating its identification capability as a smartwatch sensor. As an alternative, the gyroscope and magnetometer could be utilized to identify individuals. Similar to the WISDM-HARB dataset, the three DL classifiers were examined and assessed using smartwatch sensor data from 18 physical activities. The 1D-ResNet-SE classifier surpassed the other baseline DL classifiers in the investigation. User identification also gave valuable insights into the nature of users’ actions. Using a smartwatch to identify a user was an effective solution for this kind of action.
Even though the existing smartwatch-based user identification sensor method achieves good results, future studies might benefit from researching different replacements for the proposed solution. Another option is to include a wide range of activities, such as more complicated and transitional tasks, within a systematic framework to increase user identification. In the future, a complete smartphone and smartwatch dataset could be evaluated that includes numerous body locations for smartphone placements, since the smartwatch sensor data are only investigated in one position in the current study. Position-based user identification can be used to enhance identification outcomes in this manner.

Author Contributions

Conceptualization and model analysis, S.M.; resource and data curation, A.J.; methodology and validation, S.M.; data visualization and graphic improvement, A.J.; discussion and final editing, S.M.; writing—review and editing, S.M.; funding acquisition, A.J. and S.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research project was supported by University of Phayao (Grant No. FF65-RIM041), Thailand Science Research and Innovation Fund, National Science, Research and Innovation Fund (NSRF), and King Mongkut’s University of Technology North Bangkok with Contract no. KMUTNB-FF-65-27.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. The summary of hyperparameters for the CNN network used in this work.
Table A1. The summary of hyperparameters for the CNN network used in this work.
StageHyperparametersValues
ArchitectureConvolutionKernel Size5
Stride1
Filters64
Dropout 0.25
Max Pooling 2
Flatten -
TrainingLoss Function Cross-entropy
Optimizer Adam
Batch Size 64
Number of Epochs 200
Table A2. The summary of hyperparameters for the LSTM network used in this work.
Table A2. The summary of hyperparameters for the LSTM network used in this work.
StageHyperparametersValues
ArchitectureLSTM Unit128
Dropout0.25
Dense128
TrainingLoss FunctionCross-entropy
OptimizerAdam
Batch Size64
Number of Epochs200
Table A3. The summary of hyperparameters for the 1D-ResNet network used in this work.
Table A3. The summary of hyperparameters for the 1D-ResNet network used in this work.
StageHyperparametersValues
ArchitectureConvolutional Block
ConvolutionKernel Size5
Stride1
Filters64
Batch Normalization -
Activation ReLU
Max Pooling 2
Residual Block × 8
ConvolutionKernel Size5
Stride1
Filters32
Batch Normalization -
Activation ReLU
ConvolutionKernel Size5
Stride1
Filters64
Batch Normalization -
Global Average Pooling -
Flatten -
Dense 128
TrainingLoss Function Cross-entropy
Optimizer Adam
Batch Size 64
Number of Epochs 200
Table A4. The summary of hyperparameters for the 1D-ResNet-SE network used in this work.
Table A4. The summary of hyperparameters for the 1D-ResNet-SE network used in this work.
StageHyperparametersValues
ArchitectureConvolutional Block
ConvolutionKernel Size5
Stride1
Filters64
Batch Normalization -
Activation ELU
Max Pooling 2
SE-ResNet Block × 8
ConvolutionKernel Size5
Stride1
Filters32
Batch Normalization -
Activation ELU
ConvolutionKernel Size5
Stride1
Filters64
Batch Normalization -
SE Module -
Global Average Pooling -
Flatten -
Dense 128
TrainingLoss Function Cross-entropy
Optimizer Adam
Batch Size 64
Number of Epochs 200

References

  1. Ometov, A.; Shubina, V.; Klus, L.; Skibińska, J.; Saafi, S.; Pascacio, P.; Flueratoru, L.; Gaibor, D.Q.; Chukhno, N.; Chukhno, O.; et al. A Survey on Wearable Technology: History, State-of-the-Art and Current Challenges. Comput. Netw. 2021, 193, 108074. [Google Scholar] [CrossRef]
  2. Nickel, C.; Wirtl, T.; Busch, C. Authentication of Smartphone Users Based on the Way They Walk Using k-NN Algorithm. In Proceedings of the 2012 Eighth International Conference on Intelligent Information Hiding and Multimedia Signal Processing, Athens, Greece, 18–20 July 2012; pp. 16–20. [Google Scholar] [CrossRef]
  3. Lin, C.L.; Hwang, T. A password authentication scheme with secure password updating. Comput. Secur. 2003, 22, 68–72. [Google Scholar] [CrossRef]
  4. Merkle, R.C. A Digital Signature Based on a Conventional Encryption Function. Advances in Cryptology–CRYPTO ’87; Pomerance, C., Ed.; Springer: Berlin, Germany, 1988; pp. 369–378. [Google Scholar]
  5. Suh, G.E.; Devadas, S. Physical Unclonable Functions for Device Authentication and Secret Key Generation. In Proceedings of the 2007 44th ACM/IEEE Design Automation Conference, San Diego, CA, USA, 4–8 June 2007; pp. 9–14. [Google Scholar]
  6. Indu, I.; Anand, P.R.; Bhaskar, V. Identity and access management in cloud environment: Mechanisms and challenges. Eng. Sci. Technol. Int. J. 2018, 21, 574–588. [Google Scholar] [CrossRef]
  7. Ailisto, H.; Lindholm, M.; Mäntyjärvi, J.; Vildjiounaite, E.; Mäkelä, S.M. Identifying people from gait pattern with accelerometers. In Proceedings of the Biometric Technology for Human Identification II, Orlando, FL, USA, 28–29 March 2005; pp. 7–14. [Google Scholar]
  8. Derawi, M.O.; Nickel, C.; Bours, P.; Busch, C. Unobtrusive User-Authentication on Mobile Phones Using Biometric Gait Recognition. In Proceedings of the 2010 Sixth International Conference on Intelligent Information Hiding and Multimedia Signal Processing, Washington, DC, USA, 15–17 October 2010; pp. 306–311. [Google Scholar] [CrossRef]
  9. Sha, K.; Kumari, M. Patient Identification Based on Wrist Activity Data. In Proceedings of the 2018 IEEE/ACM International Conference on Connected Health: Applications, Systems and Engineering Technologies, Washington, DC, USA, 26–28 September 2018; Association for Computing Machinery: New York, NY, USA, 2018; pp. 29–30. [Google Scholar]
  10. Wang, C.; Wang, Y.; Chen, Y.; Liu, H.; Liu, J. User authentication on mobile devices: Approaches, threats and trends. Comput. Netw. 2020, 170, 107118. [Google Scholar] [CrossRef]
  11. Zhang, S.; Sun, L.; Mao, X.; Hu, C.; Liu, P.; Herrera, L.J. Review on EEG-Based Authentication Technology. Comput. Intell. Neurosci. 2021, 2021, 1–20. [Google Scholar] [CrossRef]
  12. Burge, M.J.; Bowyer, K.W. Handbook of Iris Recognition; Springer Publishing Company, Incorporated: Cham, Switzerland, 2013. [Google Scholar]
  13. Kavitha, S.; Sripriya, P. A Review on Palm Vein Biometrics. Int. J. Eng. Technol. 2018, 7, 407. [Google Scholar] [CrossRef] [Green Version]
  14. Saevanee, H.; Clarke, N.; Furnell, S.; Biscione, V. Continuous user authentication using multi-modal biometrics. Comput. Secur. 2015, 53, 234–246. [Google Scholar] [CrossRef] [Green Version]
  15. Kumar, R.; Phoha, V.V.; Raina, R. Authenticating users through their arm movement patterns. arXiv 2016, arXiv:cs.CV/1603.02211. [Google Scholar]
  16. Chen, L.; Hoey, J.; Nugent, C.D.; Cook, D.J.; Yu, Z. Sensor-Based Activity Recognition. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 2012, 42, 790–808. [Google Scholar] [CrossRef]
  17. Weiss, G.M.; Yoneda, K.; Hayajneh, T. Smartphone and Smartwatch-Based Biometrics Using Activities of Daily Living. IEEE Access 2019, 7, 133190–133202. [Google Scholar] [CrossRef]
  18. Ahmad, M.; Alqarni, M.; Khan, A.; Khan, A.; Chauhdary, S.; Mazzara, M.; Umer, T.; Distefano, S. Smartwatch-Based Legitimate User Identification for Cloud-Based Secure Services. Mob. Inf. Syst. 2018, 2018, 1–14. [Google Scholar] [CrossRef] [PubMed]
  19. Angrisano, A.; Bernardi, M.L.; Cimitile, M.; Gaglione, S.; Vultaggio, M. Identification of Walker Identity Using Smartphone Sensors: An Experiment Using Ensemble Learning. IEEE Access 2020, 8, 27435–27447. [Google Scholar] [CrossRef]
  20. Mekruksavanich, S.; Jitpattanakul, A. Biometric User Identification Based on Human Activity Recognition Using Wearable Sensors: An Experiment Using Deep Learning Models. Electronics 2021, 10, 308. [Google Scholar] [CrossRef]
  21. Benegui, C.; Ionescu, R.T. Convolutional Neural Networks for User Identification Based on Motion Sensors Represented as Images. IEEE Access 2020, 8, 61255–61266. [Google Scholar] [CrossRef]
  22. Neverova, N.; Wolf, C.; Lacey, G.; Fridman, L.; Chandra, D.; Barbello, B.; Taylor, G. Learning Human Identity From Motion Patterns. IEEE Access 2016, 4, 1810–1820. [Google Scholar] [CrossRef]
  23. He, K.; Sun, J. Convolutional neural networks at constrained time cost. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5353–5360. [Google Scholar] [CrossRef] [Green Version]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  25. Li, Y.; Wang, L. Human Activity Recognition Based on Residual Network and BiLSTM. Sensors 2022, 22, 635. [Google Scholar] [CrossRef] [PubMed]
  26. Ronald, M.; Poulose, A.; Han, D.S. iSPLInception: An Inception-ResNet Deep Learning Architecture for Human Activity Recognition. IEEE Access 2021, 9, 68985–69001. [Google Scholar] [CrossRef]
  27. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef] [Green Version]
  28. Mekruksavanich, S.; Jantawong, P.; Jitpattanakul, A. A Lightweight Deep Convolutional Neural Network with Squeeze-and-Excitation Modules for Efficient Human Activity Recognition using Smartphone Sensors. In Proceedings of the 2021 2nd International Conference on Big Data Analytics and Practices (IBDAP), Bangkok, Thailand, 26–27 August 2021; pp. 23–27. [Google Scholar] [CrossRef]
  29. Mekruksavanich, S.; Jitpattanakul, A. Detection of Freezing of Gait in Parkinson’s Disease by Squeeze-and-Excitation Convolutional Neural Network with Wearable Sensors. In Proceedings of the 2021 15th International Conference on Open Source Systems and Technologies (ICOSST), Lahore, Pakistan, 15–16 December 2021; pp. 1–5. [Google Scholar] [CrossRef]
  30. Mekruksavanich, S.; Jantawong, P.; Charoenphol, A.; Jitpattanakul, A. Fall Detection from Smart Wearable Sensors using Deep Convolutional Neural Network with Squeeze-and-Excitation Module. In Proceedings of the 2021 25th International Computer Science and Engineering Conference (ICSEC), Chiang Rai, Thailand, 18–20 November 2021; pp. 448–453. [Google Scholar] [CrossRef]
  31. Zou, L.; Liu, W.; Lei, M.; Yu, X. An Improved Residual Network for Pork Freshness Detection Using Near-Infrared Spectroscopy. Entropy 2021, 23, 1293. [Google Scholar] [CrossRef]
  32. Park, J.; Kim, J.k.; Jung, S.; Gil, Y.; Choi, J.I.; Son, H.S. ECG-Signal Multi-Classification Model Based on Squeeze-and-Excitation Residual Neural Networks. Appl. Sci. 2020, 10, 6495. [Google Scholar] [CrossRef]
  33. Saini, B.s.; Kaur, N.; Bhatia, K. Authenticating Mobile Phone User using Keystroke Dynamics. Int. J. Comput. Sci. Eng. 2018, 6, 372–377. [Google Scholar] [CrossRef]
  34. Shi, W.; Yang, J.; Jiang, Y.; Yang, F.; Xiong, Y. SenGuard: Passive user identification on smartphones using multiple sensors. In Proceedings of the IEEE 7th International Conference on Wireless and Mobile Computing, Networking and Communications (WiMob), Shanghai, China, 10–12 October 2011; pp. 141–148. [Google Scholar] [CrossRef]
  35. De Luca, A.; Hang, A.; Brudy, F.; Lindner, C.; Hussmann, H. Touch Me Once and i Know It’s You! Implicit Authentication Based on Touch Screen Patterns. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Austin, TX, USA, 5–10 May 2012; Association for Computing Machinery: New York, NY, USA, 2012; pp. 987–996. [Google Scholar] [CrossRef]
  36. Sae-Bae, N.; Memon, N.; Isbister, K. Investigating multi-touch gestures as a novel biometric modality. In Proceedings of the 2012 IEEE Fifth International Conference on Biometrics: Theory, Applications and Systems (BTAS), Arlington, VA, USA, 23–27 September 2012; pp. 156–161. [Google Scholar] [CrossRef]
  37. Frank, M.; Biedert, R.; Ma, E.; Martinovic, I.; Song, D. Touchalytics: On the Applicability of Touchscreen Input as a Behavioral Biometric for Continuous Authentication. IEEE Trans. Inf. Forensics Secur. 2013, 8, 136–148. [Google Scholar] [CrossRef] [Green Version]
  38. Rocha, C.C.; Lima, J.C.D.; Dantas, M.A.R.; Augustin, I. A2BeST: An adaptive authentication service based on mobile user’s behavior and spatio-temporal context. In Proceedings of the 2011 IEEE Symposium on Computers and Communications (ISCC), Corfu, Greece, 28 June–1 July 2011; pp. 771–774. [Google Scholar] [CrossRef]
  39. Sabharwal, M. Multi-Modal Biometric Authentication and Secure Transaction Operation Framework for E-Banking. Int. J. Bus. Data Commun. Netw. 2017, 13, 102–116. [Google Scholar] [CrossRef] [Green Version]
  40. Jakobsson, M.; Shi, E.; Golle, P.; Chow, R. Implicit Authentication for Mobile Devices. In Proceedings of the 4th USENIX Conference on Hot Topics in Security, Montreal, QC, Canada, 10–14 August 2009; USENIX Association: Berkeley, CA, USA, 2009. HotSec’09. p. 9. [Google Scholar]
  41. Casale, P.; Pujol, O.; Radeva, P. Personalization and User Verification in Wearable Systems Using Biometric Walking Patterns. Pers. Ubiquitous Comput. 2012, 16, 563–580. [Google Scholar] [CrossRef]
  42. Rong, L.; Jianzhong, Z.; Ming, L.; Xiangfeng, H. A Wearable Acceleration Sensor System for Gait Recognition. In Proceedings of the 2007 2nd IEEE Conference on Industrial Electronics and Applications, Harbin, China, 23–25 May 2007; pp. 2654–2659. [Google Scholar] [CrossRef]
  43. Mantyjarvi, J.; Lindholm, M.; Vildjiounaite, E.; Makela, S.M.; Ailisto, H. Identifying users of portable devices from gait pattern with accelerometers. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, Philadelphia, PA, USA, 23 March 2005; Volume 2. [Google Scholar]
  44. Parziale, A.; Carmona-Duarte, C.; Ferrer, M.A.; Marcelli, A. 2D vs 3D Online Writer Identification: A Comparative Study. In Document Analysis and Recognition—ICDAR 2021; Lladós, J., Lopresti, D., Uchida, S., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 307–321. [Google Scholar]
  45. Musale, P.; Baek, D.; Werellagama, N.; Woo, S.S.; Choi, B.J. You Walk, We Authenticate: Lightweight Seamless Authentication Based on Gait in Wearable IoT Systems. IEEE Access 2019, 7, 37883–37895. [Google Scholar] [CrossRef]
  46. Yang, J.B.; Nguyen, M.N.; San, P.P.; Li, X.L.; Krishnaswamy, S. Deep Convolutional Neural Networks on Multichannel Time Series for Human Activity Recognition. In Proceedings of the 24th International Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015; AAAI Press: Cambridge, MA, USA, 2015; pp. 3995–4001. [Google Scholar]
  47. Ronao, C.A.; Cho, S.B. Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst. Appl. 2016, 59, 235–244. [Google Scholar] [CrossRef]
  48. Jiang, W.; Yin, Z. Human Activity Recognition Using Wearable Sensors by Deep Convolutional Neural Networks. In Proceedings of the 23rd ACM International Conference on Multimedia, Brisbane, Australia, 26–30 October 2015; Association for Computing Machinery: New York, NY, USA, 2015; pp. 1307–1310. [Google Scholar]
  49. Huang, J.; Lin, S.; Wang, N.; Dai, G.; Xie, Y.; Zhou, J. TSE-CNN: A Two-Stage End-to-End CNN for Human Activity Recognition. IEEE J. Biomed. Health Inform. 2020, 24, 292–299. [Google Scholar] [CrossRef]
  50. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:cs.CV/1505.04597. [Google Scholar]
  51. Zhang, Y.; Zhang, Z.; Zhang, Y.; Bao, J.; Zhang, Y.; Deng, H. Human Activity Recognition Based on Motion Sensor Using U-Net. IEEE Access 2019, 7, 75213–75226. [Google Scholar] [CrossRef]
  52. Mahmud, T.; Sazzad Sayyed, A.Q.M.; Fattah, S.A.; Kung, S.Y. A Novel Multi-Stage Training Approach for Human Activity Recognition From Multimodal Wearable Sensor Data Using Deep Neural Network. IEEE Sens. J. 2021, 21, 1715–1726. [Google Scholar] [CrossRef]
  53. Tang, Y.; Teng, Q.; Zhang, L.; Min, F.; He, J. Layer-Wise Training Convolutional Neural Networks With Smaller Filters for Human Activity Recognition Using Wearable Sensors. IEEE Sens. J. 2021, 21, 581–592. [Google Scholar] [CrossRef]
  54. Yang, Z.; Wang, Y.; Liu, C.; Chen, H.; Xu, C.; Shi, B.; Xu, C.; Xu, C. LegoNet: Efficient Convolutional Neural Networks with Lego Filters. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 09–15 June 2019; Chaudhuri, K., Salakhutdinov, R., Eds.; International Machine Learning Society: Stroudsburg, PA, USA, 2019; Volume 97, pp. 7005–7014. [Google Scholar]
  55. Zeng, M.; Gao, H.; Yu, T.; Mengshoel, O.J.; Langseth, H.; Lane, I.; Liu, X. Understanding and Improving Recurrent Networks for Human Activity Recognition by Continuous Attention. In Proceedings of the 2018 ACM International Symposium on Wearable Computers, Singapore, 8–12 October 2018; Association for Computing Machinery: New York, NY, USA, 2018; pp. 56–63. [Google Scholar]
  56. Barut, O.; Zhou, L.; Luo, Y. Multitask LSTM Model for Human Activity Recognition and Intensity Estimation Using Wearable Sensor Data. IEEE Internet Things J. 2020, 7, 8760–8768. [Google Scholar] [CrossRef]
  57. Aljarrah, A.A.; Ali, A.H. Human Activity Recognition using PCA and BiLSTM Recurrent Neural Networks. In Proceedings of the 2019 2nd International Conference on Engineering Technology and its Applications (IICETA), Al-Najef, Iraq, 27–28 August 2019; pp. 156–160. [Google Scholar] [CrossRef]
  58. Steven Eyobu, O.; Han, D.S. Feature Representation and Data Augmentation for Human Activity Classification Based on Wearable IMU Sensor Data Using a Deep LSTM Neural Network. Sensors 2018, 18, 2892. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  59. Chen, Z.; Jiang, C.; Xiang, S.; Ding, J.; Wu, M.; Li, X. Smartphone Sensor-Based Human Activity Recognition Using Feature Fusion and Maximum Full A Posteriori. IEEE Trans. Instrum. Meas. 2020, 69, 3992–4001. [Google Scholar] [CrossRef]
  60. Chen, Z.; Wu, M.; Gao, K.; Wu, J.; Ding, J.; Zeng, Z.; Li, X. A Novel Ensemble Deep Learning Approach for Sleep-Wake Detection Using Heart Rate Variability and Acceleration. IEEE Trans. Emerg. Top. Comput. Intell. 2021, 5, 803–812. [Google Scholar] [CrossRef]
  61. Cho, K.; van Merrienboer, B.; Gulcehre, C.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP 2014), Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar]
  62. Xu, C.; Chai, D.; He, J.; Zhang, X.; Duan, S. InnoHAR: A Deep Neural Network for Complex Human Activity Recognition. IEEE Access 2019, 7, 9893–9902. [Google Scholar] [CrossRef]
  63. Chen, Z.; Wu, M.; Cui, W.; Liu, C.; Li, X. An Attention Based CNN-LSTM Approach for Sleep-Wake Detection With Heterogeneous Sensors. IEEE J. Biomed. Health Inform. 2021, 25, 3270–3277. [Google Scholar] [CrossRef]
  64. Chen, K.; Yao, L.; Zhang, D.; Wang, X.; Chang, X.; Nie, F. A Semisupervised Recurrent Convolutional Attention Model for Human Activity Recognition. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 1747–1756. [Google Scholar] [CrossRef]
  65. Xia, K.; Huang, J.; Wang, H. LSTM-CNN Architecture for Human Activity Recognition. IEEE Access 2020, 8, 56855–56866. [Google Scholar] [CrossRef]
  66. Alo, U.R.; Nweke, H.F.; Teh, Y.W.; Murtaza, G. Smartphone Motion Sensor-Based Complex Human Activity Identification Using Deep Stacked Autoencoder Algorithm for Enhanced Smart Healthcare System. Sensors 2020, 20, 6300. [Google Scholar] [CrossRef]
  67. Peng, L.; Chen, L.; Ye, Z.; Zhang, Y. AROMA: A Deep Multi-Task Learning Based Simple and Complex Human Activity Recognition Method Using Wearable Sensors. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 2. [Google Scholar] [CrossRef]
  68. Liu, L.; Peng, Y.; Liu, M.; Huang, Z. Sensor-based human activity recognition system with a multilayered model using time series shapelets. Knowl.-Based Syst. 2015, 90, 138–152. [Google Scholar] [CrossRef]
  69. Chen, L.; Liu, X.; Peng, L.; Wu, M. Deep learning based multimodal complex human activity recognition using wearable devices. Appl. Intell. 2021, 51, 1–14. [Google Scholar] [CrossRef]
  70. Shoaib, M.; Bosch, S.; Incel, O.D.; Scholten, H.; Havinga, P.J.M. Complex Human Activity Recognition Using Smartphone and Wrist-Worn Motion Sensors. Sensors 2016, 16, 426. [Google Scholar] [CrossRef] [PubMed]
  71. Shoaib, M.; Scholten, H.; Havinga, P.J.M.; Incel, O.D. A hierarchical lazy smoking detection algorithm using smartwatch sensors. In Proceedings of the 2016 IEEE 18th International Conference on e-Health Networking, Applications and Services (Healthcom), Munich, Germany, 14–16 September 2016; pp. 1–6. [Google Scholar] [CrossRef]
  72. Shoaib, M.; Incel, O.; Scholten, H.; Havinga, P. Smokesense: Online activity recognition framework on smartwatches. In Mobile Computing, Applications, and Services—9th International Conference, MobiCASE 2018; Ohmura, R., Murao, K., Inoue, S., Gotoh, Y., Eds.; Springer: Amsterdam, The Netherlands, 2018; Lecture Notes of the Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering, LNICST; pp. 106–124. [Google Scholar] [CrossRef]
  73. Harris, E.J.; Khoo, I.H.; Demircan, E. A Survey of Human Gait-Based Artificial Intelligence Applications. Front. Robot. AI 2022, 8, 1–28. [Google Scholar] [CrossRef]
  74. Nguyen, B.; Coelho, Y.; Bastos, T.; Krishnan, S. Trends in human activity recognition with focus on machine learning and power requirements. Mach. Learn. Appl. 2021, 5, 100072. [Google Scholar] [CrossRef]
  75. Scheurer, S.; Tedesco, S.; O’Flynn, B.; Brown, K.N. Comparing Person-Specific and Independent Models on Subject-Dependent and Independent Human Activity Recognition Performance. Sensors 2020, 20, 3647. [Google Scholar] [CrossRef] [PubMed]
  76. Reiss, A.; Stricker, D. Introducing a New Benchmarked Dataset for Activity Monitoring. In Proceedings of the 2012 16th International Symposium on Wearable Computers, Newcastle, UK, 18–22 June 2012; pp. 108–109. [Google Scholar] [CrossRef]
  77. Roggen, D.; Calatroni, A.; Rossi, M.; Holleczek, T.; Förster, K.; Tröster, G.; Lukowicz, P.; Bannach, D.; Pirkl, G.; Ferscha, A.; et al. Collecting complex activity datasets in highly rich networked sensor environments. In Proceedings of the 2010 Seventh International Conference on Networked Sensing Systems (INSS), Kassel, Germany, 15–18 June 2010; pp. 233–240. [Google Scholar] [CrossRef] [Green Version]
  78. Mostayed, A.; Kim, S.; Mazumder, M.M.G.; Park, S.J. Foot Step Based Person Identification Using Histogram Similarity and Wavelet Decomposition. In Proceedings of the 2008 International Conference on Information Security and Assurance (ISA 2008), Busan, Korea, 24–26 April 2008; pp. 307–311. [Google Scholar] [CrossRef]
  79. Anguita, D.; Ghio, A.; Oneto, L.; Parra Perez, X.; Reyes Ortiz, J.L. A Public Domain Dataset for Human Activity Recognition using Smartphones. In Proceedings of the Networks, Computational Intelligence and Machine Learning (ESANN 2013), Bruges, Belgium, 24–26 April 2013; pp. 437–442. [Google Scholar]
  80. Kwapisz, J.R.; Weiss, G.M.; Moore, S.A. Activity Recognition Using Cell Phone Accelerometers. SIGKDD Explor. Newsl. 2011, 12, 74–82. [Google Scholar] [CrossRef]
  81. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer Series in Statistics; Springer: Berlin, Germany, 2009. [Google Scholar]
  82. Monti, R.P.; Tootoonian, S.; Cao, R. Avoiding Degradation in Deep Feed-Forward Networks by Phasing out Skip-Connections; ICANN: Los Angeles, CA, USA, 2018. [Google Scholar]
  83. Muqeet, A.; Iqbal, M.T.B.; Bae, S.H. HRAN: Hybrid Residual Attention Network for Single Image Super-Resolution. IEEE Access 2019, 7, 137020–137029. [Google Scholar] [CrossRef]
  84. Bisong, E. Building Machine Learning and Deep Learning Models on Google Cloud Platform: A Comprehensive Guide for Beginners; Apress: New York, NY, USA, 2019. [Google Scholar]
  85. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. arXiv 2016, arXiv:cs.DC/1603.04467. [Google Scholar]
  86. Ismail Fawaz, H.; Forestier, G.; Weber, J.; Idoumghar, L.; Muller, P.A. Deep learning for time series classification: A review. Data Min. Knowl. Discov. 2019, 33, 917–963. [Google Scholar] [CrossRef] [Green Version]
  87. Benavoli, A.; Corani, G.; Mangili, F. Should We Really Use Post-Hoc Tests Based on Mean-Ranks? J. Mach. Learn. Res. 2016, 17, 1–10. [Google Scholar]
  88. Hodges, J.L.; Lehmann, E.L. Rank Methods for Combination of Independent Experiments in Analysis of Variance. Ann. Math. Stat. 1962, 33, 403–418. [Google Scholar] [CrossRef]
  89. Finner, H. On a Monotonicity Problem in Step-Down Multiple Test Procedures. J. Am. Stat. Assoc. 1993, 88, 920–923. [Google Scholar] [CrossRef]
Figure 1. The proposed CHM-UserIden framework for smartwatch-based user identification used in this study.
Figure 1. The proposed CHM-UserIden framework for smartwatch-based user identification used in this study.
Sensors 22 03094 g001
Figure 2. Architecture of the 1D-ResNet-SE.
Figure 2. Architecture of the 1D-ResNet-SE.
Sensors 22 03094 g002
Figure 3. Structure of the residual block.
Figure 3. Structure of the residual block.
Sensors 22 03094 g003
Figure 4. Structure and functionality of the squeeze-and-excitation block.
Figure 4. Structure and functionality of the squeeze-and-excitation block.
Sensors 22 03094 g004
Figure 5. Confusion matrix for a multiclass classification problem.
Figure 5. Confusion matrix for a multiclass classification problem.
Sensors 22 03094 g005
Table 1. A summary of existing literature on user identification based on the sensor data.
Table 1. A summary of existing literature on user identification based on the sensor data.
Work (Year)ClassifierSensorsDevicePerformance (% Accuracy)ContributionNo. of Users
Parziale et al.(2021) [44]Random Forest1 Acc.Smartwatch89.77User identification based on writing activity performed in air98
Mekruksavanich et al. (2021) [20]CNN-LSTM2 Acc. 1 Gyro.Smartphone94.57User identification based on smartphone sensor from dynamic activities (walking, walking upstairs, and walking downstairs)30
Benegui et al. (2020) [21]CNN1 Acc. 1 Gyro.Smartphone90.75User identification based on motion sensor data of tapping on screen motion from smartphone50
Angrisano et al. (2020) [19]Random Forest1 Acc. 1 Gyro.Smartphone93.8User identification based on walking activities using ensemble machine learning32
Weiss et al. (2019) [17]Random Forest1 Acc. 1 Gyro.Smartphone92.7User identification based on simple activities and complex activities using machine learning approaches51
1 Acc. 1 Gyro.Smartwatch71.7
Musale et al. (2019) [45]Random Forest1 Acc. 1 Gyro.Smartwatch91.8User identification based on statistical features and human-action-related features from sensor data51
Ahmad et al. (2018) [18]Decision Tree1 Acc. 1 Gyro. 1 Mag.Smartwatch98.68User identification based on ambulatory activities using machine learning6
Nevero et al. (2016) [22]CNN1 Acc. 1 Gyro. 1 Mag.Smartphone69.41User identification based on walking activities using dense convolutional clockwork RNNs587
Table 2. Characteristics of the selected activity-based datasets.
Table 2. Characteristics of the selected activity-based datasets.
DatasetCategoryActivityDescriptionRaw Sensor DataPercentage
UT-SmokeSimpleSittingSitting649,00014.28%
StandingStanding649,00014.29%
ComplexSmokingSmoking1,298,00028.57%
EatingEating649,00014.29%
DrinkingDrinking1,298,00028.57%
WISDM-HARBSimpleWalkingWalking192,5315.60%
JoggingJogging187,8335.46%
StairsWalking upstairs and downstairs180,4165.24%
SittingSitting195,0505.67%
StandingStanding194,1035.64%
KickingKicking a soccer ball191,5355.57%
ComplexDribblingDribbling a basketball194,8455.66%
CatchPlaying catch a tennis ball187,6845.46%
TypingTyping187,1755.44%
WritingWriting197,4035.74%
ClappingClapping190,7765.55%
TeethBrushing teeth190,7595.54%
FoldingFolding clothes193,3735.62%
PastaEating pasta189,6095.51%
SoupEating soup187,0575.44%
SandwichEating a sandwich190,1915.53%
ChipsEating chips192,0855.58%
DrinkingDrinking from a cup197,9175.75%
Table 3. Performance metrics for a multiclass confusion matrix.
Table 3. Performance metrics for a multiclass confusion matrix.
MetricsFormulas
Accuracy
A c c u r a c y = i = 1 n T P ( C i ) i = 1 n j = 1 n c i j
Recall of class C i
R e c a l l ( C i ) = T P ( C i ) T P ( C i ) + F N ( C i )
Precision of class C i
P r e c i s i o n ( C i ) = T P ( C i ) T P ( C i ) + F P ( C i )
F1-score of class C i
F 1 s c o r e ( C i ) = 2 × P r e c i s i o n ( C i ) × R e c a l l ( C i ) P r e c i s i o n ( C i ) + R e c a l l ( C i )
Recall
R e c a l l = 1 n i = 1 n R e c a l l ( C i )
Precision
P r e c i s i o n = 1 n i = 1 n P r e c i s i o n ( C i )
F1-score
F 1 s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
False Acceptance Rate of class C i
F A R ( C i ) = F N ( C i ) F N ( C i ) + T P ( C i )
False Rejection Rate of class C i
F R R ( C i ) = F P ( C i ) F P ( C i ) + T N ( C i )
Table 4. Identification effectiveness on classifier evaluation of deep learning models using various combinations of sensor data from UT-Smoke dataset.
Table 4. Identification effectiveness on classifier evaluation of deep learning models using various combinations of sensor data from UT-Smoke dataset.
SensorActivityRecognition Achievement of DL Models Using Sensors Data from UT-Smoke Dataset
CNNLSTM1D-ResNet-SE
Accuracy ± SDF1 ± SDEERAccuracy ± SDF1 ± SDEERAccuracy ± SDF1 ± SDEER
Acc.Smoking69.42 (±0.530)69.35 (±0.545)32.41(±1.17)65.47 (±3.885)65.00 (±4.409)31.26 (±1.12)89.06 (±0.560)89.04 (±0.577)13.49 (±0.87)
Eating80.85 (±0.523)80.64 (±0.602)21.22 (±0.70)79.92 (±2.653)79.81 (±3.136)18.35 (±1.26)88.33 (±5.448)88.20 (±5.608)14.60 (±1.29)
Drinking66.41 (±1.574)65.86 (±1.746)33.20 (±1.35)66.20 (±2.704)65.75 (±3.436)43.19 (±1.25)84.81 (±1.052)84.77 (±1.090)15.14 (±1.08)
Inactive14.61 (±0.560)5.54 (±1.068)91.19 (±0.32)14.82 (±0.109)6.77 (±0.564)90.66 (±0.41)14.54 (±0.273)8.01 (±1.120)90.44 (±0.56)
Avg. (all)57.8255.3544.5156.6054.3345.8769.1967.5333.42
Avg. (active)72.2371.9528.9470.5370.1930.9387.4087.3414.41
Gyro.Smoking51.28 (±0.380)51.07 (±0.418)50.53 (±0.69)45.81 (±0.649)45.19 (±1.065)56.99 (±1.86)68.64 (±5.345)68.63 (±4.833)33.74 (±7.32)
Eating67.03 (±0.394)66.36 (±0.431)35.29 (±1.27)61.62 (±1.548)59.92 (±1.717)44.20 (±2.25)80.42 (±4.736)80.77 (±4.417)21.07 (±7.20)
Drinking43.00 (±0.657)42.67 (±0.515)57.71 (±0.54)37.09 (±3.286)34.79 (±3.426)65.06 (±2.30)46.40 (±5.949)45.72 (±6.773)54.97 (±3.14)
Inactive11.49 (±0.650)7.03 (±0.904)92.97 (±0.42)15.08 (±0.168)4.40 (±0.250)91.14 (±0.05)13.11 (±0.705)7.21 (±0.479)90.80 (±0.35)
Avg. (all)43.2041.7859.1339.9036.0864.3552.1450.5850.15
Avg. (active)53.7753.3747.8448.1746.6355.4265.1565.0436.59
Mag.Smoking54.77 (±3.998)53.79 (±4.619)48.86 (±2.92)57.85 (±2.779)56.77 (±3.134)40.62 (±1.43)78.06 (±10.695)78.18 (±10.496)21.49 (±8.35)
Eating76.10 (±3.206)76.10 (±3.163)25.01 (±1.02)78.77 (±2.733)78.34 (±2.959)22.21 (±3.35)89.81 (±3.266)89.85 (±3.194)15.08 (±5.63)
Drinking44.09 (±4.993)42.58 (±6.440)53.45 (±3.64)63.66 (±1.984)63.69 (±2.074)40.68 (±1.62)76.82 (±7.711)76.92 (±7.568)24.07 (±4.91)
Inactive15.41 (±0.006)4.11 (±0.003)90.92 (±0.02)14.59 (±0.230)8.64 (±1.611)90.43 (±0.51)14.34 (±0.343)8.69 (±1.485)89.95 (±0.67)
Avg. (all)47.5944.1554.5653.7251.8648.4964.7663.4137.65
Avg. (active)58.3257.4942.4466.7666.2734.5081.5681.6520.21
Acc.+ Gyro.Smoking71.11 (±0.928)70.92 (±1.067)29.13 (±0.78)69.24 (±2.088)69.29 (±2.119)9.58 (±1.61)91.20 (±0.610)91.19 (±0.612)9.74 (±1.94)
Eating83.27 (±0.420)83.10 (±0.425)17.74 (±1.23)84.12 (±2.734)84.42 (±2.471)15.51 (±0.38)91.65 (±1.781)91.69 (±1.756)8.42 (±1.66)
Drinking67.92 (±0.905)67.58 (±1.027)32.90 (±1.46)65.87 (±3.575)66.06 (±3.352)31.38 (±2.72)84.78 (±2.265)84.73 (±2.367)13.63 (±0.41)
Inactive14.21 (±0.829)6.77 (±1.379)91.48 (±0.34)14.61 (±0.352)8.01 (±0.966)90.96 (±0.11)14.14 (±0.608)8.47 (±1.316)90.24 (±0.65)
Avg. (all)59.1357.0942.8158.4656.9536.8670.4469.0230.51
Avg. (active)74.1073.8726.5973.0873.2618.8289.2189.2010.60
Acc.+Mag.Smoking87.14 (±1.329)87.14 (±1.329)11.82 (±0.98)82.27 (±2.381)82.24 (±2.449)18.64 (±2.91)96.63 (±0.489)96.63 (±0.490)3.07 (±0.51)
Eating96.08 (±0.462)96.07 (±0.449)4.81 (±0.58)93.34 (±3.324)93.40 (±3.181)5.66 (±2.15)97.98 (±0.088)97.96 (±0.082)2.88 (±0.79)
Drinking86.80 (±0.974)86.90 (±0.906)12.66 (±1.00)78.65 (±6.253)78.69 (±6.262)20.23 (±4.56)96.47 (±0.237)96.46 (±0.245)3.41 (±0.36)
Inactive15.26 (±0.181)4.54 (±0.710)90.94 (±0.04)14.81 (±0.197)7.05 (±0.460)89.98 (±0.42)14.54 (±0.481)8.73 (±0.651)8.38 (± 1.82)
Avg. (all)71.3290.0430.0667.2784.7833.6376.4174.954.44
Avg. (active)90.0154.939.7684.7564.9714.8497.0397.023.12
Gyro.+Mag.Smoking60.64 (±1.497)60.54 (±1.645)40.23 (±3.16)61.35 (±2.468)60.59 (±2.604)38.22 (±4.16)86.76 (±3.827)86.79 (±3.738)17.74 (±4.98)
Eating81.14 (±1.512)81.17 (±1.449)20.61 (±1.47)81.67 (±1.875)81.50 (±1.917)20.29 (±1.46)87.24 (±4.701)87.52 (±4.434)7.71 (±1.80)
Drinking49.13 (±16.952)46.86 (±21.449)47.88 (±3.49)63.93 (±2.815)63.79 (±2.591)40.80 (±2.80)78.58 (±2.384)78.60 (±2.304)20.04 (±4.20)
Inactive15.10 (±0.577)4.61 (±0.967)90.91 (±0.01)15.01 (±0.206)6.34 (±1.098)90.63 (±0.23)14.43 (±0.162)8.56 (±1.152)89.93 (±0.45)
Avg. (all)51.5048.3049.9168.9853.0647.4966.7565.3733.86
Avg. (active)63.6462.8636.2444.3968.6333.1084.1984.3015.16
Acc.+Gyro.+Mag.Smoking88.68 (±0.492)88.68 (±0.477)11.60 (±1.10)80.40 (±2.335)80.39 (±2.450)21.50 (±6.94)97.24 (±0.280)97.24 (±0.280)2.85 (±0.46)
Eating95.84 (±1.247)95.83 (±1.246)5.02 (±0.57)95.18 (±0.658)95.18 (±0.613)5.93 (±2.22)98.15 (±0.178)98.13 (±0.179)2.32 (±0.27)
Drinking88.16 (±0.916)88.27 (±0.888)14.15 (±1.65)80.18 (±4.504)80.36 (±4.613)19.53 (±3.39)96.54 (±0.511)96.54 (±0.496)3.39 (±0.22)
Inactive15.34 (±0.136)4.13 (±0.038)90.90 (±0.01)14.89 (±0.325)8.38 (±1.965)90.57 (±0.56)14.62 (±0.330)9.22 (±1.107)89.92 (±0.68)
Avg. (all)72.0169.2330.4267.6666.0834.3876.6475.2824.62
Avg. (active)90.8990.9310.2685.2585.3115.6597.3197.302.85
Table 5. Recognition effectiveness on classifier evaluation of deep learning models using WIDSM-HARB dataset (Acc. and Gyro. sensors).
Table 5. Recognition effectiveness on classifier evaluation of deep learning models using WIDSM-HARB dataset (Acc. and Gyro. sensors).
ActivityIdentification Performance on Classifier Evaluation of DL Models Using WIDSM-HARB Dataset (Acc. and Gyro.).
CNNLSTM1D-ResNet-SE
Accuracy ± SDF1 ± SDEERAccuracy ± SDF1 ± SDEERAccuracy ± SDF1 ± SDEER
Simple MotionWalking68.14 (±1.230)67.41 (±1.295)31.35 (±3.92)78.00 (±3.059)77.30 (±3.504)20.84 (±3.30)93.26 (±3.302)93.35 (±3.131)10.82 (±7.16)
Jogging74.66 (±2.131)73.84 (±2.309)27.69 (±2.16)86.84 (±2.473)86.62 (±2.518)14.74 (±2.32)96.25 (±2.100)96.20 (±2.088)2.43 (±0.49)
Stairs41.84 (±3.244)42.13 (±2.911)56.01 (±2.29)58.04 (±3.625)56.90 (±3.718)44.54 (±2.57)82.83 (±7.179)82.53 (±7.345)13.01 (±8.27)
Sitting68.47 (±1.366)67.62 (±1.484)31.42 (±1.68)67.03 (±2.887)64.80 (±2.883)33.28 (±3.25)71.66 (±4.868)69.90 (±5.767)25.47 (±5.56)
Standing59.47 (±2.546)58.81 (±2.539)41.81 (±1.63)61.53 (±1.138)58.65 (±1.455)38.48 (±2.43)64.12 (±8.317)61.68 (±9.090)36.11 (±4.95)
Kicking37.53 (±2.361)36.92 (±1.994)64.52 (±3.56)46.56 (±2.854)43.75 (±3.225)54.15 (±1.65)84.28 (±5.572)84.36 (±5.362)17.35 (±8.92)
Hand Complex MovementDribbling58.55 (±6.655)57.98 (±6.575)40.81 (±2.00)75.24 (±1.676)74.43 (±1.958)26.65 (±2.45)93.43 (±4.231)93.35 (±4.284)9.43 (±11.91)
Catch52.55 (±3.207)51.03 (±3.199)53.87 (±3.78)64.22 (±4.014)62.49 (±4.472)33.74 (±4.20)94.37 (±4.718)94.31 (±4.806)4.89 (±1.39)
Typing76.98 (±1.429)76.26 (±1.701)22.08 (±1.56)70.74 (±4.555)67.79 (±5.454)28.46 (±2.52)84.69 (±2.886)83.74 (±3.188)18.34 (±3.93)
Writing72.06 (±2.128)71.10 (±2.416)32.28 (±2.80)70.23 (±1.821)68.06 (±2.037)31.26 (±3.88)81.67 (±8.008)81.10 (±8.612)38.34 (±20.69)
Clapping78.59 (±3.466)77.96 (±4.025)17.48 (±3.24)88.89 (±2.398)88.55 (±2.590)12.69 (±1.32)95.99 (±2.523)95.76 (±2.800)3.14 (±2.36)
Teeth68.09 (±2.737)67.14 (±3.101)31.68 (±2.35)68.64 (±4.315)67.29 (±4.261)29.63 (±2.89)95.31 (±1.966)95.16 (±2.079)5.31 (±1.57)
Folding37.80 (±1.622)36.31 (±1.761)59.93 (±2.56)48.84 (±2.396)46.15 (±2.343)51.28 (±1.90)76.12 (±11.614)75.19 (±12.463)22.84 (±8.31)
Pasta56.11 (±2.703)55.01 (±2.727)43.67 (±1.52)62.89 (±2.47361.09 (±2.401)37.88 (±2.83)82.81 (±5.391)82.50 (±5.598)15.31 (±4.55)
Soup64.68 (±2.369)64.21 (±2.512)34.25 (±2.91)71.31 (±2.613)69.99 (±3.035)30.06 (±0.76)88.02 (±6.895)87.92 (±7.091)10.69 (±1.49)
Sandwich50.83 (±2.568)49.43 (±2.252)47.98 (±4.74)56.83 (±1.992)53.25 (±2.512)45.50 (±1.13)78.08 (±1.063)77.83 (±0.953)21.99 (±2.49)
Chips50.52 (±2.702)49.53 (±2.206)52.60 (±2.54)58.42 (±3.183)55.94 (±3.239)38.07 (±2.41)81.20 (±6.487)80.88 (±6.662)16.35 (±2.72)
Drinking60.11 (±2.413)59.51 (±2.591)40.90 (±4.19)61.94 (±2.148)59.57 (±2.682)37.74 (±1.33)81.80 (±1.485)81.38 (±1.675)17.02 (±1.90)
Average59.8359.0140.5766.4564.5934.9484.7784.2916.05
Table 6. Recognition effectiveness on classifier evaluation of deep learning models using WIDSM-HARB dataset (Acc. sensor).
Table 6. Recognition effectiveness on classifier evaluation of deep learning models using WIDSM-HARB dataset (Acc. sensor).
ActivityRecognition Effectiveness on Classifier Evaluation of DL Models Using WIDSM-HARB Dataset (Acc.).
CNNLSTM1D-ResNet-SE
Accuracy ± SDF1 ± SDEERAccuracy ± SDF1 ± SDEERAccuracy ± SDF1 ± SDEER
Simple MotionWalking52.14 (±1.384)52.16 (±1.541)44.87 (±3.05)66.05 (±2.426)64.99 (±3.000)38.45 (±4.01)91.91 (±1.623)91.65 (±1.755)12.01 (±4.26)
Jogging54.80 (±3.332)54.32 (±3.556)46.36 (±3.85)77.74 (±3.976)77.48 (±4.231)25.76 (±2.69)93.42 (±1.980)93.28 (±2.238)7.10 (±2.47)
Stairs32.28 (±2.522)31.01 (±2.780)66.85 (±1.16)45.61 (±3.929)42.67 (±4.630)52.93 (±2.56)78.01 (±8.410)77.24 (±8.826)20.46 (±7.40)
Sitting64.98 (±2.314)64.45 (±2.472)35.50 (±1.93)60.35 (±3.066)57.48 (±3.784)39.59 (±2.06)71.55 (±5.009)70.41 (±5.788)28.85 (±3.33)
Standing61.24 (±1.972)59.61 (±2.221)38.37 (±2.54)56.53 (±2.148)52.08 (±2.355)43.65 (±1.15)62.47 (±3.619)60.96 (±3.654)36.99 (±5.82)
Kicking28.93 (±2.614)28.05 (±2.555)68.49 (±1.20)41.58 (±2.884)38.70 (±3.460)59.30 (±3.39)72.67 (±5.045)71.68 (±5.070)27.05 (±2.53)
Hand Complex MovementDribbling44.34 (±1.771)43.48 (±1.764)54.72 (±1.35)63.01 (±6.270)61.11 (±6.266)38.50 (±1.26)90.96 (±3.829)90.86 (±3.856)12.20 (±14.56)
Catch37.40 (±3.813)36.62 (±3.898)65.13 (±2.53)57.04 (±4.027)55.43 (±4.389)42.97 (±2.62)90.66 (±1.680)90.72 (±1.544)19.85 (±14.15)
Typing70.12 (±7.315)68.85 (±7.755)29.26 (±4.93)68.77 (±2.418)65.68 (±2.560)2.18 (±2.71)77.22 (±5.938)75.81 (±6.881)25.46 (±7.33)
Writing64.71 (±1.891)64.06 (±2.342)37.97 (±3.06)63.23 (±4.084)60.84 (±4.719)37.34 (±3.96)54.76 (±16.852)52.27 (±17.678)35.14 (±10.76)
Clapping71.63 (±5.277)70.78 (±5.572)29.46 (±2.11)79.82 (±3.143)79.16 (±3.505)20.45 (±2.23)93.09 (±2.551)93.00 (±2.561)7.03 (±2.25)
Teeth59.26 (±2.407)58.51 (±2.259)37.73 (±3.23)63.27 (±2.111)61.47 (±1.890)38.69 (±3.05)90.37 (±3.344)90.34 (±3.345)7.91 (±1.38)
Folding36.24 (±3.944)35.30 (±3.842)63.66 (±2.34)44.42 (±0.754)41.62 (±0.719)54.68 (±4.95)79.54 (±3.411)79.24 (±3.461)22.39 (±2.81)
Pasta52.12 (±2.994)51.58 (±2.995)48.38 (±2.25)56.60 (±2.922)54.34 (±3.274)43.61 (±3.34)82.39 (±2.943)81.93 (±2.987)27.56 (±11.70)
Soup59.27 (±3.055)58.43 (±3.125)42.73 (±4.47)60.63 (±4.157)58.12 (±4.833)39.18 (±1.01)82.68 (±6.134)81.78 (±7.039)17.38 (±7.21)
Sandwich50.46 (±2.102)49.33 (±1.517)50.20 (±2.12)53.28 (±1.784)50.06 (±1.866)23.80 (±1.76)74.28 (±1.979)73.70 (±2.101)23.80 (±1.76)
Chips47.76 (±2.208)46.53 (±2.561)55.04 (±3.66)52.29 (±2.488)50.14 (±2.893)24.22 (±2.76)76.91 (±1.069)76.47 (±1.202)24.22 (±2.76)
Drinking55.08 (±2.754)54.03 (±2.571)45.96 (±2.47)56.44 (±1.570)53.36 (±1.842)24.81 (±3.13)74.23 (±3.473)73.21 (±3.614)24.81 (±3.13)
Average52.3851.5147.8259.2653.9336.1279.8479.1421.12
Table 7. Recognition effectiveness on classifier evaluation of deep learning models using WIDSM-HARB dataset. (Gyro. sensor).
Table 7. Recognition effectiveness on classifier evaluation of deep learning models using WIDSM-HARB dataset. (Gyro. sensor).
ActivityRecognition Effectiveness on Classifier Evaluation of DL Models Using WIDSM-HARB Dataset (Gyro.).
CNNLSTM1D-ResNet-SE
Accuracy ± SDF1 ± SDEERAccuracy ± SDF1 ± SDEERAccuracy ± SDF1 ± SDEER
Simple MotionWalking54.72 (±1.465)54.18 (±1.506)45.97 (±1.50)49.69 (±4.723)46.49 (±4.400)51.33 (±4.89)84.74 (±13.344)83.92 (±14.802)7.89 (±3.86)
Jogging58.73 (±2.690)57.80 (±2.572)42.06 (±2.87)61.87 (±4.805)59.95 (±5.992)38.91 (±4.88)92.74 (±11.612)92.15 (±12.797)8.80 (±6.26)
Stairs23.57 (±1.631)22.37 (±1.524)76.48 (±1.59)24.48 (±3.510)21.656 (±2.971)75.56 (±3.52)83.57 (±8.468)83.22 (±8.659)15.12 (±6.85)
Sitting18.53 (±1.513)17.97 (±1.147)82.53 (±1.63)09.57 (±0.354)7.31 (±0.710)91.07 (±0.29)19.61 (±1.308)18.00 (±1.280)81.55 (±2.27)
Standing14.76 (±2.124)14.22 (±2.384)86.09 (±1.98)8.118 (±0.802)05.87 (±1.002)92.83 (±1.15)20.76 (±6.836)19.57 (±7.979)73.52 (±8.14)
Kicking17.07 (±2.259)16.01 (±2.424)83.26 (±1.97)25.31 (±4.754)22.65 (±4.271)75.62 (±4.49)80.47 (±5.510)80.06 (±5.909)18.14 (±3.50)
Hand Complex MovementDribbling49.88 (±1.678)48.88 (±1.400)50.08 (±1.88)60.96 (±4.594)58.01 (±5.165)39.32 (±4.55)91.75 (±11.119)91.51 (±11.570)3.30 (±1.73)
Catch33.93 (±1.015)33.61 (±0.900)66.97 (±0.84)32.37 (±5.311)29.080 (±5.411)68.47 (±5.36)92.457 (±7.628)92.17 (±8.210)3.76 (±0.92)
Typing26.54 (±1.069)25.02 (±1.136)74.55 (±1.22)33.27 (±3.089)29.61 (±3.073)67.45 (±3.26)82.84 (±11.221)82.54 (±11.622)25.63 (±6.83)
Writing23.84 (±1.954)22.86 (±2.152)78.04 (±1.46)34.76 (±5.479)32.16 (±5.968)66.74 (±5.36)58.64 (±15.074)54.18 (±17.788)25.54 (±20.07)
Clapping77.48 (±2.574)77.14 (±2.449)22.94 (±2.61)73.66 (±1.376)71.98 (±1.735)26.95 (±1.46)95.37 (±6.640)95.20 (±6.964)3.77 (±2.49)
Teeth49.38 (±1.549)48.38 (±1.764)50.47 (±1.48)40.86 (±4.84)37.24 (±4.594)59.50 (±4.63)92.65 (±2.852)92.57 (±2.887)11.24 (±5.35)
Folding11.67 (±1.063)11.21 (±0.950)88.86 (±0.97)14.41 (±2.488)12.81 (±2.683)86.88 (±2.37)74.17 (±1.693)73.70 (±2.128)25.83 (±7.36)
Pasta23.67 (±2.899)22.61 (±3.101)77.48 (±2.74)15.32 (±1.698)11.90 (±1.699)85.23 (±1.69)81.90 (±2.340)81.38 (±2.690)14.71 (±2.77)
Soup34.58 (±2.684)33.87 (±2.541)65.77 (±2.71)26.41 (±4.782)22.86 (±4.585)74.54 (±4.90)85.69 (±7.019)85.09 (±7.844)11.14 (±9.50)
Sandwich19.41 (±0.922)18.56 (±0.911)80.71 (±0.69)16.10 (±2.210)12.42 (±2.770)84.23 (±2.25)53.65 (±9.941)51.81 (±10.759)48.61 (±17.12)
Chips18.25 (±2.270)17.57 (±1.829)82.18 (±2.31)14.88 (±4.596)11.26 (±4.318)85.71 (±4.51)73.12 (±10.084)73.25 (±10.004)27.18 (±4.45)
Drinking25.06 (±1.836)23.83 (±1.812)76.07 (±1.79)17.44 (±8.376)14.10 (±8.025)83.43 (±8.37)55.03 (±15.631)53.43 (±16.501)48.35 (±14.77)
Average32.2831.4568.3631.0828.1966.0273.2972.4325.23
Table 8. Wilcoxon test based on the accuracy metrics for model performance comparison on the UT-Smoke dataset.
Table 8. Wilcoxon test based on the accuracy metrics for model performance comparison on the UT-Smoke dataset.
Hand Complex MovementTest ModelsAccuracy ± SDWilcoxon Test
p-Value H 0
Smoking1D-ResNet93.61 (±0.45)0.043reject
1D-ResNet-SE97.24 (±0.28)
Eating1D-ResNet97.19 (±0.21)0.079accept
1D-ResNet-SE98.15 (±0.18)
Drinking1D-ResNet92.71 (±0.51)0.022reject
1D-ResNet-SE96.62 (±0.33)
Table 9. Wilcoxon test based on the accuracy metrics for model performance comparison on the WISDM-HARB dataset.
Table 9. Wilcoxon test based on the accuracy metrics for model performance comparison on the WISDM-HARB dataset.
Hand Complex MovementTest ModelsAccuracy ± SDWilcoxon Test
p-Value H 0
Dribbling1D-ResNet92.95 (±1.07)0.686accept
1D-ResNet-SE93.43 (±4.23)
Catch1D-ResNet94.36 (±0.58)0.691accept
1D-ResNet-SE94.37 (±4.12)
Typing1D-ResNet76.67 (±13.19)0.039reject
1D-ResNet-SE84.69 (±2.89)
Writing1D-ResNet53.30 (±15.65)0.012reject
1D-ResNet-SE81.67 (±8.01)
Clapping1D-ResNet94.78 (±0.93)0.043reject
1D-ResNet-SE95.99 (±2.52)
Teeth1D-ResNet95.00 (±2.66)0.642accept
1D-ResNet-SE95.31 (±1.97)
Folding1D-ResNet75.41 (±2.57)0.633accept
1D-ResNet-SE76.12 (±11.61)
Pasta1D-ResNet81.01 (±2.44)0.345accept
1D-ResNet-SE82.81 (±5.39)
Soup1D-ResNet87.35 (±2.08)0.728accept
1D-ResNet-SE88.02 (±6.90)
Sandwich1D-ResNet77.83 (±2.80)0.043reject
1D-ResNet-SE78.08 (±1.06)
Chips1D-ResNet80.40 (±5.86)0.138accept
1D-ResNet-SE81.20 (±6.49)
Drinking1D-ResNet78.86 (±4.68)0.025reject
1D-ResNet-SE81.80 (±1.49)
Table 10. Friedman aligned ranking test and Finner post-hoc test based on the accuracy metrics of the 1D-ResNet-SE model using different sensor data on the UT-Smoke dataset.
Table 10. Friedman aligned ranking test and Finner post-hoc test based on the accuracy metrics of the 1D-ResNet-SE model using different sensor data on the UT-Smoke dataset.
ActivityModelsFriedman Aligned Ranking TestFinner Post-Hoc Test
p-Value H 0
SmokingAcc.+Gyro.1.1526--
Acc.2.62640.0000021reject
Gyro.4.35190.0000203reject
EatingAcc.+Gyro.1.2728--
Acc.2.61680.0000036reject
Gyro.4.59620.0016952reject
DrinkingAcc.+Gyro.1.0607--
Acc.2.82840.0000001reject
Gyro.4.06070.0000001reject
Table 11. Friedman aligned ranking test and Finner post-hoc test based on the accuracy metrics of the 1D-ResNet-SE model using different sensor data on the WISDM-HARB dataset.
Table 11. Friedman aligned ranking test and Finner post-hoc test based on the accuracy metrics of the 1D-ResNet-SE model using different sensor data on the WISDM-HARB dataset.
ActivityModelsFriedman Aligned Ranking TestFinner Post-Hoc Test
p-Value H 0
DribblingAcc.+Gyro.1.7678--
Acc.3.39410.346accept
Gyro.3.32340.786accept
CatchAcc.+Gyro.1.0606--
Acc.3.88900.043reject
Gyro.3.53550.225accept
TypingAcc.+Gyro.1.9091--
Acc.3.67690.345accept
Gyro.2.89910.686accept
WritingAcc.+Gyro.2.1920--
Acc.3.60620.501accept
Gyro.2.68700.345accept
ClappingAcc.+Gyro.1.3435--
Acc.3.81840.079accept
Gyro.3.32340.893accept
TeethAcc.+Gyro.1.6971--
Acc.2.75770.043reject
Gyro.4.03050.138accept
FoldingAcc.+Gyro.2.9698--
Acc.2.40410.982accept
Gyro.3.11130.502accept
PastaAcc.+Gyro.1.4142--
Acc.3.46480.079accept
Gyro.3.60620.892accept
SoupAcc.+Gyro.1.8384--
Acc.3.53550.041reject
Gyro.3.11120.501accept
SandwichAcc.+Gyro.1.0607--
Acc.3.53550.138accept
Gyro.3.88910.042reject
ChipsAcc.+Gyro.1.4849--
Acc.2.40420.041reject
Gyro.4.59620.015reject
DrinkingAcc.+Gyro.1.1313--
Acc.3.11130.043reject
Gyro.4.24260.021reject
Table 12. The comparison results of the proposed 1D-ResNet-SE model and the previous work model.
Table 12. The comparison results of the proposed 1D-ResNet-SE model and the previous work model.
ActivityIdentification Performance (%Accuracy)
Acc.Gyro.Acc.+Gyro.
Random Forest [17]1D-ResNet-SERandom Forest [17]1D-ResNet-SERandom Forest [17]1D-ResNet-SE
Simple MotionWalking75.1091.9167.0084.7478.9093.26
Jogging75.0093.4274.3092.7482.1096.25
Stairs52.4078.0139.2083.5758.7082.83
Sitting70.4071.5530.1019.6169.3071.66
Standing64.1062.4727.0020.7661.2064.12
Kicking54.3072.6738.3080.4759.8084.28
Average65.2278.3445.9863.6568.3382.07
Hand Complex MovementDribbling72.3090.9674.8091.7580.3093.43
Catch69.1090.6671.3092.4675.4094.37
Typing81.2077.2251.2082.8484.2084.69
Writing79.6054.7647.6058.6479.1081.67
Clapping83.4093.0973.9095.3785.3095.99
Teeth70.0090.3756.3092.6576.1095.31
Folding60.0079.5438.8074.1763.0076.12
Pasta67.2082.3938.1081.9071.6082.81
Soup74.1082.6850.4085.6976.6088.02
Sandwich61.9074.2837.6053.6562.1078.08
Chips62.6076.9138.7073.1262.4081.20
Drinking63.9074.2341.3055.0365.3081.80
Average70.4480.5951.6778.1173.4586.12
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mekruksavanich, S.; Jitpattanakul, A. Deep Residual Network for Smartwatch-Based User Identification through Complex Hand Movements. Sensors 2022, 22, 3094. https://doi.org/10.3390/s22083094

AMA Style

Mekruksavanich S, Jitpattanakul A. Deep Residual Network for Smartwatch-Based User Identification through Complex Hand Movements. Sensors. 2022; 22(8):3094. https://doi.org/10.3390/s22083094

Chicago/Turabian Style

Mekruksavanich, Sakorn, and Anuchit Jitpattanakul. 2022. "Deep Residual Network for Smartwatch-Based User Identification through Complex Hand Movements" Sensors 22, no. 8: 3094. https://doi.org/10.3390/s22083094

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop