Next Article in Journal
A Non-Contact Detection Method for Multi-Person Vital Signs Based on IR-UWB Radar
Next Article in Special Issue
Automatic Gender and Age Classification from Offline Handwriting with Bilinear ResNet
Previous Article in Journal
Bayesian Estimation of Oscillator Parameters: Toward Anomaly Detection and Cyber-Physical System Security
Previous Article in Special Issue
Fitness Movement Types and Completeness Detection Using a Transfer-Learning-Based Deep Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Recognition of Uni-Stroke Characters with Hand Movements in 3D Space Using Convolutional Neural Networks

1
Department of Artificial Intelligence, Pukyong National University, Busan 48513, Korea
2
Softbrain Co. Ltd., Tokyo 103-0027, Japan
3
School of Computer Science and Engineering, The University of Aizu, Fukushima 965-8580, Japan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(16), 6113; https://doi.org/10.3390/s22166113
Submission received: 14 July 2022 / Revised: 12 August 2022 / Accepted: 14 August 2022 / Published: 16 August 2022
(This article belongs to the Special Issue Vision and Sensor-Based Sensing in Human Action Recognition)

Abstract

:
Hand gestures are a common means of communication in daily life, and many attempts have been made to recognize them automatically. Developing systems and algorithms to recognize hand gestures is expected to enhance the experience of human–computer interfaces, especially when there are difficulties in communicating vocally. A popular system for recognizing hand gestures is the air-writing method, where people write letters in the air by hand. The arm movements are tracked with a smartwatch/band with embedded acceleration and gyro sensors; a computer system then recognizes the written letters. One of the greatest difficulties in developing algorithms for air writing is the diversity of human hand/arm movements, which makes it difficult to build signal templates for air-written characters or network models. This paper proposes a method for recognizing air-written characters using an artificial neural network. We utilized uni-stroke-designed characters and presented a network model with inception modules and an ensemble structure. The proposed method was successfully evaluated using the data of air-written characters (Arabic numbers and English alphabets) from 18 people with 91.06% accuracy, which reduced the error rate of recent studies by approximately half.

1. Introduction

Hand gestures along with speech are important tools in human communication. Notably, hand gestures have allowed people to express and describe their thoughts more accurately. In addition, gestures are more compelling in articulating one’s opinions and beliefs, especially when there are difficulties in voice communication, the environment is noisy, or the communication occurs in an environment where silence is required.
In recent decades, studies have used gestures as communication tools between humans and computers. Air writing, where researchers write in the air with hand gestures, is recorded, analyzed, and recognized automatically using computer algorithms. The most popular method for recording hand gestures is the camera-based approach using a pen and an optical camera. Detecting a pen-tip or fingertip is one of the most important tasks in such computer vision-based approaches, as the shape of the written characters is very similar to conventional handwriting on paper. Handwriting recognition on paper is a traditional research topic that has been studied in the last decades. Various methods, such as linear discriminant analysis, multi-layer perceptron, tangent distance, and support vector machine [1,2], have been previously proposed, and the accuracies of handwritten character recognition using convolutional neural networks are higher than 99% in recent papers [3,4].
In [5], the pen tip position was easily obtained with color information, as the color of the pen was fixed in the experiments. In [6], a method to detect hand regions by predefining the color of hands was demonstrated, which requires a controlled environment. Hsieh et al. proposed a method to detect hand trajectory by obtaining the difference between consequent video frames [7].
The detection of the hand or fingertips can be enhanced by substituting the optical camera with a depth sensor. For example, [8] used the software development kit (SDK) of a Kinect sensor to detect the hand position, and [9] calculated the position of the fingertip from the hand position and depth information. In [10], the position of the fingertips was traced, and tapping gestures were recognized using the Kinect sensor to detect the fingertips. Similarly, a hand-tracking method by measuring depth with an ultrasonic transceiver was proposed in [11].
A drawback of the camera- or the depth sensor-based method is that they mostly require users to stand or sit in front of a camera or a sensor. Though head-mounted cameras, proposed in [6], mitigate this limitation to a certain extent, the inconvenience persists.
Another approach to recording hand gestures is using accelerometer/gyro sensors, which provide users freedom of movement. Because accelerometer/gyro sensors are embedded in most present-day smartwatches/bands, users can easily utilize them as input devices. This approach is also stable irrespective of the light conditions.
One of the biggest issues with the accelerometer/gyro sensor-based approach is tracing the position of the sensors in the global xyz coordinates. Because the sensors move or rotate while obtaining the signals, the accelerometer values follow the local coordinates of the sensors. Therefore, [12] proposed a method for the automatic conversion of coordinates; however, it requires a ground truth of the traces, which is not commonly available.
Therefore, the accelerometer/gyro sensor-based methods have lower accuracy than camera-based methods. For example, the maximum accuracies of camera-based approaches are 99.571% for Arabic numbers and 16 directional symbols [7], and 95% for English letters and Arabic numbers [8]. By contrast, the accelerometer/gyro sensor-based methods achieved 79.6% accuracy for Japanese Katakana (46 letters) with the K-nearest neighbor method [13], 88.1% for Japanese Hiragana (46 letters) with the hidden Markov model (HMM) [14], and 89.2% by combining K-NN with dynamic time warping (DTW) techniques in [15].
Another issue with the sensor-based approach is the diversity of the signals obtained among individuals. Because the speed and rotating angles of a hand differ for each person, it is difficult to recognize the signal of an unknown person. For example, recognition accuracy dropped significantly from 89.2% to 83.2% in [15] and from 95.3% to 81.9% in [16] when the system was tested independent of the writer.
This study focused on increasing the accuracy of the air-writing method for this writer-independent case. We proposed a method to process the obtained signal by designing a novel network structure, training method, and boosting algorithm for the trained network model.
The remainder of this paper is organized as follows: Section 2 explains the experimental data and proposed method. The experimental results and related discussions are presented in Section 3. Finally, Section 4 concludes the paper.

2. Materials and Methods

2.1. Device

A Huawei Watch 2 and an Android tablet (Nexus 7) were used to collect the data. The watch has a built-in 3-axis accelerometer and gyro sensors and can transfer the data into the Android tablet using Bluetooth (see Figure 1).
The data were recorded on the tablet as there was insufficient space to record data in the watch. A user wears the watch on a wrist and virtually draws characters with hand movements (see Figure 2).

2.2. Data

Twenty people aged from 19 to 24 participated in the air-writing activity. Each person virtually wrote the characters freely in the air five times each, after having the shape of each character explained to them. There was no guidance on arm posture or speed, and the participants were allowed to rest freely during the experiments. Data from the smartwatch were transferred to the tablet and recorded. Data from two participants were excluded from further evaluation because of recording errors found after the experiments. All participants were right-handed.
Figure 3 shows the Graffiti design of the characters used in this experiment. This was developed for Palm OS-based PDA [17] and was used in the air-writing system in [8]. Because the air-written characters are not visible, the uni-stroke design makes it convenient for users to move their hands to write characters.
The gestures of some characters are the same or slightly different (see Figure 4). The gestures of Arabic numerals “0” and “1” are the same as the gestures of letters “O” and “I,” respectively; the gestures of “4” and “7” are very close to that of “L” and “T,” respectively. Each of the 36 characters were recorded five times by each participant, but the patterns with similar shapes were categorized as a single class during the pattern recognition phase. Therefore, the number of classes of gestures used in the experiment was 32.
The data were transferred and recorded at 40 Hz from the watch to the tablet. The data include accelerometer and gyro values in three axes and the gravity-removed accelerometer values provided by the watch using a linear filter.
An example of the collected data in a tabular form is shown in Figure 5. The first column has the time stamp, the 2nd to 4th columns have the acceleration data, the 5th to 7th columns have acceleration data after gravity removal, and the 8th to 10th columns have the gyro data.
Figure 6 shows examples of a character as graphs written by different writers. The trends in the signals of the same gestures are similar, but there is diversity among writers.

2.3. Preprocessing

The collected data were smoothed using a Savitzky–Golay filter [18]; this was conducted as the precision of the raw data was low because of quantization. The window length and polynomial order were set to 13 and 5, respectively. This also removed high-frequency noise. However, some of the signals were corrupted during the recording phase. Therefore, signals shorter than the median length of the same gesture by each participant were removed from the dataset. A total of 18 out of 3240 signals were removed from 18 participants’ data. The removed signals were distributed over 12 characters, where the maximum number of the removed signals per character was 4 of Arabic digit 3.
Examples of normal and corrupted signals of the accelerometer are shown in Figure 7. The accelerometer values were adjusted (shifted along the y axis) such that the value at the beginning is zero.
Because the obtained signals include the non-movement data, the following algorithm was developed to trim the signals using wavelet transforms:
(1)
Calculate detail coefficients from single-level wavelet transform for each channel of signals;
(2)
Interpolate and smooth with a Savitzky–Golay filter [18] as the length of the coefficient signal is half of the original signal;
(3)
Calculate the standard deviation of the signal for each channel, and set the minimum value among the channels as the threshold for the data;
(4)
Signal regions where the detail coefficients of all the channels are less than or equal to the threshold are marked as non-movement regions;
(5)
Remove the signals in the non-movement regions.
This algorithm is applied for each datum separately. Removing the non-movement region means that the signal changes in this region are ignored and the data points are removed. Figure 8 shows an example of non-movement region removal. The regions at the beginning and end were removed, and only the hand movement regions were retained.
As the lengths of the processed signal varied, all signals were resampled to have the same number of data points. This process was necessary to input the data into a convolutional layer. The resampled signal length was 129, which is the maximum length of the hand movement regions for all the data.
The data were normalized between 0 and 1, which is the recommended input for artificial neural networks. Normalization was conducted for each channel group of accelerometers, linear accelerometers (gravity-removed accelerometers), and gyro values. The minimum and maximum values were calculated for each group and then normalized to between 0 and 1. The minimum and maximum values were then calculated for each air-writing datum.

2.4. Convolutional Neural Network

The proposed network model with inception architecture [19], inspired by GoogLeNet [20], is shown in Figure 9. The two-dimensional (2D) convolutions are substituted by 1D convolutions and the inception modules of GoogLeNet were simplified for our purposes. The inception modules are expected to extract features in different frequency levels. Different sizes of convolutions were located parallelly and the results were concatenated into a tensor after the convolutions. The sizes of the convolution filters are 1, 3, 5 and 7. The number of filters is 4 for the first inception module, and the numbers of the succeeding modules are 8, 16, 32 and 64.
Fully connected layers were connected with the dropout layers to prevent overfitting, as the amount of data is limited in our experiment. Parameters such as the number of filters, filter sizes, and dropout ratios are shown in Figure 9.
The accuracy of artificial neural networks may vary because of the randomness of the training algorithms, especially when the size of data is small. We employed an ensemble model to stabilize and increase the overall accuracy [21,22]. We trained five different models in the training phase and calculated the median values from the five outputs to derive the final output (see Figure 10).
The proposed method was validated independently for writers by training the network with the data of 17 participants and then testing using the data of the last participant. This process was repeated 18 times such that each participant’s data were used as test data at least once.
The number of epochs and batch size were set to 400 and 16, respectively. The ADAM optimizer with a learning rate of 0.001 and the AMSGrad option [23] was used for training the network. Weights were initialized with Xavier method. These parameters and options were determined experimentally.

2.5. User-Independent Evaluation

The proposed method was evaluated in a user-independent manner, following the leave-one-subject-out cross-validation (LOSOCV) approach [24]. In this approach, data of a participant were used for testing, and the other data were used for training (20% of the training data were selected for validation in the training phase). The training and test were repeated 18 times, such that all the data were used for the test. The biggest merit of LOSOCV is that it enables user-independent evaluations. Because the data from the same participant were separated, it can be assumed that the accuracies with a new user’s data would be similar to the results of the evaluation [25].

3. Results

The experiment was conducted on a computer with Intel i9 and Nvidia RTX 3090ti. The GPU memory was 24GB and all the program codes were written in Python with the TensorFlow library.
The proposed method was evaluated with and without the preprocessing trimming algorithm, as it is often reported that CNNs show improved performance without intensive preprocessing algorithms. The accuracy with and without the trimming procedure was 88.27% and 91.06%, respectively, indicating the ineffectiveness of the trimming algorithm in our dataset. This may have been caused by writing speed changes, as removing the non-movement regions was successful in our visual examinations. The convolutional operators of the networks could fail to detect signal changes when the writing speed changes when adjusting the length of the signals after trimming.
The ensemble structure was effective and stable according to the experiment. The accuracy dropped to 89.51% when the ensemble structure was removed from the proposed method and did not improve even with repeated experiments. This is because we utilized the 18-fold validation policy where the accuracies were averaged from 18 different trained models.
Table 1 compares the results of the proposed method and the conventional methods in the literature. This indicates that the proposed method increased the accuracy by utilizing the uni-stroke design and a deep neural network, reducing error rates by approximately half, from 16.8% as reported in [15] to 8.94%. It was reported that the error rate was reduced to 4.4% with a word-level correction algorithm in [15], but this was not reported in Table 1 to compare the accuracies of a character-level recognition.
It is remarkable that the proposed method achieved accuracy higher than 90% with data from accelerometer and gyro sensors in a user-independent manner. Higher accuracies have been reported with depth sensors or optical cameras, but the best accuracies with accelerometer and gyro sensors have been limited to 83.2% when the systems were validated user-independently. The high accuracy was achieved by two factors: the dataset and network structure. We employed the uni-stroke characters designed for PDA, which are simpler than the original alphabets. Because the uni-stroke characters were designed for finger gestures, we expect that it was easier for the participants to be accustomed to air writing. The design of the proposed network structure also affected the accuracy. A simplified inception structure and the sizes and numbers of filters affect the accuracy of the neural networks.
The evaluation was conducted in a user-independent manner to overcome the limitations of the small size of the dataset. The same accuracies are expected to be achieved when new users’ data are included for additional evaluations. Because of the posture variations during the air writing, the accuracies of the user-independent validations were lower than the user-dependent validations [15,16].
Ninety models were trained during the experiment as the five trained models were utilized for each fold. Figure 11 shows the training and validation curves when we trained the same structure of the proposed model five times. The curves of the five models had similar trends. The accuracies of the training and validation increased sharply during the first 100 epochs and were stable after the 200th epoch.
The five models were trained with the same network structure and data, but the predicted outputs were different because of the randomness of the weight initializations. Accuracy without the ensemble algorithm was slightly reduced to 89.51% on average.
Table 2 shows the accuracy changes according to the number of epochs. Accuracy increased from 89.48% to 91.06% as the number of epochs increased from 50 to 400, but it gradually decreased after 400 epochs.
The differences in the accuracies were observed for each character. Table 3 shows the precision, recall, and F1 scores for each character. The means of F1 score, precision, and recall were 90.91%, 91.09% and 90.81%, respectively. Some characters were considered as being in the same categories because they have little difference in gesture (0–O, 1–I, 4–L, and 7–T). Some characters showed relatively lower scores (with a standard deviation of 5.07). Only a single character had an accuracy under 80% (“Z” was 79.78%).
One of the reasons for this low accuracy can be found in the confusion matrix in Figure 12. The main confusion characters are Z–2 (four errors) and Z–7/T (four errors). These errors can be explained by the similar shapes of these character pairs in the 3D acceleration space. For example, D and P have similar acceleration changes, except for the power along the vertical axis when drawing an arc at the end.
Figure 13 shows the misclassified signals of characters together with the correctly classified signals. The misclassified characters generally showed different signal trends. For example, misclassified T and 6 had additional movements in the Y-axis in comparison to the correctly classified characters. This can be caused by unexpected wrist rotation during the air writing or a loosened strap meaning the watch moved freely during the air writing. Some signals were completely different signal shapes as shown in the signals of C. It is expected that the participant may have moved their arm in the wrong direction during the air writing.
Accuracies also varied according to the participants (see Figure 14). The accuracies of the 13 participants were higher than 90%; four participants had accuracies between 80% and 90%, and one (participant #8) had 45.56%, which is approximately half of the average. The reason for this low accuracy is possibly a different method of hand/arm movements during air writing. Accuracy may increase with additional training data for various postures.
It should be noted that the data of two participants were excluded from the experiments because of a recording failure, since similar situations could occur during real-time application in the future. Firstly, this recording issue could be recognized by checking the data communication status, as the data are corrupted by network failure. Secondly, recording failure can be recognized by checking the length of the obtained data. Because recording failure can occur for unknown reasons, we need to check if all the data were transferred by checking the length of the obtained signals.

4. Conclusions

This paper proposed a method for recognizing the air-written characters of English alphabets and Arabic numbers from accelerometer and gyro sensor data. Accuracies of air writing with accelerometer and gyro sensor data have been lower than depth sensors or optical cameras because of data variations among users. This study utilized a uni-stroke design for the characters and a deep neural network structure to solve this problem. The network structure was newly designed by utilizing simplified inception modules. The method was evaluated using user-independent data from 18 people to validate whether the proposed method overcomes the variation problems among users. The proposed user-independent air-writing system achieved 91.06% accuracy, which is higher than what has been reported in previous studies using acetometer/gyro sensor data, reducing the error rates by approximately half.
One of the limitations of the proposed system is the variation in accuracies according to the characters and the participants. Designing new shapes for English alphabets could be a solution because much confusion occurred between characters with similar shapes in the air writing. Combining boosting algorithms such as AdaBoost with the convolutional neural network [27] could be another solution. The use of a larger database for the network training would reduce the error rates for unknown styles of air writing.

Author Contributions

Conceptualization, W.-D.C.; data curation, A.M.; methodology, K.-T.K.; project administration, W.-D.C. and J.S.; software, A.M. and K.-T.K.; supervision, W.-D.C. and J.S.; validation, K.-T.K.; writing—original draft, K.-T.K.; writing—review and editing, W.-D.C. and J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a grant from the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2020R1F1A1077162).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hastie, T.; Simard, P.Y. Metrics and Models for Handwritten Character Recognition. Stat. Sci. 1998, 13, 54–65. [Google Scholar] [CrossRef]
  2. Singh, D.; Khan, M.A.; Bansal, A.; Bansal, N. An Application of SVM in Character Recognition with Chain Code. In Proceedings of the International Conference Communication, Control and Intelligent Systems, CCIS 2015; Institute of Electrical and Electronics Engineers Inc.: New York, NY, USA, 2016; pp. 167–171. [Google Scholar]
  3. Alqudah, A.; Alqudah, A.M.; Alquran, H.; Al-zoubi, H.R.; Al-qodah, M.; Al-khassaweneh, M.A. Recognition of Handwritten Arabic and Hindi Numerals Using Convolutional Neural Networks. Appl. Sci. 2021, 11, 1573. [Google Scholar] [CrossRef]
  4. Bora, M.B.; Daimary, D.; Amitab, K.; Kandar, D. Handwritten Character Recognition from Images Using CNN-ECOC. In Proceedings of the Procedia Computer Science; Elsevier B.V.: Amsterdam, The Netherlands, 2020; Volume 167, pp. 2403–2409. [Google Scholar]
  5. Roy, P.; Ghosh, S.; Pal, U. A CNN Based Framework for Unistroke Numeral Recognition in Air-Writing. In Proceedings of the International Conference on Frontiers in Handwriting Recognition (ICFHR), Niagara Falls, NY, USA, 5–8 August 2018; pp. 404–409. [Google Scholar] [CrossRef]
  6. Sonoda, T.; Muraoka, Y. A Letter Input System Based on Handwriting Gestures. Electron. Commun. Jpn. Part III Fundam. Electron. Sci. (Engl. Transl. Denshi Tsushin Gakkai Ronbunshi) 2006, 89, 53–64. [Google Scholar] [CrossRef]
  7. Hsieh, C.-H.; Lo, Y.-S.; Chen, J.-Y.; Tang, S.-K. Air-Writing Recognition Based on Deep Convolutional Neural Networks. IEEE Access 2021, 9, 142827–142836. [Google Scholar] [CrossRef]
  8. Murata, T.; Shin, J. Hand Gesture and Character Recognition Based on Kinect Sensor. Int. J. Distrib. Sens. Netw. 2014, 2014, 1–6. [Google Scholar] [CrossRef]
  9. Kane, L.; Khanna, P. Vision-Based Mid-Air Unistroke Character Input Using Polar Signatures. IEEE Trans. Hum. -Mach. Syst. 2017, 47, 1077–1088. [Google Scholar] [CrossRef]
  10. Shin, J.; Kim, C.M. Non-Touch Character Input System Based on Hand Tapping Gestures Using Kinect Sensor. IEEE Access 2017, 5, 10496–10505. [Google Scholar] [CrossRef]
  11. Saez-Mingorance, B.; Mendez-Gomez, J.; Mauro, G.; Castillo-Morales, E.; Pegalajar-Cuellar, M.; Morales-Santos, D.P. Air-Writing Character Recognition with Ultrasonic Transceivers. Sensors 2021, 21, 6700. [Google Scholar] [CrossRef]
  12. Otsubo, Y.; Matsuki, K.; Nakai, M. A Study on Restoration of Acceleration Feature Vectors for Aerial Handwritten Character Recognition. In Proceedings of the Forum on Information Technology, Tottori, Japan, 4–6 September 2013; pp. 147–148. [Google Scholar]
  13. Kuramochi, K.; Tsukamoto, K.; Yanai, H.F. Accuracy Improvement of Aerial Handwritten Katakana Character Recognition. In Proceedings of the 2017 56th Annual Conference of the Society of Instrument and Control Engineers of Japan (SICE), Kanazawa, Japan, 19–22 September 2017; pp. 116–119. [Google Scholar] [CrossRef]
  14. Matsuki, K.; Nakai, M. Aerial Handwritten Character Recognition Using an Acceleration Sensor and Gyro Sensor. In Proceedings of the Forum on Information Technology 2011, Hakodate, Japan, 7–9 September 2011; pp. 237–238. [Google Scholar]
  15. Yanay, T.; Shmueli, E. Air-Writing Recognition Using Smart-Bands. Pervasive Mob. Comput. 2020, 66, 101183. [Google Scholar] [CrossRef]
  16. Amma, C.; Gehrig, D.; Schultz, T. Airwriting Recognition Using Wearable Motion Sensors. ACM Int. Conf. Proc. Ser. 2010, 10, 1–8. [Google Scholar] [CrossRef]
  17. Graffiti (Palm OS). Available online: https://en.wikipedia.org/wiki/Graffiti_%28Palm_OS%29#/media/File:Palm_Graffiti_gestures.png (accessed on 8 June 2022).
  18. Schafer, R.W. What Is a Savitzky-Golay Filter? [Lecture Notes]. IEEE Signal Processing Mag. 2011, 28, 111–117. [Google Scholar] [CrossRef]
  19. Szegedy, C.; Vanhoucke, V.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar]
  20. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  21. Livieris, I.E.; Pintelas, E.; Stavroyiannis, S.; Pintelas, P. Ensemble Deep Learning Models for Forecasting Cryptocurrency Time-Series. Algorithms 2020, 13, 121. [Google Scholar] [CrossRef]
  22. Ganaie, M.A.; Hu, M.; Tanveer, M.; Suganthan, P.N. Ensemble Deep Learning: A Review. Eng. Appl. Artif. Intell. 2021, 115. [Google Scholar] [CrossRef]
  23. Reddi, S.J.; Kale, S.; Kumar, S. On the Convergence of Adam and Beyond. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, 30 April –3 May 2018; pp. 1–23. [Google Scholar]
  24. Esterman, M.; Tamber-Rosenau, B.J.; Chiu, Y.C.; Yantis, S. Avoiding Non-Independence in FMRI Data Analysis: Leave One Subject Out. Neuroimage 2010, 50, 572–576. [Google Scholar] [CrossRef] [PubMed]
  25. Gholamiangonabadi, D.; Kiselov, N.; Grolinger, K. Deep Neural Networks for Human Activity Recognition with Wearable Sensors: Leave-One-Subject-Out Cross-Validation for Model Selection. IEEE Access 2020, 8, 133982–133994. [Google Scholar] [CrossRef]
  26. Xu, C.; Pathak, P.H.; Mohapatra, P. Finger-Writing with Smartwatch. In Proceedings of the 16th International Workshop on Mobile Computing Systems and Applications, Santa Fe, NM, USA, 12–13 February 2015; pp. 9–14. [Google Scholar] [CrossRef]
  27. Taherkhani, A.; Cosma, G.; McGinnity, T.M. AdaBoost-CNN: An Adaptive Boosting Algorithm for Convolutional Neural Networks to Classify Multi-Class Imbalanced Datasets Using Transfer Learning. Neurocomputing 2020, 404, 351–366. [Google Scholar] [CrossRef]
Figure 1. Huawei Watch 2 and tablet. The waveforms transferred from the device are displayed and recorded on the tablet.
Figure 1. Huawei Watch 2 and tablet. The waveforms transferred from the device are displayed and recorded on the tablet.
Sensors 22 06113 g001
Figure 2. Example of an air-writing activity.
Figure 2. Example of an air-writing activity.
Sensors 22 06113 g002
Figure 3. Uni-stroke characters (Graffiti in Palm OS). Reprinted with permission from [17].
Figure 3. Uni-stroke characters (Graffiti in Palm OS). Reprinted with permission from [17].
Sensors 22 06113 g003
Figure 4. The pairs of the same or slightly different gestures. Reprinted with permission from [17].
Figure 4. The pairs of the same or slightly different gestures. Reprinted with permission from [17].
Sensors 22 06113 g004
Figure 5. Example of collected data (digit 0) from a smartwatch.
Figure 5. Example of collected data (digit 0) from a smartwatch.
Sensors 22 06113 g005
Figure 6. Acceleration signals of three characters by three participants.
Figure 6. Acceleration signals of three characters by three participants.
Sensors 22 06113 g006
Figure 7. Example of normal (left) and corrupted (right) signals of Arabic numeral 9 written by a participant.
Figure 7. Example of normal (left) and corrupted (right) signals of Arabic numeral 9 written by a participant.
Sensors 22 06113 g007
Figure 8. Example of non-movement region removal for Arabic numeral 9. (a) Accelerometer signal before the removal, (b) signal after the removal.
Figure 8. Example of non-movement region removal for Arabic numeral 9. (a) Accelerometer signal before the removal, (b) signal after the removal.
Sensors 22 06113 g008
Figure 9. Proposed network structure. Conv1D[N].
Figure 9. Proposed network structure. Conv1D[N].
Sensors 22 06113 g009
Figure 10. Ensemble method.
Figure 10. Ensemble method.
Sensors 22 06113 g010
Figure 11. Example of training and validation curves when a model is trained five times.
Figure 11. Example of training and validation curves when a model is trained five times.
Sensors 22 06113 g011
Figure 12. Confusion matrix.
Figure 12. Confusion matrix.
Sensors 22 06113 g012
Figure 13. Examples of misclassified and correctly classified signals of characters.
Figure 13. Examples of misclassified and correctly classified signals of characters.
Sensors 22 06113 g013
Figure 14. Accuracies according to the participants.
Figure 14. Accuracies according to the participants.
Sensors 22 06113 g014
Table 1. Comparison of the studies on air-writing hand gesture recognition.
Table 1. Comparison of the studies on air-writing hand gesture recognition.
ReferenceRecording DevicesAlgorithmsPatternsNumber of ParticipantsAccuracy
[8]KinectDP matchingAlphanumeric characters1095.0% (Alphanumeric character)
98.9% (Arabic digits)
[9]KinectBenchmarksArabic digits, alphabets, and symbols896.5%
[10]KinectTemplate matchingJapanese (Hiragana) and alphabets-94.3%
[7]Optical cameraConvolutional neural network (CNN)Arabic digits and 16 directional symbols1499.571% (writer-independent)
[5]Optical cameraCNNArabic digits2097.7%
[11]Ultrasonic transceiverCNN with long short-term memory (LSTM)Arabic digits (1 to 4) and alphabets (A to D)-98.28%
[6]Wearable optical/infrared cameraDP matchingAlphanumeric characters575.5%
[26]SmartwatchNaive BayesAlphabets190.00% (Naive Bayes)
Logistic regression94.62% (logistic regression)
decision tree88.08% (decision tree)
[16]Wearable motion sensorsHidden Markov model (HMM)Alphabets1095.3% (writer-dependent)
81.9% (writer-independent)
[15]Smart bandsK-nearest neighbor (k-NN) with dynamic-time-warping (DTW)Alphabets5589.2% (writer-dependent, k-NN + DTW)
CNN83.2% (writer-independent, CNN, character level)
[14]Wii Remote ControllerHidden Markov model (HMM)Japanese (Hiragana)588.1%
[13]Laser pointer with accelerometerK-NNJapanese (Katakana)1079.6%
-ProposedCNN with ensemble structureAlphanumeric characters1891.06% (writer-independent)
Table 2. Accuracies with different numbers of epochs.
Table 2. Accuracies with different numbers of epochs.
Epochs50100150200300400500600
Accuracy89.4889.6690.4190.3590.7591.0690.7890.29
Table 3. Accuracies according to the characters.
Table 3. Accuracies according to the characters.
CharacterPrecisionRecallF1 Score
0/O89.8993.8991.85
1/I91.2692.7892.01
288.3784.4586.36
392.0593.1092.57
4/L92.8694.4193.63
594.1288.8991.43
690.8087.7889.27
7/T91.6291.1191.36
893.2692.2292.74
984.0984.0984.09
A82.1892.2286.91
B95.4093.2694.32
C96.6295.5696.09
D85.7180.9083.24
E88.5496.5992.39
F95.6096.6796.13
G96.6797.7597.20
H100.0097.7598.86
J97.83100.0098.90
K100.0097.7598.86
M95.6597.7896.70
N86.9688.8987.91
P85.2385.2385.23
Q92.7785.5689.02
R83.9182.0282.95
S83.8786.6785.25
U86.6786.6786.67
V96.5994.4495.51
W94.0587.7890.80
X91.0191.0191.01
Y92.8687.6490.17
Z78.4981.1179.78
Avg.91.0990.8190.91
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chang, W.-D.; Matsuoka, A.; Kim, K.-T.; Shin, J. Recognition of Uni-Stroke Characters with Hand Movements in 3D Space Using Convolutional Neural Networks. Sensors 2022, 22, 6113. https://doi.org/10.3390/s22166113

AMA Style

Chang W-D, Matsuoka A, Kim K-T, Shin J. Recognition of Uni-Stroke Characters with Hand Movements in 3D Space Using Convolutional Neural Networks. Sensors. 2022; 22(16):6113. https://doi.org/10.3390/s22166113

Chicago/Turabian Style

Chang, Won-Du, Akitaka Matsuoka, Kyeong-Taek Kim, and Jungpil Shin. 2022. "Recognition of Uni-Stroke Characters with Hand Movements in 3D Space Using Convolutional Neural Networks" Sensors 22, no. 16: 6113. https://doi.org/10.3390/s22166113

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop