Next Article in Journal
ReS2tAC—UAV-Borne Real-Time SGM Stereo Optimized for Embedded ARM and CUDA Devices
Previous Article in Journal
Towards 6G IoT: Tracing Mobile Sensor Nodes with Deep Learning Clustering in UAV Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Foot Gesture Recognition Using High-Compression Radar Signature Image and Deep Learning

1
Division of Automotive Technology, Daegu Gyeongbuk Institute of Science and Technology (DGIST), Daegu 42988, Korea
2
Department of Interdisciplinary Engineering, Daegu Gyeongbuk Institute of Science and Technology (DGIST), Daegu 42988, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(11), 3937; https://doi.org/10.3390/s21113937
Submission received: 26 April 2021 / Revised: 30 May 2021 / Accepted: 4 June 2021 / Published: 7 June 2021
(This article belongs to the Section Radar Sensors)

Abstract

:
Recently, Doppler radar-based foot gesture recognition has attracted attention as a hands-free tool. Doppler radar-based recognition for various foot gestures is still very challenging. So far, no studies have yet dealt deeply with recognition of various foot gestures based on Doppler radar and a deep learning model. In this paper, we propose a method of foot gesture recognition using a new high-compression radar signature image and deep learning. By means of a deep learning AlexNet model, a new high-compression radar signature is created by extracting dominant features via Singular Value Decomposition (SVD) processing; four different foot gestures including kicking, swinging, sliding, and tapping are recognized. Instead of using an original radar signature, the proposed method improves the memory efficiency required for deep learning training by using a high-compression radar signature. Original and reconstructed radar images with high compression values of 90%, 95%, and 99% were applied for the deep learning AlexNet model. As experimental results, movements of all four different foot gestures and of a rolling baseball were recognized with an accuracy of approximately 98.64%. In the future, due to the radar’s inherent robustness to the surrounding environment, this foot gesture recognition sensor using Doppler radar and deep learning will be widely useful in future automotive and smart home industry fields.

1. Introduction

In recent years, gesture recognition has been playing an important role in opening new ways of human and machine interaction in a wide variety of applications such as wearable devices [1,2,3], smart phones [4], autonomous vehicles [5,6,7], and health care [8].
Existing sensors for a gesture recognition of movements of hands and arms include ultrasound [9,10,11,12], camera based vision [13,14,15,16,17], and radar [18,19,20,21,22,23,24,25,26,27,28]. Ultrasonic sensors have the advantage of relatively low price, but they have short detection distance. Camera based image sensors are a very common gesture recognition approach that use various popular CNN (Convolution Neural Network)-based deep learning models [13,14,15,16,17]. However, because camera based sensors are strongly dependent on surrounding environment factors like lighting, dust, and so on, it is necessary to design sensors very precisely for out-of-vehicle or outdoor use. Radar has the advantage of being strong against noise signals such as those from moisture, dust, and vibration. In addition, radar sensors are suitable for human motion recognition because they have excellent sensitivity and Doppler resolution to changes of moving objects as well as strong immunity to ambient noise.
Among various types of radar sensor, Doppler radar is a good candidate for gesture recognition because it has high sensitivity to human movement. Based on machine-learning and deep learning techniques involving the time-frequency Doppler spectrum, Doppler radar-based gesture recognition has been intensively studied for static and dynamic hand gestures [21,22,24,25] and arm gestures [23,26]. In conventional supervised machine-learning, well-defined extracted features are essential. Typical machine learning methods are the k-nearest neighbor (KNN), support vector machine (SVM), and random forest methods [29,30,31,32,33]. However, the recognition performance is strongly dependent on the predefined features. On the other hand, deep learning approaches based on multi-layer networks such as the convolutional neural network (CNN) are promising for overcoming such feature selection problems without advance need for feature sets. GoogleNet, AlexNet, VGGNet, ResNet, and DenseNet are good examples of deep learning models [34,35,36,37,38,39,40].
Among various methods of gesture recognition, foot gesture recognition is a very useful tool because it allows simple command and control when the hand cannot be freely used. Existing sensors use simple solutions by means of motion detection. In conventional sensors, unexpected door opening may occur due to the unwanted movement of small animals or objects around the sensor. To overcome this problem, the function has been improved in a way that recognizes the user’s foot movement or intention after user authentication through an additional external device such as a smart key fob or smart phone. Nevertheless, the problem of unpredictable door opening remains because it is difficult to grasp the user’s exact intention or foot gesture. In order to solve this misrecognition problem in a more sophisticated smart opening system, it is necessary to recognize foot gestures only with foot motions without a external device.
Conventional foot gesture recognition is performed using a dual channel surface electromyography (sEMG) wearable band, which has been proposed as a hands-free controller for entertainment applications [41]. In automotive applications, one method involves predicting foot movements toward pedals based on capacitive proximity sensing and the hidden Markov model for robust driver foot tracking and recognition [42,43]. Another method involves detecting only simple kick gestures using Doppler radar and a machine-leaning model as a kick-activation system [27,28]. Existing foot gesture solutions have certain limitations such as inconvenience of contact sensing or confined working zones indoors or inside vehicles.
To overcome these limitations, Doppler radar-based foot gesture recognition has recently attracted attention as a smart trunk opener for smart vehicles and as a smart door opener for smart home applications [44,45,46]. For example, it is very useful to open the trunk of a car or automatically open and close a house front door using foot motion in situations in which hands are unavailable or uncomfortable to use. Unfortunately, so far, Doppler radar-based recognition of various foot gestures is still very challenging.
To the best of our knowledge, no studies have yet deeply dealt with recognition of various foot gestures based on Doppler radar and deep learning. Therefore, in this paper, we propose such a Doppler system for foot gesture recognitions. In addition, through the proposed method, we evaluate the recognition performance of four different foot gestures and examine impacts on recognition of foot gestures of sudden moving objects such as rolling objects, passing animals, and so on. To increase the recognition performance, the deep learning-based recognition requires a vast number of radar images for deep learning training [47,48].
To overcome the requirement of large memory, a newly extracted high-compression radar signature image is utilized for deep learning based foot gesture recognition. The proposed technique creates a new high-compression radar signature by extracting a dominant feature via SVD (Singular Value Decomposition) processing and recognizes four different foot gestures via a deep learning model, without degradation of gesture recognition. Instead of a high-resolution original radar signature, the proposed method improves the required memory efficiency for deep learning training by using a high-compression radar signature.
This paper presents a new method for effectively selecting valid radar data through high memory space efficiency and data visualization in the configuration of a large-capacity model training data set required for gesture recognition using a radar signature and deep learning. In other words, by converting the radar STFT signature transparently mapped to Gray or RGB image instead of dealing with the raw beat signal of the radar output signal, the visibility of the radar signature can be clearly checked, so that valid radar data for various gesture targets can be easily selected and thus, a large radar data set is built. On the other hand, when the STFT radar signature is converted to an image, some signals of the radar signature are lost due to the quantization error. To minimize this quantization error, the STFT radar signature was converted to an RGB image instead of a gray image.
Foot gestures of interest include kicking, swinging, sliding, and tapping; also considered is a moving object, namely a baseball that rolls near the radar sensor and passes it. This object can have a significant impact on foot gesture recognition because the four different foot gestures and the movement of the rolling baseball have very similar movement patterns.
The remainder of this paper is organized as follows. Section 2 presents the signal model of the Doppler radar used for foot gesture recognition. Section 3 presents a new method using a high-compression radar signature image and deep learning model with singular value decomposition scheme for foot gesture recognition. Section 4 details various experiments and the obtained original and reconstructed radar signature images for different foot gestures. In Section 5, experimental recognition results of four different foot gestures are provided. Finally, conclusions are given in Section 6.

2. Signal Model

This section describes the system model of the Doppler radar for foot gesture recognition. Doppler radar can estimate Doppler signatures corresponding to unique foot motions. Particularly, Doppler signals include micro-Doppler frequency components due to movements of non-rigid feet and ankles and main Doppler shift components caused by translational motions of the body.
In a typical Doppler radar, a transmitted (TX) signal is reflected from a foot gesture. The reflected signal is changed into a beat signal at the receiver (RX). The sinusoidal signal’s frequency is sensitive to small movements of the foot. A system model of CW (continuous wave) radar is considered for a single human subject. The CW TX signal is transformed into a continuous wave signal by s tx t = exp j 2 π f c t , where fc is the center frequency.
For the Doppler estimation of a set of L rigid and non-rigid targets with different radial velocities { v m t , for m = 1 ,   ,   L }, foot movements will change the targets’ Doppler frequency shifts. These frequencies correspond to { f m t = 2 f c v m t / C , for m = 1 , , L } where C is the velocity of light. The beat signal of a single frame for a CW radar is obtained by mixing the transmitted and received signals. The beat signal x t at a single frame is represented by
x t = m = 1 L x m , I t I - channel + j m = 1 L x m , Q t Q - channel = m = 1 L a m cos ( 2 π f m t ) x m , I t + j a m sin ( 2 π f m t ) x m , Q t
where x m , I t and x m , Q t are the I-channel and Q-channel components, respectively, of a beat signal consisting of L human’s body motion components due to a foot gesture. a m and f m are the amplitude and the micro-Doppler frequency components, respectively. After passing through an ADC (analog-to-digital converter) with sampling rate of f s = 1 / T s , the digitized beat signal at a single frame is denoted by X n and obtained by substituting t = n T s into   X t , that is, X t = X n T s , where T s is the sampling time interval.
By transforming the digitized beat signal X n via the FFT (Fast Fourier transform), an STFT (Short-Time Fourier Transform) spectrogram X k , f can be determined as
X k , f = n = 0 M 1 X n ω k n e j 2 π f n N
where ω n is a window function and M and N are the number of samples per frame and the point of FFT, respectively.
Equations (1) and (2) ensure that beat signals have different amplitude and frequency components for each foot movement. By extracting these unique radar signatures in the time and frequency domains in a beat signal, different foot movements can be distinguished.

3. New Foot Gesture Recognition Method Using High-Compression Radar Signature Image and Deep Learning Model

Figure 1 shows a new method of using a high-compression radar signature image and deep learning for foot gesture recognition.
The newly proposed foot gesture recognition method is largely composed of a six-step procedure. The entire procedure includes (1) acquisition of beat signal for each foot gesture, (2) calculation of STFT radar spectrogram, (3) conversion of STFT radar signature RGB image, (4) selection of order of eigenvalue, (5) reconstruction of new radar signature image, and (6) deep learning processing for foot gesture recognition.
In the first step, a beat signal x t for different foot gestures is obtained using Equation (1). In the second step, the STFT spectrogram X k , f is extracted from the received beat signal using Equation (2). Then, only the Doppler frequency of interest corresponding to the foot gesture is selected through appropriate frequency filtering in the spectrogram.
In the third step, a radar signature RGB image I is obtained by using RGB transformation and color mapping to convert the STFT spectrogram. To construct radar data sets for deep learning training as easily as possible, visualization of radar data is quite useful. In other words, instead of storing the beat signal (raw data) to select valid data for various foot gestures, the visibility of the radar signal can be easily checked by means of the STFT radar signature RGB image, so it is easy to pick up the valid data corresponding to the short movement. On the other hand, the radar signature signal suffers from the quantization error caused by the conversion of the STFT radar signature to the RGB image.
To provide a constant-resolution analysis in both time and frequency domains for robust classification, the reconstructed radar RGB image shall satisfy shift-invariance property in time and frequency domains because the STFT radar signature satisfies this shift-invariance property. To meet this shift-invariance property, during the compression, the STFT radar signature is converted into an RGB image with a single array proportional to the magnitude through a colormap.
In the fourth step, an SVD (Singular Value Decomposition) technique is performed from the radar signature image I and the order of its eigenvalues for feature extraction is selected. Singular value decomposition (SVD) and principal component analysis (PCA) are two popular eigenvalue methods used to reduce a high-dimensional data set into fewer dimensions while retaining important information. SVD is very similar to PCA because of a data decomposition approach for feature extraction of a signal. SVD is based on a single datapoint while PCA require a vast amount of data. In other words, PCA cannot use a single data to pick up features and cannot provide details of the dominant features. Moreover, different foot gestures are usually hidden in certain specific times and frequencies, so using SVD to extract the dominant features for a single radar signal allows for the collection of more comprehensive information than using PCA.
To extract dominant feature information of a foot movement feature from the original radar image I, the SVD technique is utilized. For convenience, the value of the RGB image, expressed as an array of three numbers, is mapped to an array of a single number via color mapping. Here, a jet colormap is used. The final converted RGB image has a value of one dimension, not three dimensions as obtained via jet colormap. Therefore, the value of the converted radar RGB image I corresponds to the magnitude of the STFT radar signature. Then, the transformed image is scaled to the size of p by q   p > q for input into the deep learning model. Generally speaking, computational complexity uses the number of Floating-point Operation(FLOP). Assuming that we use radar images of size p × q ( p = q = 227) and truncated SVD, the SVD processing requires computational complexity of O( 3 p 3 + p 3 + p ) FLOP.
Singular value decomposition for radar signature RGB image is expressed as
I = U Σ V T
where U and V are the left and right matrix, respectively. The columns of orthogonal matrices U and V are eigenvectors, and Σ = diag [ σ 1 , σ 2 , σ r , σ q ] is the eigen matrix and diag [·] denotes a diagonal matrix operator. That is, U and V, and Σ , are expressed in the following matrix form:
U = [ u 1   u 2     u r     u p ] ,
Σ = [ σ 1 0 0 0 σ 2 0 σ r 0 0 σ q 0 0 0 0 0 0 ] ,
V = [ v 1   v 2     v r     v p ] .
By means of the eigenvalues σ i , and the corresponding eigenvectors u i and v i , the radar signature RGB image is rearranged into the following expression:
I = σ 1 u 1 v 1 T + σ 2 u 2 v 2 T + + σ r u r v r T + + σ q u q v q T ,
where σ i is the eigenvalue and has a descending order, i.e., σ 1 > σ 2 > > σ r > > σ q , and u i and v i are unit orthogonal column vectors corresponding to the eigenvalues. Equation (7) shows that the radar signature RGB image consists of a weighted linear combination of different features.
It is important to determine the eigenvalues and their corresponding eigenvectors so as to include sufficient feature information for each foot movement in the original radar RGB image I. By observing the knee points for all foot gestures, which change rapidly, the threshold of the order of the eigenvalue can be well selected. By means of selection of the order, a unique and sufficient feature can be extracted, and the high-compression radar image is simultaneously created in the radar signature image because the order of the eigenvalues can be small relative to the maximum number q . This allows the size of the radar signature image to be dramatically reduced.
When the STFT radar signature is converted into an RGB image, the size of the radar signature is one-dimensional, whereas the RGB image is separated into three dimensions, so the original radar signature can be distorted. To overcome this problem (to match the dimensions of the radar signature and the RGB image), it is converted into an RGB image that can be expressed as a one-dimensional single value mapped through a colormap.
Figure 2 shows the original radar image, the reconstructed radar image from a RGB image with three dimensional arrays, and the reconstructed radar image from a RGB image with a single dimensional array via a colormap.
In the fifth step, a reconstructed radar signature image with a high-impact feature is generated. Using the selected order r, a new radar signature image is reconstructed by the following equation.
I r = U r Σ r V r T
where U r = [ u 1 u 2 u r ] , Σ r = diag [ σ 1 σ 2 σ r ] , and V r = [ v 1 v 2 v r ] ; r is relatively small compared to q , r q . From Equation (8), a high-compression radar signature images with sufficient features for each foot gesture are newly created.
In the final step, foot gestures are classified by deep learning. The newly proposed radar image was trained and validated for a well-known deep learning model; foot gesture recognition is evaluated through a confusion matrix.

4. Experiments and Signature Images for Doppler Radar-Based Foot Gesture Recognition

Figure 3 shows experimental setup to recognize foot gestures and experimental scenes for obtaining radar signals for different foot gestures. Figure 3a is an experimental setup to collect radar data for foot gestures. Here, the radar sensor used for foot gesture recognition uses a CW modulated waveform. The center frequency is 24 GHz. The maximum detection distance is within about 25 m, with RCS = 0 dBm 2 . Figure 3b,c provide pictures of the utilized antenna and the amplification board of the beat signal, respectively. Here, the horizontal azimuth angle and vertical elevation angle are 80 and 12 degrees, respectively. As can be seen in Figure 3d,e, the radar was installed at a height of about 0.6 m in consideration of a typical deployment location, such as the bumper height of a passenger car or SUV vehicle, or an installation location of a smart door. The radar radiates a transmission signal toward a target. The emitted signal is reflected by movements of the target, that is, foot motion gestures, and returned to the radar. If a received signal is above any predetermined threshold, the received signal is captured via the acquisition system after amplification. Radar STFT signature is transformed by performing a short-time Fourier transform on a PC. Table 1 summarized the used radar parameters for foot gesture recognition.
Figure 4 illustrates different kinds of foot gesture used in this experiment. As can be seen in Figure 4, the four different foot gestures include kicking, swinging, sliding, and tapping; also included is a sudden movement like that of a rolling baseball. Among the different foot movements, kicking extends one foot from the front of the radar in the direction of irradiation and then returns it to the original position; swinging involves one foot moving from a starting a point approximately 45 degrees with respect to the direction of the radiation, to a point approximately 45 degrees along a curved line; sliding involves moving one foot from one side to another in a straight line; tapping involves moving foot up and down, slowly raising the foot in front of the radar and then rapidly bringing it down; the moving object motion is of a baseball rolling near the radar.
The radar signatures are specific and unique and can be distinguished for the different foot gestures. Here, clutter components caused by various stationary obstacles surrounding the radar and DC components due to leaky signals are removed by pre-signal processing before performing a STFT processing. In addition, beat signal is sampled at a sampling rate of 10 KHz for each time window of 409.6 ms. During the STFT processing, time window of 409.6 ms has four sub-time intervals of 102.4 ms. A 10 kHz sampling for 409.6 ms will produce data of 4096 points. The sampled data of 4096 points becomes data size of 1024 × 3073 by doing STFT operation having the parameters such as the point of FFT of 1024 and the size of overlap of 1023. The STFT radar image is finally extracted at a frame rate of 4.9 fps.
To configure the data set for each foot gesture, the STFT radar spectrogram was transformed into an RGB image. The STFT radar spectrogram has a size of 1024 × 3073. Only the frequency shift of a foot gesture in the STFT radar spectrogram is extracted. From various experimental results, Doppler frequencies caused by foot gestures are mainly in the range of −250 Hz to 250 Hz. The Doppler frequency of interest for foot gesture recognition is marginally chosen from 500 Hz to 500 Hz. Therefore, the STFT radar spectrogram of 1024 × 3073 size is converted to an original RGB image of 105 × 3073 size. Therefore, radar spectrum with size of 1024 × 3073 was converted to a 105 × 3073 RGB image, and the converted RGB image is rescaled to an RGB image of 227 × 227, which becomes the input for the deep learning model. Here, to convert the STFT radar spectrogram into an RGB image, the STFT radar spectrogram was transformed into an RGB image using a jet colormap.
Figure 5 shows radar signatures in the time and frequency domains and the STFT spectrogram for four different foot gestures and one rolling ball. In Figure 5, each row represents a radar signature for a specific recognition target. The first, second, and third columns show radar signals in the time domain, radar signals in the frequency spectrum, and the STFT spectrogram for each target, respectively. The STFT spectrogram is obtained by performing STFT processing with the following Kaiser window function [49,50].
ω n = I 0 β 1 n N / 2 N / 2 2 I 0 β ,   for   n = 0 , 1 , N 1
where I 0 · is the modified Bessel function of the first kind with an order of zero and β = 5 . Because the time window of STFT uses the Kaiser function, which is a unit energy window function, the STFT radar signature is a linear and shift-invariant distribution in time and frequency. Because the STFT radar signature is magnitude-wise converted into a RGB image via a colormap, the reconstructed STFT radar image satisfies shift-invariance property in time and frequency domains. Thus, it provides a constant-resolution analysis in both time and frequency domains for robust classification.
Figure 6 shows the original STFT radar RGB images for the four different foot gestures. As can be seen in Figure 4 and Figure 5, there are clear distinguishing features in the time domain, in the frequency domain, and in the STFT spectrogram for each target. Looking at the symmetry of the time domain and frequency spectrum, kicking and sliding have symmetrical properties, while swinging and tapping have asymmetric characteristics. In addition, from the point of view of bandwidth, kicking, swinging, and tapping have relatively wider bandwidths, while sliding and the rolling baseball exhibit relatively narrower bandwidth due to their different kinetic mechanisms. For example, sliding has a relatively low amplitude and many sidelobes in the frequency domain compared to kicking. Because the movement direction of swinging is almost perpendicular to the direction of radiation, the radial component of the Doppler frequency shift is relatively small. Figure 7 shows original STFT radar signature images for five different gestures measured from the radar installed at another height of about 1.5 m.
Kicking, swinging, and sliding involve motions of the foot approaching and moving away from the radar. Kicking starts with taking the foot off the floor and finishes with a motion of putting the foot back on the floor after fully extending the foot. According to the STFT spectrogram image, the kinetic mechanism allows the sign of the Doppler frequency shift to change from negative to positive and the received signal is stronger than those during swinging or sliding.
In cases of both swinging and sliding, the STFT spectrogram exhibits similar frequency bands but slightly different levels of received power. The strongest received signal is the moment when the foot approaches the radar most closely.
The radar signature in the time domain, the radar signature in the frequency domain, and the STFT spectrum for tapping motion were measured by taking a motion of slowly raising and suddenly lowering the instep. This motion has a unique STFT spectrogram that can be divided into an almost unchangeable and a relatively narrower frequency band in a relatively long interval and a relatively strong and wider frequency band in a very short interval.
The rolling ball has a distinguishable feature in that the STFT spectrogram has several narrow frequency bands due to the ball’s moving speed as well as the ball’s rotational speed during measurement.
Figure 8 shows the eigenvalue and cumulative energy of original radar signature RGB images for different foot gestures. From Figure 8, as expected, knee points for all foot gestures that change rapidly are observed; each foot gesture has different knee points, meaning that each foot gesture has a respective domain energy.
Figure 9 shows the original and reconstructed radar signature RGB images according to the different compression ratios for the kicking gesture. Here, the relationship between the order of eigenvalue to be selected and the compression ratio is calculated by
r p + q = p q 1 ϵ
where r is the order of eigenvalue to be selected and p and q are the dimensions of the original image matrix and ϵ is the compression ratio of the original image to be reconstructed. Figure 9a is the original radar signature RGB image and Figure 9b,d are the reconstructed radar images with different compression ratios ϵ = 90%, 95%, and 99%. Here, the selected order of the eigenvalue r is 11, 6 and 1, respectively.
Figure 10 shows the reconstructed radar signature RGB images according to the different foot gestures and the movement of a moving object all with the same compression ratio of ϵ = 99%. Figure 10 corresponds to kicking, swinging, sliding, and tapping movements and the movement of the rolling baseball. Figure 10 provides reconstructed radar signature RGB images for each different foot gesture, which are slightly distinguishable.

5. Recognition Results

The original radar signature RGB images for the four different foot gestures and one rolling ball were acquired with 700 images for each target. The total number of original radar images was approximately 3500. To acquire radar images for each foot movement, one woman and four men took part in this experiment. Among the acquired radar images, 600 for each foot gesture were used in a ratio of 90:10 for training and cross-validation for deep learning processing; 100 images for each foot gesture were used to evaluate foot gesture recognition.
To select a learning framework to evaluate the recognition performance of the proposed method, comparative experiments were performed on five different learning models including PCA based SVM, because PCA based SVM method is widely used in the motion classification. In this paper, the learning frameworks considered are GoogleNet, ResNet, VGG, AlexNet, and PCA based SVM [34,39,40,51,52,53]. Comparative recognition performance is summarized in Table 2. In this case, the recognition indicators are recall, precision, F1, and accuracy. Each performance indicator is defined like that recall = TP/(TP + FN), precision = TP/(TP + FP), F1 = 2 (recall × precision)/(recall + precision), and accuracy = (TP + TN)/(TP + FP + FN + TN). Here, TP, TN, FP, and FN are abbreviated as true positive, true negative, false positive, and false negative, respectively. Model training was performed on a PC with Intel (R) core (TM) i7-9800X CPU core, clock speed of 3.80 GHz, RAM of 64 GB, and two nvidia titan GPU computational performance. As can be seen from Table 2, the foot gesture recognition using the proposed high compression method was compared with the others for five different models. Five different learning models have very good performance. In this paper, the AlexNet model was selected among several deep learning models by considering the learning time and recognition performance.
Table 3 shows the accuracy and loss after training and cross-validation based on the AlexNet deep learning model for various radar signature images. With respect to different compression ratios, radar signature images are used to evaluate the recognition performance. Case # 1 is training results using original image; case # 2 uses 90% compressed radar images; cases # 3 and # 4 correspond to radar signature images having 95%, and 99% compression ratios, respectively. Here, the utilized validation frequency is 50; the maximum epochs and size of the mini batches are 30 and 128, respectively. The AlexNet deep learning model was found to have a good accuracy of above 95% even when using radar signature images with compression ratio of 95%.
Figure 11 shows foot gesture recognition performance obtained using the proposed method in cases of original and reconstructed radar images for both training and test evaluation. Here, four different foot gestures of kicking, swinging, sliding, and tapping are considered as well as the movement of an unpredictable object that may affect foot gesture recognition, a rolling baseball.
As can be seen in Figure 11, the confusion matrix represents the recognition performance of the four different foot gestures using the proposed radar signature image and AlexNet deep learning model. The horizontal and vertical axes in the confusion matrix are the input and output, respectively, of the AlexNet deep learning model.
As can be seen in Figure 11a, all different gestures have good recognition recall of about 96.4% and false recognition rate of about 3.6%. Among the four different foot gestures, tapping has the best recognition, with a recall over 98%. Compared to the other foot gestures, tapping has a short, strong, and clear feature in the radar signature image. On the other hand, sliding, swinging, and the rolling baseball were more easily confused. Sliding and swinging have considerable similarities in kinetic mechanism but a subtle difference that makes it possible to distinguish them. This small difference is caused by the linear and curved movement patterns. Further, sliding and the rolling ball show nearly the same linear movement pattern. There is a subtle difference between the two movements. Sliding involves just linear motion, without up-and-down bouncing, while the rolling ball suffers from up-and-down motion and linear motion simultaneously. This subtle difference makes it possible to clearly distinguish them.
Figure 11b,d show the performance of foot gesture recognition in the case of the proposed high-compression radar images for both training sets. In Figure 11b,c, reconstructed radar images with compression ratios of 90% and 95%, respectively, are used; in Figure 11d, reconstructed radar images with compression ratio of 99% are used.
Figure 11b,d show that the four different foot gestures have nearly identical recognition recall of approximately 96% when comparing case using reconstructed radar image with a compression ratio of 90% and case of using original radar image. The SVD processing extracts the dominant signal components corresponding to large eigenvalues and removes some background noise. So, the accuracy after compression is a little increased due to a denoising effect of the SVD processing.
Figure 11c shows the case of a reconstructed radar image with compression ratio of 95%: the recognition recall is a very high 92%. Among the four different foot gestures, kicking and tapping have especially excellent recall of more than 97%.
In Figure 11d, for a reconstructed radar image with compression rate of 99%, the recognition recall is slightly worse at approximately 90%. Especially, swinging and sliding gestures have recognition recall values of approximately 85%.
To assess the quality of recognition and a more accurate understanding, we depict comparative recognition results for each foot gesture based on reconstructed radar images with different compression ratios graphically and numerically.
Figure 12 shows graphically probabilistic performance of radar sensor’s foot gesture recognition according to original and reconstructed radar images. Table 4 shows more detail of numerically probabilistic performance. Here, recall, precision, and F1 are used as recognition performance for different target gestures.
Figure 13 shows foot gesture recognition performance when compressed and original radar signature images are used as data sets for training and testing, respectively. Since radar-based deep learning for gesture recognition requires high-volume images for training, it is necessary to use high-compression, low-capacity radar images to improve the memory efficiency. When a highly compressed radar image is used for training and a high-resolution original radar image is used as input of deep learning model, this has a considerable impact on recognition performance of radar-based foot gestures.
In Figure 13a,b, radar images with 90% and 95% compression ratios, respectively, are used for deep learning training. Figure 13a shows recognition recall of foot gestures of approximately 96% when reconstructed radar images with compression ratio of 90% are used for training and original radar image is used for test evaluation.
When radar images with different compression ratios are used for training and testing, the recognition performance of foot gestures is almost the same as in cases using all original radar images and all compressed radar images with compression ratio of 90%, respectively, for training and testing.
Figure 13b shows that when reconstructed radar images with compression ratio of 95% are used for training and original radar images are used for test evaluation, the recognition recall of all foot gestures is approximately 92%. In this case, the recognition performance of foot gestures is almost the same as in the case in which all reconstructed radar images with compression ratio of 95% are used for both training and testing.
Figure 13c shows that when reconstructed radar images with compression ratio of 99% are used for training and original radar images are used as input, the recognition recall of all the foot gestures is approximately 63%. In Figure 13c the recognition recall values for kicking and tapping are 95%, and 83%, respectively. The recognition recall for the rolling ball is approximately 97%. On the other hand, the recognition recall of swinging and sliding were 19% and 23%, respectively, values showing considerable deterioration. From Figure 13, radar-based deep learning for foot gesture recognition can achieve good recognition performance and high memory saving by using as training images radar images at compression ratio of about 90%. Regardless of the use of either compressed or original radar images as input, the recognition recall of foot gestures is over 92%.
Figure 14 shows the required memory capacity versus the number of training data according to different kinds of radar signature images for foot gesture recognition using radar-based deep learning model. Under the assumption that one pixel image is one byte, the utilized original radar RGB image has approximately 151 Kbytes. As can be seen in Figure 14, compared to using the original radar images, the required memory capacity can be considerably reduced (approximately 19 times) by using radar images with compression ratio of 95%. In other words, if we use a million radar RGB images for deep learning training, memory requirement will be approximately 144 GB. On the other hand, if we use reconstructed radar images with 95% compression ratio, memory of about 8 GB will be required for the same number of images.

6. Conclusions

In this paper, we propose a new technique for foot gesture recognition using an SVD based high-compression radar image and a deep learning model. Here, four different foot gestures including kicking, swinging, sliding, and tapping, and the movement of a rolling baseball as an example of something other than a foot movement, were considered. The Doppler radar signature in the time domain, the Doppler radar signature in the frequency domain, and the STFT spectrogram for four different foot gestures were measured and unique radar signatures corresponding to each foot gesture were obtained. In addition, to improve the required memory efficiency, a reconstructed radar signature image with high-compression and a unique dominant feature is created by appropriately selecting high-impact eigenvalues through an SVD processing; the newly proposed radar signature image is used for both training and as input for the deep learning model for foot gesture recognition. Finally, to evaluate the foot gesture recognition performance, AlexNet, a CNN-based deep learning model, was used. The four different foot gestures included kicking, swinging, sliding, and tapping. Original and reconstructed corresponding radar images with high compression values of 90%, 95%, and 99% were used for the AlexNet based deep learning model. As for experimental results, in the case of using a high-compression (95%) radar signature image, all four different foot gestures and the rolling baseball were found to have nearly the same recognition accuracy of approximately 98.64%, similar to that obtained in the original high-resolution radar images.
In the future, due to the radar’s inherent robustness to surrounding environments, foot gesture recognition sensors using Doppler radar and deep learning will be widely useful in future automotive and smart home industry fields such as smart trunk openers, smart door openers, and so on.

Author Contributions

All authors conceived and designed the system and experiments together; S.S. and J.L. mainly performed the simulations and experiments. B.K., S.K., and J.L. mainly analyzed the results. S.S. wrote the manuscript and J.L. created the main idea and concept. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the DGIST R&D Program of the Ministry of Science, ICT and Future 283 Planning, Korea (20-IT-02).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Muezzinoglu, T.; Karakose, M. An Intelligent Human–Unmanned Aerial Vehicle Interaction Approach in Real Time Based on Machine Learning Using Wearable Gloves. Sensors 2021, 21, 1776. [Google Scholar] [CrossRef]
  2. Song, X.; Liu, X.; Peng, Y.; Xu, Z.; Liu, W.; Pang, K.; Wang, J.; Zhong, L.; Yang, Q.; Meng, J. A Graphene-Coated Silk-Spandex Fabric Strain Sensor for Human Movement Monitoring and Recognition. Nanotechnology 2021, 32. [Google Scholar] [CrossRef]
  3. Zhao, T.; Liu, J.; Wang, Y.; Liu, H.; Chen, Y. Towards Low-Cost Sign Language Gesture Recognition Leveraging Wearables. IEEE Trans. Mob. Comput. 2021, 20. [Google Scholar] [CrossRef]
  4. Li, T.; Shi, C.; Li, P.; Chen, P. A Novel Gesture Recognition System Based on CSI Extracted from a Smartphone with Nexmon Firmware. Sensors 2021, 21, 222. [Google Scholar] [CrossRef]
  5. Neethu, P.S.; Suguna, R.; Palanivel, R.S. Performance Evaluation of SVM-Based Hand Gesture Detection and Recognition System Using Distance Transform on Different Data Sets for Autonomous Vehicle Moving Applications. Circuit World 2021. [Google Scholar] [CrossRef]
  6. Qian, X.; Ju, W.; Sirkin, D.M. Aladdin’s Magic Carpet: Navigation by in-Air Static Hand Gesture in Autonomous Vehicles. Int. J. Hum. Comput. Interact. 2020, 36. [Google Scholar] [CrossRef]
  7. Kim, B.S.; Jin, Y.; Lee, J.; Kim, S. Low-Complexity Music-Based Direction-of-Arrival Detection Algorithm for Frequency-Modulated Continuous-Wave Vital Radar. Sensors 2020, 20, 4295. [Google Scholar] [CrossRef] [PubMed]
  8. Mahmoud, N.M.; Fouad, H.; Soliman, A.M. Smart Healthcare Solutions Using the Internet of Medical Things for Hand Gesture Recognition System. Complex Intell. Syst. 2020. [Google Scholar] [CrossRef]
  9. Zeng, Q.; Kuang, Z.; Wu, S.; Yang, J. A Method of Ultrasonic Finger Gesture Recognition Based on the Micro-Doppler Effect. Appl. Sci. 2019, 9, 2314. [Google Scholar] [CrossRef] [Green Version]
  10. Ibrahim, E.A.; Geilen, M.; Li, M.; de Gyvez, J.P. Multi-Angle Fusion for Low-Cost Near-Field Ultrasonic in-Air Gesture Recognition. IEEE Access 2020, 8. [Google Scholar] [CrossRef]
  11. Zhou, F.; Li, X.; Wang, Z. Efficient High Cross-User Recognition Rate Ultrasonic Hand Gesture Recognition System. IEEE Sens. J. 2020, 20. [Google Scholar] [CrossRef]
  12. Saez, B.; Mendez, J.; Molina, M.; Castillo, E.; Pegalajar, M.; Morales, D.P. Gesture Recognition with Ultrasounds and Edge Computing. IEEE Access 2021. [Google Scholar] [CrossRef]
  13. Barbhuiya, A.A.; Karsh, R.K.; Jain, R. CNN Based Feature Extraction and Classification for Sign Language. Multimed. Tools Appl. 2021, 80. [Google Scholar] [CrossRef]
  14. Wang, Y.; Yang, Y.; Zhang, P. Gesture Feature Extraction and Recognition Based on Image Processing. Traitement Signal 2020, 37. [Google Scholar] [CrossRef]
  15. Gangrade, J.; Bharti, J. Vision-Based Hand Gesture Recognition for Indian Sign Language Using Convolution Neural Network. IETE J. Res. 2020. [Google Scholar] [CrossRef]
  16. Rezende, T.M.; Almeida, S.G.M.; Guimarães, F.G. Development and Validation of a Brazilian Sign Language Database for Human Gesture Recognition. Neural Comput. Appl. 2021. [Google Scholar] [CrossRef]
  17. Gadekallu, T.R.; Alazab, M.; Kaluri, R.; Maddikunta, P.K.R.; Bhattacharya, S.; Lakshmanna, K.; Parimala, M. Hand Gesture Classification Using a Novel CNN-Crow Search Algorithm. Complex Intell. Syst. 2021. [Google Scholar] [CrossRef]
  18. Smith, J.W.; Thiagarajan, S.; Willis, R.; Makris, Y.; Torlak, M. Improved Static Hand Gesture Classification on Deep Convolutional Neural Networks Using Novel Sterile Training Technique. IEEE Access 2021, 9. [Google Scholar] [CrossRef]
  19. Lee, H.R.; Park, J.; Suh, Y.J. Improving Classification Accuracy of Hand Gesture Recognition Based on 60 GHz FMCW Radar with Deep Learning Domain Adaptation. Electronics 2020, 9, 2140. [Google Scholar] [CrossRef]
  20. Zhang, K.; Lan, S.; Zhang, G. On the Effect of Training Convolution Neural Network for Millimeter-Wave Radar-Based Hand Gesture Recognition. Sensors 2021, 21, 259. [Google Scholar] [CrossRef]
  21. Yu, M.; Kim, N.; Jung, Y.; Lee, S. A Frame Detection Method for Real-Time Hand Gesture Recognition Systems Using CW-Radar. Sensors 2020, 20, 2321. [Google Scholar] [CrossRef] [Green Version]
  22. Wang, Z.; Li, G.; Yang, L. Dynamic Hand Gesture Recognition Based on Micro-Doppler Radar Signatures Using Hidden Gauss-Markov Models. IEEE Geosci. Remote Sens. Lett. 2021, 18. [Google Scholar] [CrossRef]
  23. Wang, Z.; Yu, Z.; Lou, X.; Guo, B.; Chen, L. Gesture-Radar: A Dual Doppler Radar Based System for Robust Recognition and Quantitative Profiling of Human Gestures. IEEE Trans. Hum. Mach. Syst. 2021, 51. [Google Scholar] [CrossRef]
  24. Kim, Y.; Toomajian, B. Hand Gesture Recognition Using Micro-Doppler Signatures with Convolutional Neural Network. IEEE Access 2016, 4. [Google Scholar] [CrossRef]
  25. Skaria, S.; Al-Hourani, A.; Lech, M.; Evans, R.J. Hand-Gesture Recognition Using Two-Antenna Doppler Radar with Deep Convolutional Neural Networks. IEEE Sens. J. 2019, 19. [Google Scholar] [CrossRef]
  26. Zeng, Z.; Amin, M.G.; Shan, T. Arm Motion Classification Using Time-Series Analysis of the Spectrogram Frequency Envelopes. Remote Sens. 2020, 12, 454. [Google Scholar] [CrossRef] [Green Version]
  27. Ehrnsperger, M.G.; Brenner, T.; Hoese, H.L.; Siart, U.; Eibert, T.F. Real-Time Gesture Detection Based on Machine Learning Classification of Continuous Wave Radar Signals. IEEE Sens. J. 2021, 21. [Google Scholar] [CrossRef]
  28. Shankar, Y.; Santra, A. Valid Kick Recognition in Smart Trunks Based on Hidden Markov Model Using Doppler Radar. In Proceedings of the 2019 International Radar Conference (RADAR), Toulon, France, 23–27 September 2019. [Google Scholar]
  29. Callaghan, D.; Burger, J.; Mishra, A.K. A Machine Learning Approach to Radar Sea Clutter Suppression. In Proceedings of the 2017 IEEE Radar Conference (RadarConf), Seattle, WA, USA, 8–12 May 2017. [Google Scholar]
  30. Islam, S.M.M.; Rahman, A.; Yavari, E.; Baboli, M.; Boric-Lubecke, O.; Lubecke, V.M. Identity Authentication of OSA Patients Using Microwave Doppler Radar and Machine Learning Classifiers. In Proceedings of the IEEE Radio and Wireless Symposium, RWS, San Antonio, TX, USA, 26–29 January 2020. [Google Scholar]
  31. Abedi, H.; Luo, S.; Shaker, G. On the Use of Low-Cost Radars and Machine Learning for in-Vehicle Passenger Monitoring. In Proceedings of the 2020 IEEE 20th Topical Meeting on Silicon Monolithic Integrated Circuits in RF Systems, SiRF, San Antonio, TX, USA, 26–29 January 2020. [Google Scholar]
  32. Luo, F.; Poslad, S.; Bodanese, E. Human Activity Detection and Coarse Localization Outdoors Using Micro-Doppler Signatures. IEEE Sens. J. 2019, 19. [Google Scholar] [CrossRef]
  33. Kucuk, C.; Taskin, G.; Erten, E. Paddy-Rice Phenology Classification Based on Machine-Learning Methods Using Multitemporal Co-Polar X-Band SAR Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9. [Google Scholar] [CrossRef]
  34. Ghadimi, G.; Norouzi, Y.; Bayderkhani, R.; Nayebi, M.M.; Karbasi, S.M. Deep Learning-Based Approach for Low Probability of Intercept Radar Signal Detection and Classification. J. Commun. Technol. Electron. 2020, 65. [Google Scholar] [CrossRef]
  35. Wang, W.; Zhang, C.; Tian, J.; Wang, X.; Ou, J.; Zhang, J.; Li, J. High-Resolution Radar Target Recognition via Inception-Based VGG (IVGG) Networks. Comput. Intell. Neurosci. 2020, 2020. [Google Scholar] [CrossRef]
  36. Rahman, S.; Robertson, D.A. Classification of Drones and Birds Using Convolutional Neural Networks Applied to Radar Micro-Doppler Spectrogram Images. IET Radar Sonar Navig. 2020, 14. [Google Scholar] [CrossRef] [Green Version]
  37. Ahmed, S.; Cho, S.H. Hand Gesture Recognition Using an IR-UWB Radar with an Inception Module-Based Classifier. Sensors 2020, 20, 564. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Gao, J.; Shen, L.; Gao, L. Modulation Recognition for Radar Emitter Signals Based on Convolutional Neural Network and Fusion Features. Trans. Emerg. Telecommun. Technol. 2019, 30. [Google Scholar] [CrossRef]
  39. Liu, H.; Yang, R.; Yang, Y.; Hou, C.; Hu, Z.; Jiang, T. Human–Human Interaction Recognition Based on Ultra-Wideband Radar. Signal Image Video Process. 2020, 14. [Google Scholar] [CrossRef]
  40. Anishchenko, L.; Zhuravlev, A.; Chizh, M. Fall Detection Using Multiple Bioradars and Convolutional Neural Networks. Sensors 2019, 19, 5569. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Maragliulo, S.; Lopes, P.F.A.; Osorio, L.B.; de Almeida, A.T.; Tavakoli, M. Foot Gesture Recognition through Dual Channel Wearable EMG System. IEEE Sens. J. 2019, 19. [Google Scholar] [CrossRef]
  42. Frank, S.; Kuijper, A. Robust Driver Foot Tracking and Foot Gesture Recognition Using Capacitive Proximity Sensing. J. Ambient Intell. Smart Environ. 2019, 11, 221–235. [Google Scholar] [CrossRef]
  43. Frank, S.; Kuijper, A. Enabling Driver Feet Gestures Using Capacitive Proximity Sensing. In Proceedings of the 2018 International Conference on Intelligent Environments, IE, Rome, Italy, 25–28 June 2018. [Google Scholar]
  44. Huang, Q.; Chen, H.; Zhang, Q. Joint Design of Sensing and Communication Systems for Smart Homes. IEEE Netw. 2020, 34. [Google Scholar] [CrossRef]
  45. Dong, Y.; Yao, Y.D. Secure MMWave-Radar-Based Speaker Verification for IoT Smart Home. IEEE Internet Things J. 2021, 8. [Google Scholar] [CrossRef]
  46. Issakov, V. Circuit Design Challenges of Highly-Integrated Mm-Wave Radar-Based Sensors in SOI Based Technologies. In Proceedings of the ESSCIRC 2018—IEEE 44th European Solid State Circuits Conference (ESSCIRC), Dresden, Germany, 3–6 September 2018; Institute of Electrical and Electronics Engineers: Manhattan, NY, USA, 2018; p. 260. [Google Scholar] [CrossRef]
  47. Yang, H.; Zhang, Y.; DIng, W. Multiple Heterogeneous P-DCNNs Ensemble with Stacking Algorithm: A Novel Recognition Method of Space Target ISAR Images under the Condition of Small Sample Set. IEEE Access 2020, 8. [Google Scholar] [CrossRef]
  48. Oyama, Y.; Maruyama, N.; Dryden, N.; McCarthy, E.; Harrington, P.; Balewski, J.; Matsuoka, S.; Nugent, P.; van Essen, B. The Case for Strong Scaling in Deep Learning: Training Large 3D CNNs with Hybrid Parallelism. IEEE Trans. Parallel Distrib. Syst. 2021, 32. [Google Scholar] [CrossRef]
  49. Ghavamirad, R.; Sebt, M.A. Sidelobe Level Reduction in ACF of NLFM Waveform. IET Radar Sonar Navig. 2019, 13. [Google Scholar] [CrossRef] [Green Version]
  50. Kravchenko, V.F.; Pustovoit, V.I.; Churikov, D.V. A New Class of Weight and WA Systems of the Kravchenko-Kaiser Functions. Dokl. Phys. 2014, 59. [Google Scholar] [CrossRef]
  51. Chiang, H.P.; Lai, C.F.; Lai, Y.H.; Huang, Y.M. A Sensor-Based Feet Motion Recognition of Graphical User Interface Controls. Multimed. Tools Appl. 2016, 75. [Google Scholar] [CrossRef]
  52. Hu, L.; Cui, J. Digital Image Recognition Based on Fractional-Order-PCA-SVM Coupling Algorithm. Meas. J. Int. Meas. Confed. 2019, 145. [Google Scholar] [CrossRef]
  53. Uddin, M.P.; Mamun, M.A.; Hossain, M.A. PCA-Based Feature Reduction for Hyperspectral Remote Sensing Image Classification. IETE Tech. Rev. 2020. [Google Scholar] [CrossRef]
Figure 1. New recognition method using new radar signature image and deep learning via singular value decomposition for foot gesture.
Figure 1. New recognition method using new radar signature image and deep learning via singular value decomposition for foot gesture.
Sensors 21 03937 g001
Figure 2. The original and reconstructed STFT radar signatures from general RGB image and RGB image via colormap: (a) Original Image; (b) Reconstructed image from general RGB image; (c) Reconstructed image from RGB image via a colormap.
Figure 2. The original and reconstructed STFT radar signatures from general RGB image and RGB image via colormap: (a) Original Image; (b) Reconstructed image from general RGB image; (c) Reconstructed image from RGB image via a colormap.
Sensors 21 03937 g002
Figure 3. Experimental setup and foot motion scenes for Doppler-radar based foot gesture recognition: (a) Experimental setup and data acquisition system; (b) Utilized antenna; (c) Amplification board for beat signal; (d) Side view of foot gesture; (e) Front view of same gesture.
Figure 3. Experimental setup and foot motion scenes for Doppler-radar based foot gesture recognition: (a) Experimental setup and data acquisition system; (b) Utilized antenna; (c) Amplification board for beat signal; (d) Side view of foot gesture; (e) Front view of same gesture.
Sensors 21 03937 g003
Figure 4. Illustration of different foot gestures used in experiment: (a) Kicking; (b) Swinging; (c) Sliding; (d) Tapping; (e) Rolling ball.
Figure 4. Illustration of different foot gestures used in experiment: (a) Kicking; (b) Swinging; (c) Sliding; (d) Tapping; (e) Rolling ball.
Sensors 21 03937 g004
Figure 5. Radar signature for kicking: (a1) Time domain; (b1) Frequency domain; (c1) STFT spectrogram.Radar. signature for swinging: (a2) Time domain; (b2) Frequency domain; (c2) STFT spectrogram.Radar. signature for sliding: (a3) Time domain; (b3) Frequency domain; (c3) STFT spectrogram. Radar signature for tapping: (a4) Time domain; (b4) Frequency domain; (c4) STFT spectrogram. Radar signature for a rolling baseball: (a5) Time domain; (b5) Frequency domain; (c5) STFT spectrogram.
Figure 5. Radar signature for kicking: (a1) Time domain; (b1) Frequency domain; (c1) STFT spectrogram.Radar. signature for swinging: (a2) Time domain; (b2) Frequency domain; (c2) STFT spectrogram.Radar. signature for sliding: (a3) Time domain; (b3) Frequency domain; (c3) STFT spectrogram. Radar signature for tapping: (a4) Time domain; (b4) Frequency domain; (c4) STFT spectrogram. Radar signature for a rolling baseball: (a5) Time domain; (b5) Frequency domain; (c5) STFT spectrogram.
Sensors 21 03937 g005aSensors 21 03937 g005b
Figure 6. Original STFT radar signature images at a height of about 0.6 m: (a) Kicking; (b) Swinging; (c) Sliding; (d) Tapping; (e) Rolling baseball.
Figure 6. Original STFT radar signature images at a height of about 0.6 m: (a) Kicking; (b) Swinging; (c) Sliding; (d) Tapping; (e) Rolling baseball.
Sensors 21 03937 g006
Figure 7. STFT radar signature images at a different height of about 1.5 m: (a) Kicking; (b) Swinging; (c) Sliding; (d) Tapping; (e) Rolling baseball.
Figure 7. STFT radar signature images at a different height of about 1.5 m: (a) Kicking; (b) Swinging; (c) Sliding; (d) Tapping; (e) Rolling baseball.
Sensors 21 03937 g007
Figure 8. Eigenvalue and cumulative energy in radar signature RGB image for different foot gestures: (a) Eigenvalue; (b) Cumulative energy.
Figure 8. Eigenvalue and cumulative energy in radar signature RGB image for different foot gestures: (a) Eigenvalue; (b) Cumulative energy.
Sensors 21 03937 g008
Figure 9. Original and reconstructed radar signature RGB images according to different compression ratios for kicking gesture: (a) Original radar image; (b) Reconstructed radar image with compression ratio of ϵ = 90%; (c) Reconstructed radar image with compression ratio of ϵ = 95%; (d) Reconstructed radar image with compression ratio of ϵ = 99%.
Figure 9. Original and reconstructed radar signature RGB images according to different compression ratios for kicking gesture: (a) Original radar image; (b) Reconstructed radar image with compression ratio of ϵ = 90%; (c) Reconstructed radar image with compression ratio of ϵ = 95%; (d) Reconstructed radar image with compression ratio of ϵ = 99%.
Sensors 21 03937 g009
Figure 10. Reconstructed radar signature RGB images according to different foot gestures and movement of moving object, with compression ratio of ϵ = 99%: (a) Kicking; (b) Swinging; (c) Sliding; (d) Tapping; (e) Rolling ball.
Figure 10. Reconstructed radar signature RGB images according to different foot gestures and movement of moving object, with compression ratio of ϵ = 99%: (a) Kicking; (b) Swinging; (c) Sliding; (d) Tapping; (e) Rolling ball.
Sensors 21 03937 g010
Figure 11. Confusion matrix showing performance of radar sensor’s foot gesture recognition according to original and reconstructed radar images used for both training and testing: (a) Original radar image; (b) Reconstructed radar image with compression ratio of ϵ = 90%; (c) Reconstructed radar image with compression ratio of ϵ = 95%; (d) Reconstructed radar image with compression ratio of ϵ = 99%.
Figure 11. Confusion matrix showing performance of radar sensor’s foot gesture recognition according to original and reconstructed radar images used for both training and testing: (a) Original radar image; (b) Reconstructed radar image with compression ratio of ϵ = 90%; (c) Reconstructed radar image with compression ratio of ϵ = 95%; (d) Reconstructed radar image with compression ratio of ϵ = 99%.
Sensors 21 03937 g011aSensors 21 03937 g011b
Figure 12. Probabilistic performance of radar sensor’s foot gesture recognition according to original and reconstructed radar images: (a) Recall; (b) Precision.
Figure 12. Probabilistic performance of radar sensor’s foot gesture recognition according to original and reconstructed radar images: (a) Recall; (b) Precision.
Sensors 21 03937 g012
Figure 13. Confusion matrix showing foot gesture recognition performance when compressed and original radar signature images are used as data sets for training and testing, respectively: (a) Reconstructed radar of compression ratio of 90% for training and original radar images for testing; (b) Reconstructed radar of compression ratio of 95% for training and original radar images for testing; (c) Reconstructed radar of compression ratio of 99% for training and original radar images for testing.
Figure 13. Confusion matrix showing foot gesture recognition performance when compressed and original radar signature images are used as data sets for training and testing, respectively: (a) Reconstructed radar of compression ratio of 90% for training and original radar images for testing; (b) Reconstructed radar of compression ratio of 95% for training and original radar images for testing; (c) Reconstructed radar of compression ratio of 99% for training and original radar images for testing.
Sensors 21 03937 g013
Figure 14. Required memory capacity according to number of training data.
Figure 14. Required memory capacity according to number of training data.
Sensors 21 03937 g014
Table 1. The used radar parameters.
Table 1. The used radar parameters.
Radar ParameterValueUnit
Center frequency24GHz
Radar typeCW-
Sampling rate10KHz
Azimuth beam width80Degree
Elevation beam width12Degree
Maximum detection distance25 (RCS = 0 dBsm)m
Time window409.6ms
Frame rate4.9FPS
FFT point1024Point
Overlap size1023Point
Table 2. Comparative experiments’ results for five different learning models.
Table 2. Comparative experiments’ results for five different learning models.
ModelTraining Time
(Minutes)
RecallPrecisionF1Accuracy
GoogleNet2.540.910.920.910.96
ResNet1.520.890.900.890.96
VGG12.40.950.950.950.98
AlexNet1.240.920.930.920.97
PCA-SVM2.380.920.940.920.97
Table 3. Training results based on AlexNet deep learning model: training accuracy and loss.
Table 3. Training results based on AlexNet deep learning model: training accuracy and loss.
CaseTraining AccuracyLoss
Case # 196.00.14
Case # 297.00.23
Case # 396.70.14
Case # 485.70.53
Table 4. Probabilistic performance details of radar sensor’s foot gesture recognition according to original and reconstructed radar images.
Table 4. Probabilistic performance details of radar sensor’s foot gesture recognition according to original and reconstructed radar images.
CompressionGestureRecallPrecisionF1
OriginalKicking0.970.97980.9749
Swing0.960.94120.9505
Sliding0.940.92160.9307
Tapping0.9810.9899
Object0.970.97980.9749
90Kicking0.9910.9950
Swing0.920.96840.9436
Sliding0.980.89090.9333
Tapping0.9710.9848
Object0.970.97980.9749
95Kicking0.9810.9899
Swing0.770.97470.8603
Sliding0.90.79490.8571
Tapping0.980.95150.9655
Object0.970.94180.9557
99Kicking0.970.98980.9798
Swing0.840.85710.8485
Sliding0.850.77980.8134
Tapping0.950.93140.9406
Object0.910.97850.9430
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Song, S.; Kim, B.; Kim, S.; Lee, J. Foot Gesture Recognition Using High-Compression Radar Signature Image and Deep Learning. Sensors 2021, 21, 3937. https://doi.org/10.3390/s21113937

AMA Style

Song S, Kim B, Kim S, Lee J. Foot Gesture Recognition Using High-Compression Radar Signature Image and Deep Learning. Sensors. 2021; 21(11):3937. https://doi.org/10.3390/s21113937

Chicago/Turabian Style

Song, Seungeon, Bongseok Kim, Sangdong Kim, and Jonghun Lee. 2021. "Foot Gesture Recognition Using High-Compression Radar Signature Image and Deep Learning" Sensors 21, no. 11: 3937. https://doi.org/10.3390/s21113937

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop