Next Article in Journal
Some Hermite–Hadamard and Hermite–Hadamard–Fejér Type Fractional Inclusions Pertaining to Different Kinds of Generalized Preinvexities
Next Article in Special Issue
Prediction of COVID-19 Cases Using Constructed Features by Grammatical Evolution
Previous Article in Journal
Hybrid Feature Selection of Breast Cancer Gene Expression Microarray Data Based on Metaheuristic Methods: A Comprehensive Review
Previous Article in Special Issue
An Intelligent Vision-Based Tracking Method for Underground Human Using Infrared Videos
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Driver Abnormal Behavior Recognition and Analysis Strategy and Its Application in a Practical Vehicle

1
School of Electrical and Control Engineering, North China University of Technology, Beijing 100144, China
2
School of Automation, Qingdao University, Qinhuangdao 266071, China
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(10), 1956; https://doi.org/10.3390/sym14101956
Submission received: 30 August 2022 / Revised: 15 September 2022 / Accepted: 16 September 2022 / Published: 20 September 2022

Abstract

:
In this work, a novel driver abnormal behavior analysis system based on practical facial landmark detection (PFLD) and you only look once version 5 (YOLOv5) were developed to solve the recognition and analysis of driver abnormal behaviors. First, a library for analyzing the abnormal behavior of vehicle drivers was designed, in which the factors that cause an abnormal behavior of drivers were divided into three categories according to the behavioral characteristics including natural behavioral factors, unnatural behavioral factors, and passive behavioral factors. Then, different neural network models were established through the representation of the actual scene of the three behaviors. Specifically, the abnormal driver behavior caused by natural behavioral factors was identified by a PFLD neural network model based on facial key point detection, and the abnormal driver behavior caused by unnatural behavioral factors and passive behavioral factors were identified by a YOLOv5 neural network model based on target detection. In addition, in a test of the driver abnormal behavior analysis system in an actual vehicle, the precision rate was greater than 95%, which meets the requirements of practical application.

1. Introduction

In recent years, with the rapid development of the automobile industry and vehicle technology, road traffic safety has become a worldwide problem [1,2]. It is estimated that the number of road traffic fatalities reached 1.35 million in 2016 and is increasing every year [3]. To solve the problem of traffic safety and improve driving safety, many methods have been proposed to detect dangerous driving including the classification of driver context [4,5,6], vehicle context [7,8,9], and road context [10,11,12]. It has been reported that more than 90% of vehicle traffic accidents in the United States are caused by driver error [13], which is similar to other countries around the world. Compared with other objective factors, driver behavior is subjective and can be standardized. Therefore, it is necessary to study the behaviors that lead to driver operational errors.
There are many studies on behaviors that may lead to driver operational errors, with research content classifying abnormal behaviors including fatigued driving, distracted driving, and dangerous driving.
For the study of fatigued driving, some studies believe that a driver’s fatigue behavior is a kind of human physiological state, so by finding the physiological indicators related to the abnormal behavior of the driver, fatigued driving can be determined. For example, in [14], by detecting the EEG signal of a driver’s physiological indicators, it is possible to predict whether the driver has fatigued driving behavior. However, the physiological detection method for a driver, similar to an EEG signal, is difficult to implement in a real vehicle because a sensor device needs to be placed such as a pole piece into the driver, which easily causes a psychological burden to the driver. Therefore, Guanglong Du et al. proposed a fatigue detection method characterized by the facial information of the levels of eye and mouth opening [15]; Bappaditya Mandal et al. proposed predicting driver fatigue behavior by estimating the pose of the eyes [16] including eye opening estimation and eyelid closure percentage estimation. Moreover, methods of detecting fatigued driving based on the key information features of the driver’s face will not affect the driver and they are widely used in the detection of abnormal driver behavior.
For research on distracted driving, Alberto Fernándezs et al. believe that the driver’s use of mobile phones is a distracted driving behavior; therefore, a traditional learning strategy was proposed [17]. First, the key features of the mobile phone were manually extracted from image samples to build a detection model. Second, the mobile phone target was detected using a SVM detector. Finally, whether the driver exhibits distracted driving behavior is predicted [18]. However, the manual feature extraction method of this traditional learning strategy is challenging for the selection of excellent features, and the detection process is time-consuming. Landu Jiang et al. believe that a driver’s smoking and phone calling behavior is a kind of distracted driving behavior, and a deep convolutional neural network (CNN) method was designed. A CNN network can automatically extract key features in image samples and classify the detected smoking and phone calls [19]. This deep learning-based method does not require manual feature extraction, and the detection speed of the target is fast. Therefore, in the detection of abnormal behaviors, deep learning-based methods outperform traditional learning strategies [20].
In the research of dangerous driving, some studies have determined whether the driver has dangerous driving behaviors using a vehicle data (e.g., speed, acceleration) detection method. For example, [21] proposed identifying dangerous driving by constructing a set of vehicle trajectory histograms (e.g., control point histogram and velocity histogram). Chen et al. applied the SAX method to detect dangerous driving behavior by analyzing acceleration [22]. However, detection methods that analyze vehicle data are not real-time and cannot intervene in a timely manner in dangerous driving situations. Therefore, Galarza et al. proposed the use of smart devices and a CNN to detect driver drowsiness, dangerous driving behavior, and real-time monitoring and timely reminders for dangerous drowsy driver behavior [23]. This method of combining smart devices with a CNN solves the problem that vehicle data analysis cannot be carried out in real-time and provides a new method for the detection of dangerous driving behavior.
The above method for detecting the abnormal behavior of a driver using the key features of the driver’s face to detect fatigued driving behavior is operable in an actual vehicle environment. In the detection of distracted driving behavior and dangerous driving behavior, the detection method using a deep CNN has the advantages of fast detection speed, and when combined with the use of smart devices, the output detection results can be obtained in real-time. Therefore, a multiangle analysis of driver abnormal behavior is very important to solve actual traffic safety problems.
Regarding the abnormal behavior of a driver during the driving process, the reasons for the specific behavior are also different according to their different performance characteristics. An abnormal behavior analysis problem caused by various factors can be realized by constructing different deep learning models. To the best of our knowledge, there is no mature work on the multiangle analysis of abnormal driver behavior at home or abroad.
The contributions of this paper are as follows:
  • We established an abnormal behavior analysis library, in which natural behavioral factors include: (1) eyes closed, (2) yawning, and (3) nodding; unnatural behavioral factors include: (1) smoking, (2) phone, and (3) drinking water; and passive behavioral factors include disturbance. The abnormal behavior library provides the basis for the construction of a driver abnormal behavior recognition model.
  • A driver abnormal behavior analysis strategy is proposed. According to the specific and key abnormal behavior characteristics of a driver, the abnormal behavior of the driver is analyzed from three aspects: natural behavioral factors, unnatural behavioral factors, and passive behavioral factors.
  • A driver abnormal behavior recognition system was developed. The system was tested and applied in real vehicles, and a series of experimental results showed that this research has high accuracy in the task of the real-time detection of abnormal driver behavior.
The rest of this paper is organized as follows. In Section 2, the analysis and model design of abnormal behavior problems are studied. In Section 3, the design of the driver abnormal behavior recognition model is introduced. In Section 4, the experimental details and results of the driver abnormal behavior analysis system are described. Section 5 summarizes the main conclusions of this paper and discusses future work.

2. Problem Analysis and Model Design

2.1. Design of the Driver Abnormal Behavior Library

During driving, there are many abnormal behaviors that lead to driving errors, not only due to a specific factor, but most abnormal behavior research points to one cause branch (such as fatigue, distraction, etc.), which is disconnected from the reality base, resulting in an incomplete analysis of anomalous behavior. This paper proposes a standard that can comprehensively analyze the abnormal behavior of drivers, taking the driver as the starting point, considering the physiological behavior of the driver in the natural state, the possible autonomous behavior in the unnatural state, and external disturbances. Stressful behaviors are divided into natural behavioral factors, unnatural behavioral factors, and passive behavioral factors.
Drivers have some key abnormal behaviors during driving, and these key abnormal behaviors are classified. Key behaviors such as closing eyes for a long time, frequent yawning, and frequent nodding are classified as natural behavioral factors; key behaviors such as smoking, making a telephone call, and drinking water are classified as unnatural behavioral factors; and key behaviors that interfere with driving due to other reasons are classified as passive behavioral factors. The driver abnormal behavior analysis library is shown in Figure 1.

2.2. Structure Design

Abnormal driver behavior under natural behavior factors, unnatural behavior factors, and passive behavior factors have specific manifestations—facial features and the appearance of objective things. This paper used two models to address the driver abnormal behavior recognition task. Figure 2 shows the logical structure diagram of the driver abnormal behavior recognition system. The real-time driving status of the driver was obtained through an image capture sensor installed in the cab, and the real-time monitoring image was transmitted to the system through the reasoning of model A and model B to identify the abnormal behavior of the driver. The identified abnormal behavior information is displayed on the system visualization interface.
The abnormal behavior under the driver’s natural behavior factor is reflected in the driver’s facial features, and the establishment of a deep neural network model for the facial features of the human face is the key to solving such problems. Yi Sun et al. proposed a neural network for the detection of five facial key points (left eye, right eye, nose, left mouth corner, right mouth corner) [24]. Erjin Zhou et al. proposed the detection of 68 facial key points (such as facial lines, nose lines, lip lines, and the outer line of the eye) [25]. The detection of 68 facial key points is currently a widely used facial key point detection method [26,27]. Therefore, the deep network model A detects facial key points, and then according to the position changes of the key points, determines whether the abnormal driver behavior under natural behavior factors occurs.
Driver abnormal behavior under unnatural behavioral factors and passive behavioral factors is usually accompanied by the appearance of other objects or people such as mobile phones and water bottles. Detecting abnormal behavior using the detection of other objects or people associated with it has become a general way to solve such problems [28,29]. Mainstream target detection deep neural networks include R-CNN, YOLO, and SSD [30,31]. The abnormal behavior recognition system is a practical application system. Considering the weight of the model, detection accuracy, detection speed, etc., the YOLOv5 neural network model has obvious advantages [32]. Therefore, the design of deep network model B detects objects and other people around the driver, and then according to the positional relationship with the driver, determines whether driver abnormal behavior under unnatural or passive behavior factors occurs.

3. Design of the Driver Abnormal Behavior Recognition Model

Figure 3 shows the structural design of the driver abnormal behavior detection model. Video obtained from a vehicle camera is used as the input, the abnormal behavior of the driver is inferred and predicted through Module A and Module B, respectively, and the judgment result is output to the system monitoring the visualization interface in real-time.
The design of Module A adopted the structure of a PFLD deep neural network [33], which was divided into the input, backbone, and prediction layers. The video stream data were obtained through an on-board camera installed in the cab, and the video frame picture was obtained by sampling the video frame-by-frame.
The red-green-blue (RGB) image of the input layer is the frame picture of the collected video. The image enters the backbone layer to perform 2, 4, and 8 times downsampling operations to obtain the key point feature information of the input image. The prediction layer performs multiscale operations, which can predict key points of three different scales. The obtained feature map information is processed using a full connection, which predicts the coordinate information of 68 key points on the driver’s face and uses the coordinate information to determine whether the driver has abnormal behaviors such as long-term eye closure, frequent yawning, or frequent nodding caused by natural behavioral factors.
The design of Module B adopted the structure of the YOLOv5 deep neural network [34], which was divided into the input, backbone, neck, and prediction layers. It also uses the frame picture of the video collected from the vehicle camera as the input image. Anchors were introduced at this point to optimize the prediction of the width and height of the ground-truth (GT) bounding box to improve the accuracy.
The backbone layer performed 8, 16 and 32 times downsampling to aggregate images at different granularities and extract the image features. The neck layer fuses the image features twice through operations such as 2 times upsampling and concatenation and then passes the transformed image features to the prediction layer, which applies anchors to the feature map to generate bounding boxes with associated class probabilities. According to the predicted category information and bounding box coordinate information, it determines whether the driver has abnormal behaviors such as smoking, making a telephone call, or drinking water or disturbed driving caused by unnatural and passive behavioral factors.

3.1. PFLD Network Model Design

3.1.1. Design Principles

The module was designed to use facial key point detection tasks to identify abnormal behaviors caused by natural driver behavioral factors. Sixty-eight facial key point detection is the most common solution today including eyebrows, eyes, nose, mouth, and contours. The location information of facial key point annotation is shown in Figure 4.
There were five key points for unilateral eyebrows, which were uniformly sampled from the right to the left, for a total of 5 × 2 = 10 key points. The eyes were divided into six key points, and the upper and lower eyelids were uniformly sampled from the right to the left boundary, for a total of 6 × 2 = 12 key points. The lips were divided into 20 key points, except for the two corners of the mouth, which were divided into the upper and lower lips, the outer boundary of the upper and lower lips, each of which was uniformly sampled by five key points, and the inner boundary of the upper and lower lips, each of which was uniformly sampled by three key points, so a total of 20 key points. The nose included four key points of the bridge of the nose and five points evenly collected from the tip of the nose, for a total of nine key points. The face contours were uniformly sampled with 17 key points.
For the detection problem of the 68 key points on a driver’s face, we built a nonlinear model that satisfies Equation (1):
s = f ( x )
where s = [ s 0 ,   s 1 ,   s 2 , ,   s 67 ] T , with s i being the detected i + 1 th key point. Here, f is an unknown nonlinear function, and x is the input image extracted from the video frame. Two parameters were selected for the output s : a and b, which are the abscissa and ordinate of s , respectively.
The abnormal behavior was identified on the output key point position information, and the behavior detection from three consecutive frames was determined as the completion of one action. According to Equation (2), it can be determined that if the ratio of the height and width of the eyes in three consecutive frames is less than the value, one eye closure has occurred. In the experiment, 60 continuous eye-closing actions were set as the long-term eye-closing behavior. According to Equation (4), it can be determined that if the ratio of the height and width of the mouth in the three-frame detection is less than the value, a yawn has occurred. In the experiment, if the yawning action occurred 15 times, the recognition result was frequent yawning behavior. Head pose estimation was used to obtain the attitude angle of the head through a facial image and calculate the three Euler angles of the pitch, yaw, and roll [35]. Because the nodding behavior is only the action of nodding up and down, only the size of the pitch is considered, and if the pitch for three consecutive frames is less than the set threshold of 0.3, nodding has occurred. In the experiment, if nodding occurred 15 times, the recognition result was frequent nodding behavior.
r = r 1 + r 2 2 < α
where r 1 and r 2 are the ratio of the height and width of the right eye and the left eye, respectively; r is the average value of the height and width ratio of the two eyes; and α is the threshold parameter of the height and width ratio of the eye, which was set to 0.2 after actual adjustment. Taking the eye as an example, the calculation equation is:
r 1 = A + B 2 * C
where A = x 37 x 41 2 + y 37 y 41 2 , B = x 38 x 40 2 + y 38 y 40 2 , and C = x 36 x 39 2 + y 36 y 39 2 , as shown in Figure 5a, the coordinates ( x 36 , y 36 ) , ( x 37 , y 37 ) , ( x 38 , y 38 ) , ( x 39 , y 39 ) , ( x 40 , y 40 ) , and ( x 41 , y 41 ) are the coordinates of the six key points of the right eye.
q = X + Y 2 * Z < β
where X = x 50 x 58 2 + y 50 y 58 2 , Y = x 52 x 56 2 + y 52 y 56 2 , and Z = x 48 x 54 2 + y 48 y 54 2 ; q is the average value of the height and width ratio of the mouth; and β is the threshold parameter of the height and width ratio of the eyes, which was 0.5 after actual adjustment. As shown in Figure 5b, the coordinates ( x 48 , y 48 ) , ( x 50 , y 50 ) , ( x 52 , y 52 ) , ( x 54 , y 54 ) , ( x 56 , y 56 ) , and ( x 58 , y 58 ) were the coordinates of the six key points on the outer boundary of the lips.

3.1.2. Network Structure

Figure 6 shows the PFLD network structure diagram. The input is an RGB image, and the output is the predicted facial key points. The network structure is divided into two parts: the main network structure and the auxiliary network structure. The red box is the 68 key points used by the main network structure to predict the face; the black box is the auxiliary network structure. Previous studies have shown that adding appropriate auxiliary constraints can make the detection of facial landmarks more stable and robust [36,37]. The spatial geometry of the head pose pitch, yaw, and roll is used in the process of training key points. The information is used as a constraint to improve the robustness of key point detection.
The bottleneck structure in the main network uses MobileNet [38] blocks instead of traditional convolution operations. The basic unit of MobileNet is depthwise separable convolution, which is a factorized convolution that can be decomposed into two smaller operations: depthwise convolution and pointwise convolution. Depthwise convolution is different from standard convolution. For standard convolution, the convolution kernel is used on all input channels (input channels), while depthwise convolution uses different convolution kernels for each input channel; that is, each convolution kernel corresponds to an input channel. Pointwise convolution is actually an ordinary convolution, but it uses a 1 × 1 convolution kernel. Depthwise separable convolution first uses depthwise convolution to convolve different input channels separately and then uses pointwise convolution to combine the above outputs. The overall effect is similar to a standard convolution, but it will greatly reduce the amount of calculation and number of model parameters, making the model smaller and faster.
The auxiliary network structure takes the fourth-layer feature map of the main network as input, obtains a 1 × 1 × 128 feature map through four convolution operations, and obtains the spatial geometric information of the head pose pitch, yaw, and roll (i.e., the Euler angles) through the full connection operation [39]. The purpose of the network is to estimate the three-dimensional Euler angles for each input face sample, and the ground truth is estimated from the face key point information in the training data so that large angles and large deviations can be used for large samples by passing a larger loss and supervised and assisted training convergence.
To solve the data imbalance problem, the ability of accurate detection is improved by imposing a large penalty on the small number of training samples [40]. In mathematical terms, the loss function is expressed as follows:
L : = 1 M m = 1 M n = 1 N γ n d n m
where M is the number of samples; m is the number of feature points; γ n is different weights; and d n m is the distance measure (L1 or L2 distance) of the feature points. Considering the balance of geometric constraints and data, a new loss function was designed as follows:
L : = 1 M m = 1 M n = 1 N ( c = 1 C ω n c k = 1 K ( 1 cos θ n k ) ) d n m 2 2
where c = 1 C ω n c k = 1 K ( 1 cos θ n k ) is the final sample weight. K = 3 represents the three dimensions of the face pose estimation, that is, the yaw, pitch, and roll angles. From the calculation formula, the higher the angle, the greater the weight. C is the number of different face categories, and the face is divided into six categories: side face, front face, head up, head down, expression, and occlusion. The weight ω n c was set according to a small number of samples belonging to category C, and if there were few samples in a certain category, the given weight was greater.

3.2. YOLOv5 Network Model Design

3.2.1. Design Principles

The module was designed to use object detection tasks to identify abnormal behaviors caused by unnatural and passive driver behavioral factors. Considering the limitations of label definition for target detection [41], the possible target classification labels were divided into four categories: (1) face, (2) cigarette, (3) mobile phone, and (4) water bottle. We built a nonlinear model that satisfies Equation (7):
y = g ( x )
where y = [ y 1 ,   y 2 ,   y 3 ,   y 4 ] T , with y 1 denoting a detected bounding box of face; y 2 denoting a detected bounding box of cigarette; y 3 denoting a detected bounding box of mobile phone; and y 4 denoting a detected bounding box of water bottle. Here, g is an unknown nonlinear function, and x is the image input extracted from the video frame.
The output is expressed as y i = [ a , b , w , h , c , p 0 , p 1 , p 2 , p 3 ] T , i = 1 , 2 , 3 , 4 , and nine parameters are contained. Among them, a and b are the abscissa and ordinate of y , respectively; w and h are the width and height of y , respectively; c is the confidence of belonging to the current class; and [ p 0 , p 1 , p 2 , p 3 ] T denotes the classification results of the four label categories, where [ 1 , 0 , 0 , 0 ] T represents a face label, [ 0 , 1 , 0 , 0 ] T represents a cigarette label, [ 0 , 0 , 1 , 0 ] T represents a mobile phone label, and [ 0 , 0 , 0 , 1 ] T represents a water bottle label.
According to Equation (7), the output label value is identified as abnormal behavior. If label 1 and label 2 exist at the same time, the recognition result is smoking behavior. The recognition result of label 1 and label 3 is making a telephone call behavior. If label 1 and label 4 exist at the same time, the recognition result is drinking water behavior. If label 1 appears twice, and the label position distance is less than the safety threshold, there are other people in the cab, which may interfere with driving. At this time, the recognition result is disturbed driving behavior.
i f y 1 , y 2   and   [ 1 , 0 , 0 , 0 ] T [ 0 , 1 , 0 , 0 ] T = [ 1 , 1 , 0 , 0 ] T Smoking   y 1 , y 3   and   [ 1 , 0 , 0 , 0 ] T [ 0 , 0 , 1 , 0 ] T = [ 1 , 0 , 1 , 0 ] T Making   a   telephone   call   y 1 , y 4   and   [ 1 , 0 , 0 , 0 ] T [ 0 , 0 , 0 , 1 ] T = [ 1 , 0 , 0 , 1 ] T Drinking   water   y 1 = 2   and   ( a 1 1 a 2 1 ) 2 + ( b 1 1 b 2 1 ) 2 < θ Disturbed   driving
where a 1 1 , b 1 1 are the horizontal and vertical coordinate values of the detected first face label and a 2 1 , b 2 1 are the horizontal and vertical coordinate values of the detected second face label, and the Euclidean distance is used to calculate the distance between the two labels. θ is the distance threshold parameter, which is actually adjusted to 226.

3.2.2. Network Structure

Figure 7 shows the YOLOv5 network structure diagram. The input is an RGB image, and the output is three tensors representing three different scale bounding boxes. The parameter information contained in each tensor produces the classification result of object detection. The network model defaults to the 640 × 640 image input, and multiple combined CBS+CSP operations are designed in the network to extract deeper feature maps, in which the CBS consists of a convolution operation (Conv), batch normalization (BN) layer, and SiLU activation function [42]. Deeper feature maps have stronger semantic features and weaker location information, while shallower feature maps have stronger location information and weaker semantic features. Correspondingly, these two types of features can undergo information transfer and fusion to obtain feature maps that capture more complex feature information.
The cross-stage partial network (CSPNet) [43] divides the input feature map into two branches: one branch is convolutional, and the other branch is concatenated with the convolution result of the previous step, which reduces the information loss in the inference process. This structure can be effectively applied in the backbone to enhance the learning ability of neural networks. The SPPF structure passes the input serially through multiple 5 × 5 MaxPool layers. Compared with the original SPP structure [44], two 5 × 5 convolution operations were used to replace a 9 × 9 convolution operation, three 5 × 5 convolution operations replaced one 13 × 13 convolution operation, and the receptive field remained unchanged. However, the calculation time was reduced, which helped to solve the matching problem between the anchor points and feature maps.
The generalized intersection over union (GIOU) loss indicator in the prediction layer was used to evaluate the position loss between the target and predicted bounding boxes [45], and the binary cross-entropy (BCE) logit loss function [46] was used to evaluate the object classification loss. The GIOU loss function is:
L o s s _ G I O U = 1 G I O U = 1 ( I O U b o x _ r ( b o x _ p b o x _ g t ) b o x _ r )
Among them, the IOU loss evaluation index [47] evaluates the accuracy of the position regression according to the intersection ratio (intersection/union) of the target bounding box and the predicted bounding box. box_p is the prediction bounding box, box_gt is the ground-truth box, and box_r is the minimum bounding rectangle of box_p and box_gt.
Due to the imbalance of the number of positive and negative samples, the BCE logit loss was used to increase the number of positive sample anchors to balance the number of positive and negative samples, which can accelerate convergence in practice. The BCE logit loss function is:
B E C L o s s = 1 n i = 1 n ( y i log y i + ( 1 y i ) log ( 1 y i ) )
where y i = S i g m o i d ( x i ) = 1 1 + e x i ; n represents the total number of categories; x i is the predicted value of the current category; y i is the probability of the current category; and y i * is the true value of the current category.

4. Experiment

This section presents the experimental platform design and results of this study. In Section 4.1, the design of the experimental platform is introduced. In Section 4.2, the abnormal behavior recognition PFLD model of natural behavioral factors, abnormal behavior recognition YOLOv5 model of unnatural and passive behavioral factors experimental results, the deployment of the PFLD and YOLOv5 models using Nvidia Jetson Xavier modules, and the test of the experiment platform in a real vehicle scene were analyzed.

4.1. Experimental Platform Design

Figure 8 shows a flow chart of the experimental design. The experimental design was divided into three parts: test vehicle, test platform, and visual interface.
The process of driver abnormal behavior recognition is represented by the green line in Figure 8, in which the video capture card of the experiment platform is connected with the on-vehicle cameras on the experimental vehicle through wiring, the real-time monitoring data in the cab can be collected, and the data can be collected through the serial input to the Nvidia Jetson Xavier (NX) module. The NX module transmits the image of the video stream data into the deployed abnormal behavior recognition model. If there is an abnormal driver behavior, the visual interface of the system displays a picture of the moment when the behavior occurs and outputs the alarm information.
At the same time, considering the complex bus environment of the experimental vehicle, on-vehicle cameras need to meet certain practical requirements such as high performance, low power consumption, high video frame rate, and adaptation to outdoor environments. The experiment platform is the key to the driver abnormal behavior analysis system. Its basic structure includes a carrier board, video capture card, and NX module. The details of the three basic structures of the test platform are as follows:
  • Carrier board
The system uses an RTSO-6002E carrier board, which is an industrial-level carrier board with low power consumption and high safety. Its operating temperature range is −40–+80 °C, making it suitable for complex outdoor environments.
2.
Video capture card
The system uses an RTSV-6911i mini-PCIe video capture card to meet the needs of the visual processing system. This video capture card supports the full frame rate, multichannel video capture, and hardware color space conversion, and can convert the captured video data into the format required for system display, image analysis, image processing, and other applications. Therefore, it provides powerful front-end processing support.
3.
Nvidia Jetson Xavier module
The Nvidia Jetson Xavier [48] is a processor from Nvidia for deep machine learning. The NX has a deep learning inference computing framework that is natively supported by the entire Jetson platform, optimizing the use of GPU resources. It has strong computing power, fast processing speed, and a small module area. Therefore, it can achieve high computing performance in a compact modular system.

4.2. Experimental Results

In this study, there were three factors that caused the abnormal behavior of drivers, namely, natural behavior factors, unnatural behavior factors, and passive behavior factors. We constructed two abnormal behavior detection models according to the characteristics of the three factors. Abnormal behaviors under the influence of natural behavioral factors were identified using the PFLD model, and abnormal behaviors influenced by unnatural behavioral factors and passive behavioral factors were identified using the YOLOv5 model. Due to the different models, the indicators for training and evaluating each model were also different. In terms of evaluating the indicators of each model, considering the practical application problems, the experiment only took the detection accuracy of the model and the detection speed of the model as the reference.

4.2.1. Experimental Results for Natural Behavior Factors

The PFLD model uses the 300 W Common dataset, of which 3148 images were used as a training set and 689 images were used as a test set. In the data augmentation part, methods of flipping, rotating the image at 5° intervals (−30°~30°), and randomly occluding 20% of the face area were adopted. The parameter settings of the training process were shown in Table 1. The entire network was trained on Nvidia GeForce RTX 3090 GPUs.
Interocular normalization (ION) is a commonly used index to evaluate the accuracy of the face key points, and its mathematical equation is:
e i = x i x i * 2 d
where x i represents the predicted coordinates of the i th point, and x i * represents the ground truth of the i th point. d is the distance between the left and right corners of the eyes. Clearly, this value is used to indicate the accuracy of the prediction, and the smaller the value, the better. Table 2 compares the PFLD model with other excellent models on the 300 W Common dataset.
The PFLD model had a high accuracy, the ION could reach 3.11 on the 300 W Common dataset, and the running speed was very fast, up to 3.5 ms on the GPU GTX 1080 Ti processor.

4.2.2. Experimental Results of Unnatural and Passive Behavioral Factors

The dataset for the YOLOv5 model was made locally by us. The goal of the dataset was to simulate specific abnormal behaviors, collect multiple videos, cut and frame these videos, convert the videos into sample images, and then classify and annotate the sample images. The resulting labeled samples were used as a dataset of abnormal bus driver behavior. Figure 9 shows examples of the labeled images for various types of abnormal driver behaviors represented in the dataset. Among them, 7965 images were used as the training set, and 2433 images were used as the test set.
In the data augmentation part, the mosaic data augmentation method was adopted. The main idea is to randomly crop and scale four pictures and then randomly arrange and splice them to form a picture. The parameter settings of the training process are shown in Table 3. The entire network was trained on Nvidia GeForce RTX 3090 GPUs.
In the evaluation of the YOLOv5 model, the precision refers to the proportion of positive examples that are correctly identified and proportional to the number of positive examples, indicating the proportion of the target detected by the model as the real target object, as shown in the confusion matrix in Table 4 and Equation (12). The recall is the number of positive examples that are correctly identified as sorted out of all the positive samples tested, which represents the proportion of all real targets detected by the model, as shown in Equation (13). The average precision (AP), expressed as the integration of the precision to recall on the precision–recall curve, as shown in Equation (14), and the mean average precision (mAP) refers to the AP value. The mean under all categories was evaluation for the entire dataset.
P r e s i t i o n = T P T P + F P
R e c a l l = T P T P + F N
A P = 0 1 P R d R
The weight of the YOLOv5 model obtained in the experiment was small, the mAP could reach 0.87 on the test dataset, and the running speed was very fast. The detection rate on the Tesla P100 processor was 140 f/s.

4.2.3. Deployment of the PFLD and YOLOv5 Models on the NX

The optimal model was obtained by modifying and adjusting the PFLD and YOLOv5 networks. The optimal model needs to be deployed in the NX module of the experimental platform before it can be applied in actual vehicles. To obtain the real-time frame rate of the in-vehicle video, the NX used TensorRT to deploy the model.
TensorRT is a deep learning inference computing framework that optimizes the use of GPU resources [52]. TensorRT can be divided into two phases: the build phase and the compile phase. In the build phase, TensorRT receives externally provided network structure weights and hyperparameters, optimizes the network compiled on the NX, and generates an inference engine. In the compile phase, the GPU computing resources are called by running the inference engine, and the process is shown in Figure 10.
In fact, TensorRT only provides inference and does not require real-time feedback. Network deployment does not affect the computing power and memory resources. Resources are usually not particularly tight, but they can speed up the acquisition of video frame rates on the NX side. Using the TensorRT inference engine in the experiment, the NX side can obtain a video frame rate of approximately 40 fps, which meets the real-time requirements.

4.3. Real Vehicle Test

The performance of the model deployed to the NX side was evaluated, and the performance of the model on the application side was evaluated by calculating the precision rate (PR), missed alarm rate (MR) and false alarm rate (FR), and the PR expectation on the application side could reach 95%. The PR, TR, and FR are shown in the confusion matrix in Table 4 and Equations (15)–(17):
P R = T P + T N T P + N P + T F + N F
M R = F P T P + F P
F R = F N T N + F N
This experiment verified the performance of the model on the NX side by randomly selecting the data in the local scene many times. Figure 11 shows the PR, MR, and FR of the NX side model. Among them, (a) randomly selecting the detection results of 1417 local image data, PR = 95.6%, MR = 6.8%, FR = 0.2%; (b) randomly selecting the detection results of 1016 local image data, PR = 95.9%, MR = 5.2%, FR = 2.7%. After multiple verifications, the PR was above 95%.
We developed an interface display for the identification of abnormal driver behavior in the form of an external window. Figure 12 shows the visual interface of the system operation. The visual interface can output multichannel camera images, display the captured video frames of abnormal driver behavior, and output alarm information in real-time.
This paper proposed a novel driver abnormal behavior recognition and analysis strategy, developed a driver abnormal behavior analysis system based on deep learning, and applied it in practical vehicles. The driver abnormal behavior analysis system identified driver abnormal behavior through the real-time acquisition of in-vehicle video and outputs the recognized results to the visual interface of the system operation. Moreover, the system was applied in the real bus, and the state was stable.

5. Conclusions and Discussion

In this work, a novel driver abnormal behavior analysis system was developed to solve the recognition and analysis of driver abnormal behaviors. There are many kinds of abnormal behaviors of drivers, so a library for analyzing the abnormal behaviors of drivers was established. The driver abnormal behavior library, according to the behavior performance of each abnormal action, divides the factors that cause driver abnormal behavior into three categories: natural behavior factors, unnatural behavior factors, and passive behavior factors. Among them, the abnormal behaviors of closing eyes, yawning, and dozing off are natural behavioral factors; smoking, making phone calls, and drinking water are unnatural behavioral factors; and disturbed driving is a passive behavioral factor. Then, two neural network models were established through the representation of the actual scene of the three behaviors, and the driver abnormal behavior caused by natural behavioral factors was identified by a PFLD neural network model based on the detection of facial features. Abnormal driver behavior caused by unnatural behavioral factors and passive behaviors were identified by a target-based YOLOv5 neural network model. Furthermore, an embedded system for analyzing the abnormal behavior of drivers was developed. The purpose of the driver’s abnormal behavior recognition system is to apply to the actual vehicle, therefore, the accuracy and speed of the mobile terminal need to be considered in the research. After a test of the practical application, the precision rate of the abnormal behavior recognition proposed in this paper was more than 95%, in addition, the detection speed could also meet the requirements of real-time.
There are many kinds of abnormal driver behaviors, and the system did not consider other abnormal behaviors of drivers except the seven typical abnormal behaviors of drivers in this study. At the same time, there are many options for the identification of different abnormal behaviors. However, selecting an appropriate detection method for a specific abnormal behavior poses a great challenge to the driver’s abnormal behavior recognition system.
In the future, we will further study the abnormal behavior of drivers based on the actual situation, in addition to considering the abnormal behavior of the driver’s hands off the steering wheel.

Author Contributions

Conceptualization, H.J. and S.L.; Methodology, X.W. and S.L.; Software, L.W.; Validation, X.W., S.L. and H.J.; Investigation, X.W. and S.L. and Z.H.; Writing—original draft preparation, X.W.; Writing—review and editing, X.W. and S.L.; Supervision, L.W.; Project administration, H.J.; Funding acquisition, S.L. and L.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Beijing Municipal Natural Science Foundation under Grant 4212035, the National Natural Science Foundation of China under Grants 61903004, 61833001, R&D Program of Beijing Municipal Education Commission (KM202210009011), North China University of Technology YuYou Talent Training Program, and in part by the North China University of Technology Scientific Research Foundation, and Beijing Municipal Great Wall Scholar Program (CIT&TCD20190304).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Conflicts of Interest

The author declares that there are no conflict of interest.

References

  1. Jain, M.; Kulkarni, P. Application of AI, IOT and ML for Business Transformation of The Automotive Sector. In Proceedings of the 2022 International Conference on Decision Aid Sciences and Applications (DASA), Chiang Rai, Thailand, 23–25 March 2022. [Google Scholar]
  2. Das, D.; Pargal, S.; Chakraborty, S.; Mitra, B. Why Slammed the Brakes On? Auto-annotating Driving Behaviors from Adaptive Causal Modeling. In Proceedings of the 2022 IEEE International Conference on Pervasive Computing and Communications Workshops and other Affiliated Events (PerCom Workshops), Pisa, Italy, 21–25 March 2022. [Google Scholar]
  3. Global Status Report on Road Safety. 2018. Available online: https://www.who.int/violence_injury_prevention/road_safety_status/2018/%en/ (accessed on 20 October 2021).
  4. Bergasa, M.L.; Nuevo, J.; Sotelo, M.A.; Barea, R.; Lopez, M.E. Real-time system for monitoring driver vigilance. IEEE Trans. Intell. Transp. Syst. 2006, 7, 63–77. [Google Scholar] [CrossRef]
  5. Eriksson, M.; Papanikotopoulos, N.P. Eye-tracking for detection of driver fatigue. In Proceedings of the Conference on Intelligent Transportation Systems, Boston, MA, USA, 12 November 1997. [Google Scholar]
  6. Singh, S.; Papanikolopoulos, N.P. Monitoring driver fatigue using facial analysis techniques. In Proceedings of the 199 IEEE/IEEJ/JSAI International Conference on Intelligent Transportation Systems (Cat. No.99TH8383), Tokyo, Japan, 5–8 October 1999. [Google Scholar]
  7. Jacob, B.; Violette, E. Vehicle Trajectory Analysis: An Advanced Tool for Road Safety. Procedia Soc. Behav. Sci. 2012, 48, 1805–1814. [Google Scholar] [CrossRef]
  8. Dong, Y.; Uchimura, K.; Murayama, N. Driver Inattention Monitoring System for Intelligent Vehicles: A Review. IEEE Trans. Intell. Transp. Syst. 2011, 12, 596–614. [Google Scholar] [CrossRef]
  9. Subirats, P.; Goyat, Y.; Jacob, B.; Violette, E. A New Road Safety Indicator Based on Vehicle Trajectory Analysis. Transp. Res. Procedia 2016, 14, 4267–4276. [Google Scholar] [CrossRef]
  10. Zhou, H.; Itoh, M.; Inagaki, T. Influence of cognitively distracting activity on driver’s eye movement during preparation of changing lanes. In Proceedings of the 2008 SICE Annual Conference, Tokyo, Japan, 20–22 August 2008. [Google Scholar]
  11. Ericsson, E. Variability in urban driving patterns. Transp. Res. Part D Transp. Environ. 2000, 5, 337–354. [Google Scholar] [CrossRef]
  12. Mudgal, A.; Hallmark, S.; Carriquiry, A.; Gkritza, K. Driving behavior at a roundabout: A hierarchical Bayesian regression analysis. Transp. Res. Part D Transp. Environ. 2014, 26, 20–26. [Google Scholar] [CrossRef]
  13. Xing, Y.; Lv, C.; Zhang, Z.; Wang, H.; Na, X.; Cao, D.; Velenis, E.; Wang, F.Y. Identification and Analysis of Driver Postures for In-Vehicle Driving Activities and Secondary Tasks Recognition. IEEE Trans. Comput. Soc. Syst. 2018, 5, 95–108. [Google Scholar] [CrossRef]
  14. Lin, Z.; Qiu, T.; Liu, P.; Zhang, L.; Zhang, S.; Mu, Z. Fatigue driving recognition based on deep learning and graph neural network. Biomed. Signal Processing Control. 2021, 68, 102598. [Google Scholar] [CrossRef]
  15. Du, G.; Li, T.; Li, C.; Liu, P.X.; Li, D. Vision-Based Fatigue Driving Recognition Method Integrating Heart Rate and Facial Features. IEEE Trans. Intell. Transp. Syst. 2021, 22, 3089–3100. [Google Scholar] [CrossRef]
  16. Mandal, B.; Li, L.; Wang, G.S.; Lin, J. Towards Detection of Bus Driver Fatigue Based on Robust Visual Analysis of Eye State. IEEE Trans. Intell. Transp. Syst. 2017, 18, 545–557. [Google Scholar] [CrossRef]
  17. Nakisa, B.; Rastgoo, M.N.; Tjondronegoro, D.; Chandran, V. Evolutionary computation algorithms for feature selection of EEG-based emotion recognition using mobile sensors. Expert Syst. Appl. 2018, 93, 143–155. [Google Scholar] [CrossRef]
  18. Fernández, A.; Usamentiaga, R.; Carús, J.L.; Casado, R. Driver Distraction Using Visual-Based Sensors and Algorithms. Sensors 2016, 16, 1805. [Google Scholar] [CrossRef]
  19. Jiang, L.; Xie, W.; Zhang, D.; Gu, T. Smart Diagnosis: Deep Learning Boosted Driver Inattention Detection and Abnormal Driving Prediction. IEEE Internet Things J. 2022, 9, 4076–4089. [Google Scholar] [CrossRef]
  20. Hossain, M.U.; Rahman, A.M.; Islam, M.M.; Akhter, A.; Uddin, A.M.; Paul, M.B. Automatic driver distraction detection using deep convolutional neural networks. Intell. Syst. Appl. 2022, 14, 200075. [Google Scholar] [CrossRef]
  21. Chen, Z.; Rahman, M.A.; Islam, M.M.; Akhter, A. Dangerous driving behavior detection using video-extracted vehicle trajectory histograms. J. Intell. Transp. Syst. 2017, 21, 409–421. [Google Scholar] [CrossRef]
  22. Chen, C.; Egas, F.D.; Silva, F.M.; Paul, P.M. Effects of Eurycoma longifolia Jack supplementation combined with resistance training on isokinetic muscular strength and power, anaerobic power, and urinary testosterone: Epitestosterone ratio in young males. Int. J. Prev. Med. 2019, 10, 118. [Google Scholar] [CrossRef]
  23. Galarza, E.E.; Wang, X.; Tang, X. Real Time Driver Drowsiness Detection Based on Driver’s Face Image Behavior Using a System of Human Computer Interaction Implemented in a Smartphone. In Proceedings of the International Conference on Information Technology & Systems (ICITS 2018), Libertad, Ecuador, 10–12 January 2018. [Google Scholar]
  24. Sun, Y.; Wang, X.; Tang, X. Deep Convolutional Network Cascade for Facial Point Detection. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013. [Google Scholar]
  25. Zhou, E.; Fan, H.; Cao, Z.; Jiang, Y.; Yin, Q. Extensive Facial Landmark Localization with Coarse-to-Fine Convolutional Network Cascade. In Proceedings of the 2013 IEEE International Conference on Computer Vision Workshops, Sydney, NSW, Australia, 2–8 December 2013. [Google Scholar]
  26. Taskiran, M.; Kahraman, N.; Erdem, C.E. Face recognition: Past, present and future (a review). Digit. Signal Processing 2020, 106, 102809. [Google Scholar] [CrossRef]
  27. Xia, J.; Zhang, H.; Wen, S.; Yang, S.; Xu, M. An efficient multitask neural network for face alignment, head pose estimation and face tracking. Expert Syst. Appl. 2022, 205, 117368. [Google Scholar] [CrossRef]
  28. Xing, Y.; Lv, C.; Cao, D. (Eds.) Chapter 6—Application of Deep Learning Methods in Driver Behavior Recognition. In Advanced Driver Intention Inference; Elsevier: Amsterdam, The Netherlands, 2020; pp. 135–156. [Google Scholar]
  29. Li, L.; Zhong, B.; Hutmacher, C.; Liang, Y.; Horrey, W.J.; Xu, X. Detection of driver manual distraction via image-based hand and ear recognition. Accid. Anal. Prev. 2020, 137, 105432. [Google Scholar] [CrossRef]
  30. Zaidi, S.S.A.; Ansari, M.S.; Aslam, A.; Kanwal, N.; Asghar, M.; Lee, B. A survey of modern deep learning based object detection models. Digit. Signal Processing 2022, 126, 103514. [Google Scholar]
  31. Tong, K.; Wu, Y. Deep learning-based detection from the perspective of small or tiny objects: A survey. Image Vis. Comput. 2022, 123, 104471. [Google Scholar] [CrossRef]
  32. Dong, X.; Yan, S.; Duan, C. A lightweight vehicles detection network model based on YOLOv5. Eng. Appl. Artif. Intell. 2022, 113, 104914. [Google Scholar] [CrossRef]
  33. Guo, X.; Li, S.; Yu, J.; Zhang, J.; Ma, J.; Ma, L.; Liu, W.; Ling, H. PFLD: A Practical Facial Landmark Detector. arXiv 2019, arXiv:1902.10859. [Google Scholar]
  34. Jocher, G. Yolov5. 2020. Available online: https://github.com/ultralytics/yolov5 (accessed on 12 September 2020).
  35. Kong, S.G.; Mbouna, R.O. Head Pose Estimation From a 2D Face Image Using 3D Face Morphing with Depth Parameters. IEEE Trans. Image Processing 2015, 24, 1801–1808. [Google Scholar] [CrossRef]
  36. Jourabloo, A.; Liu, X. Pose-Invariant 3D Face Alignment. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  37. Kumar, A.; Chellappa, R. Disentangling 3D Pose in a Dendritic CNN for Unconstrained 2D Face Alignment. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  38. Yadav, N.K.; Singh, S.K.; Dubey, S.R. MobileAR-GAN: MobileNet-Based Efficient Attentive Recurrent Generative Adversarial Network for Infrared-to-Visual Transformations. IEEE Trans. Instrum. Meas. 2022, 71, 1–9. [Google Scholar] [CrossRef]
  39. Sun, J.; Lu, S. An Improved Single Shot Multibox for Video-Rate Head Pose Prediction. IEEE Sens. J. 2020, 20, 12326–12333. [Google Scholar] [CrossRef]
  40. Xing, L.-M.; Zhang, Y.-J. Forecasting crude oil prices with shrinkage methods: Can nonconvex penalty and Huber loss help? Energy Econ. 2022, 110, 106014. [Google Scholar] [CrossRef]
  41. Cao, Z.; Liao, T.; Song, W.; Chen, Z.; Li, C. Detecting the shuttlecock for a badminton robot: A YOLO based approach. Expert Syst. Appl. 2021, 164, 113833. [Google Scholar] [CrossRef]
  42. Meyer, P.J. Reachability Analysis of Neural Networks Using Mixed Monotonicity. IEEE Control. Syst. Lett. 2022, 6, 3068–3073. [Google Scholar] [CrossRef]
  43. Cheng, J.; Liu, P.; Zhang, Q.; Ma, H.; Wang, F.; Zhang, J. Real-Time and Efficient 6-D Pose Estimation from a Single RGB Image. IEEE Trans. Instrum. Meas. 2021, 70, 1–14. [Google Scholar] [CrossRef]
  44. Qi, J.; Liu, X.; Liu, K.; Xu, F.; Guo, H.; Tian, X.; Li, M.; Bao, Z.; Li, Y. An improved YOLOv5 model based on visual attention mechanism: Application to recognition of tomato virus disease. Comput. Electron. Agric. 2022, 194, 106780. [Google Scholar] [CrossRef]
  45. Yang, Q.; Yu, D.; Zhang, Z.; Yao, Y.; Chen, L. Spatiotemporal Trident Networks: Detection and Localization of Object Removal Tampering in Video Passive Forensics. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 4131–4144. [Google Scholar] [CrossRef]
  46. Kong, W.; Hong, J.; Jia, M.; Yao, J.; Cong, W.; Hu, H.; Zhang, H. YOLOv3-DPFIN: A Dual-Path Feature Fusion Neural Network for Robust Real-Time Sonar Target Detection. IEEE Sens. J. 2020, 20, 3745–3756. [Google Scholar] [CrossRef]
  47. Huang, Y.; Tang, Z.; Chen, D.; Su, K.; Chen, C. Batching Soft IoU for Training Semantic Segmentation Networks. IEEE Signal Processing Lett. 2020, 27, 66–70. [Google Scholar] [CrossRef]
  48. NVIDIA. Jetson Xavier NX. 2019. Available online: https://developer.nvidia.com/embedded/jetson-xavier-nx (accessed on 14 September 2020).
  49. Jourabloo, A.; Ye, M.; Liu, X.; Ren, L. Pose-Invariant Face Alignment with a Single CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  50. Xiao, S.; Feng, J.; Xing, J.; Lai, H.; Yan, S.; Kassim, A. Robust Facial Landmark Detection via Recurrent Attentive-Refinement Networks. In Computer Vision—ECCV 2016; Springer International Publishing: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  51. Dong, X.; Yan, Y.; Ouyang, W.; Yang, Y. Style Aggregated Network for Facial Landmark Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  52. Liu, M.; Li, Z.; Li, Y.; Liu, Y. A Fast and Accurate Method of Power Line Intelligent Inspection Based on Edge Computing. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
Figure 1. A diagram of the driver abnormal behavior analysis library.
Figure 1. A diagram of the driver abnormal behavior analysis library.
Symmetry 14 01956 g001
Figure 2. The structure diagram of the driver abnormal behavior recognition system.
Figure 2. The structure diagram of the driver abnormal behavior recognition system.
Symmetry 14 01956 g002
Figure 3. The structure of the driver abnormal behavior detection model.
Figure 3. The structure of the driver abnormal behavior detection model.
Symmetry 14 01956 g003
Figure 4. The location information of facial key point annotation.
Figure 4. The location information of facial key point annotation.
Symmetry 14 01956 g004
Figure 5. Key points for the right eye and mouth using the facial key point location information annotation: (a) the right eye contains six key points and (b) the mouth contains six key points.
Figure 5. Key points for the right eye and mouth using the facial key point location information annotation: (a) the right eye contains six key points and (b) the mouth contains six key points.
Symmetry 14 01956 g005
Figure 6. The PFLD network structure diagram.
Figure 6. The PFLD network structure diagram.
Symmetry 14 01956 g006
Figure 7. The YOLOv5 network structure diagram.
Figure 7. The YOLOv5 network structure diagram.
Symmetry 14 01956 g007
Figure 8. A flow chart of the experimental design.
Figure 8. A flow chart of the experimental design.
Symmetry 14 01956 g008
Figure 9. Labeled images representing abnormal driver behavior: (a) smoking, (b) making a phone call, (c) drinking water, and (d) obstructed driving.
Figure 9. Labeled images representing abnormal driver behavior: (a) smoking, (b) making a phone call, (c) drinking water, and (d) obstructed driving.
Symmetry 14 01956 g009
Figure 10. The TensorRT execution process.
Figure 10. The TensorRT execution process.
Symmetry 14 01956 g010
Figure 11. The PR, MR and FR values for the NX side model: (a) the PR, MR, and FR values of 1417 pieces of randomly selected data; (b) the PR, MR, and FR values of 1016 pieces of randomly selected data.
Figure 11. The PR, MR and FR values for the NX side model: (a) the PR, MR, and FR values of 1417 pieces of randomly selected data; (b) the PR, MR, and FR values of 1016 pieces of randomly selected data.
Symmetry 14 01956 g011
Figure 12. The visual interface of the system operation.
Figure 12. The visual interface of the system operation.
Symmetry 14 01956 g012
Table 1. The parameter settings of the PFLD model training.
Table 1. The parameter settings of the PFLD model training.
ParameterSetting
Images_size112 × 112
FrameworkPytorch
Batch_size256
Weight decay10−6
Momentum0.9
Learning rate10−4
Table 2. The comparison results on the 300 W Common dataset.
Table 2. The comparison results on the 300 W Common dataset.
ModulePIFA-CNNRDRPCD-CNNSANPFLD
Interocular normalization (ION)5.47 [49]5.03 [50]3.67 [37]3.34 [51]3.11
Table 3. The parameter settings of the YOLOv5 model training.
Table 3. The parameter settings of the YOLOv5 model training.
ParameterSetting
Images_size640 × 640
FrameworkPytorch
Batch_size64
Weight decay5 × 10−4
Momentum0.937
Learning rate10−2
Table 4. The confusion matrix.
Table 4. The confusion matrix.
Predictive ValueActual Value
PositiveNegative
TrueTPTN
FalseFPFN
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, S.; Wang, X.; Ji, H.; Wang, L.; Hou, Z. A Novel Driver Abnormal Behavior Recognition and Analysis Strategy and Its Application in a Practical Vehicle. Symmetry 2022, 14, 1956. https://doi.org/10.3390/sym14101956

AMA Style

Liu S, Wang X, Ji H, Wang L, Hou Z. A Novel Driver Abnormal Behavior Recognition and Analysis Strategy and Its Application in a Practical Vehicle. Symmetry. 2022; 14(10):1956. https://doi.org/10.3390/sym14101956

Chicago/Turabian Style

Liu, Shida, Xuyun Wang, Honghai Ji, Li Wang, and Zhongsheng Hou. 2022. "A Novel Driver Abnormal Behavior Recognition and Analysis Strategy and Its Application in a Practical Vehicle" Symmetry 14, no. 10: 1956. https://doi.org/10.3390/sym14101956

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop