Next Article in Journal
Feelings of Polish and Palestinian Students after Receiving Vaccinations against COVID-19
Next Article in Special Issue
Identifying Risk Factors for Autos and Trucks on Highway-Railroad Grade Crossings Based on Mixed Logit Model
Previous Article in Journal
Physical Fitness, Screen Time and Sleep Habits According to Obesity Levels in Schoolchildren: Findings from the Health Survey of the Extreme South of Chile
Previous Article in Special Issue
Comparative Analysis of Influencing Factors on Crash Severity between Super Multi-Lane and Traditional Multi-Lane Freeways Considering Spatial Heterogeneity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparing Resampling Algorithms and Classifiers for Modeling Traffic Risk Prediction

1
School of Highway, Chang’an University, Xi’an 710064, China
2
School of Civil and Environmental Engineering, Nanyang Technological University, Singapore 639798, Singapore
3
Engineering Research Center of Highway Infrastructure Digitalization, Ministry of Education, Xi’an 710000, China
4
School of Engineering, STEM College, RMIT University, Melbourne, VIC 3001, Australia
5
College of Transportation Engineering, Chang’an University, Xi’an 710064, China
*
Authors to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2022, 19(20), 13693; https://doi.org/10.3390/ijerph192013693
Submission received: 24 September 2022 / Revised: 18 October 2022 / Accepted: 18 October 2022 / Published: 21 October 2022
(This article belongs to the Special Issue Road Traffic Safety Risk Analysis)

Abstract

:
Road infrastructure has significant effects on road traffic safety and needs further examination. In terms of traffic crash prediction, recent studies have started to develop deep learning classification algorithms. However, given the uncertainty of traffic crashes, predicting the traffic risk potential of different road sections remains a challenge. To bridge this knowledge gap, this study investigated a real-world expressway and collected its traffic crash data between 2013 and 2020. Then, according to the time-spatial density ratio (Pts), road sections were assigned into three classes corresponding to low, medium, and high risk levels of traffic. Next, different classifiers were compared that were trained using the transformed and resampled feature data to construct a traffic crash risk prediction model. Last, but not least, partial dependence plots (PDPs) were employed to interpret the results and analyze the importance of individual features describing the geometry, pavement, structure, and weather conditions. The results showed that a variety of data balancing algorithms improved the performance of the classifiers, the ensemble classifier superseded the others in terms of the performance metrics, and the combined SMOTEENN and random forest algorithms improved the classification accuracy the most. In the future, the proposed traffic crash risk prediction method will be tested in more road maintenance and design safety assessment scenarios.

1. Introduction

The expressway (which is a class of very high standard road in China) is an essential part of road transportation systems and is also the road type with the highest casualty rate from traffic crashes. According to relevant statistics for China, the annual death toll per 1000 km of expressways is 43.68, which is 4 times that of ordinary roads, and the direct property damage per traffic crash on expressways is US$6500, which is 10 times that of ordinary roads [1]. In particular, the safety level of highways in mountainous areas is worse due to poor alignment conditions [2], high proportion of tunnels [3], complex weather [4], etc. In 2018 alone, there were more than 70,000 crashes in the mountainous areas in China, resulting in more than 100,000 casualties and millions of dollars in losses [5].
Traffic risk evaluation is an important basis for road safety management, which is an important subject in road safety research. Identification of the priorities of road section management through traffic risk evaluation can help designers and managers to efficiently allocate resources and improve road safety [6,7]. The operating environment of an expressway is not affected by pedestrians and personal mobility devices such as bicycles and scooters. There is a high correlation between traffic crashes and road design features [8]. Presently, there are two principal methods for traffic risk evaluation. One method involves the identification of crash-prone sections based on crash data [9]. Sufficient observation time and sample size are the keys to accurate identification of crash-prone road sections. Such an approach entails a certain lag and is hence difficult to incorporate into the design and early stages of roadway operation. The other approach involves the identification of high-risk sections based on road feature analysis [10]. However, though high-risk sections have a greater potential of becoming crash-prone sections, some may not develop into crash-prone sections. Furthermore, this approach only considers individual factors such as road alignment, weather, pavement, etc., and the identified high-risk road sections are often not sufficiently comprehensive. Given the constraints in the afore-mentioned methods, the development of road traffic risk assessment methods that make better use of crash data and with more comprehensive consideration of risk elements is of significant research and application interest.
Newer modeling techniques have been developed in the analysis of motor vehicle crash data, including two recent methods of clustering and classification [11,12,13]. Clustering methods are mainly used to discover potential risk factors and quantify their weight [14]. Traffic risk prediction generally uses decision trees and classifiers. Commonly used classification models include the decision tree classifier [15], rule induction PART [16], lazy classifier [17], Bayes classifier [18], etc. Compared with traditional traffic risk prediction methods, the application of decision tree and classifiers is exploratory rather than inferential. Hereby, more risk factors can be considered, including potential risk factors with unclear impact mechanisms.
The number of road sections at different risk levels typically varies. The spatial distribution of traffic crashes on roads shows different degrees of aggregation given that the occurrence of crashes is affected by various risk sources. Many scholars have taken note of the inherent imbalance in crash datasets, and have proposed balancing methods such as under-sampling, oversampling, and mixed sampling. Using these three methods to deal with imbalanced datasets, Mujalli et al. [19] used the Bayesian classifier to predict the severity of traffic crashes, which mainly considered the number of vehicles, crash mode, speed, lighting, and road conditions. Fiorentini et al. [20] used the random under-sampling of majority class (RUMC) technique to deal with imbalanced datasets to predict the severity of crashes wherein the reliability of the predictions was compared for the four models of random tree, K-nearest neighbor (KNN), logistic regression, and random forest. Danesh et al. [21] applied biogeography-based optimization and invasive weed optimization to deal with imbalanced datasets, and applied the three algorithms of decision tree, support vector machine (SVM), and KNN to construct a prediction model of crash severity. Chen et al. [22] innovatively proposed the ENN-SMOTE-Tomek Link (EST) algorithm to deal with the imbalance problem in the prediction of vehicle spatial sequence features.
At present, classification algorithms are used to predict the severity of traffic collisions with traffic crash datasets with imbalanced characteristics. There is a lack of relevant research on the identification of high-risk road sections [23]. Compared with the prediction of collision severity, the imbalance problem in road section risk prediction is more serious. Data imbalance is not only due to the inherent imbalance of the risk level of road sections but also the existence of potentially high-risk sections that are yet to develop into crash-prone sections [24]. Therefore, new methods need to be proposed to deal with imbalanced crash datasets for road section risk prediction using machine learning algorithms.
In this study, a prediction model of traffic crash risk potential was established. Three types of data-balancing methods (11 algorithms in total) were introduced to deal with traffic-related data. The processed balanced dataset was applied on 7 classifiers and 84 classification models were obtained. Finally, the performance of the established models was compared, and the optimal resampling algorithm and classifier for expressway risk potential prediction were proposed.
The remaining paper is organized as follows: Section 2 presents a brief description of the data used, data-balancing methods, and classifiers. Section 3 presents the findings. Section 4 gives the conclusion.

2. Materials and Methods

In this study, traffic crash data and road data on Chinese expressways were collected, which resulted in a dataset that is inherently imbalanced. Unbalanced data can easily lead to the prediction results being biased towards the majority class samples, resulting in a falsely high accuracy of the prediction results [25]. Herein, resampling algorithms and classifiers were applied to improve the negative impact of the class imbalance problem on the model performance. Considering that the potential risk of some roads may be improved by traffic safety facilities, resulting in the predicted road risk level being higher than the actual risk level, a new performance measure was proposed to evaluate the performance of the trained model. Figure 1 shows the procedures employed.

2.1. Data

In this study, traffic crash data and road data were collected for the GYX Expressway in southwest China, which were provided by GYX Expressway’s traffic police and expressway operating companies. A total of 4187 traffic crash data records were collected covering a period of 8 years (2013–2020). Crash data included details such as the crash time, crash location, etc. The total length of the two-way four-lane expressway is 210 km. Roadway data included road alignment, pavement condition, tunnels, and expressway facilities. In addition, daily weather data for 8 years (2013–2020) was obtained from the weather website.
The expressway was divided into 420 sections according to the driving direction and kilometers. In order to study the traffic risk of road sections under different weather conditions, four weather conditions were considered: sunny, cloudy, overcast, and rainy. Therefore, there were a total of 1680 sampled units within this study as shown in Table 1. Feature variables (road and weather information) and target variables (crash information) were characterized for each sampled unit.

2.1.1. Data Integrity and Accuracy

Traffic crash data is an important basis for road safety research. The recording of traffic crashes is generally the responsibility of road management companies and traffic police. In China, each expressway has a dedicated agency responsible for managing traffic crashes. Before the advent of mobile phones and monitoring systems, it was difficult to contact the traffic police on the expressway in a timely manner. Drivers would often resolve crashes themselves without going through the traffic police whenever minor traffic crashes occurred. Therefore, in the past, there was a serious lack of traffic crash data for minor crashes. With the popularity of mobile phones and monitoring systems, drivers can now readily contact the traffic police department, resulting in the timely detection of traffic crashes by the police authority. At the same time, traffic crash insurance compensation in China requires the traffic police to issue a liability determination letter. Therefore, practically all traffic crashes on the expressway are recorded by the traffic police department.
The crash data in this study were provided by traffic police and highway management companies, respectively, and included details of the time, location, direction, crash type, and casualties. Herein, the traffic police provided traffic crash data from 2013–2020, with the time and the cause of the crash being recorded. However, the highway management company only provided data for the two years of 2017 and 2018, with recording of the amount of road damage and compensation. The analysis of the crash data entailed two aspects of completeness and accuracy, as follows:
(1). Completeness of crashes: In the comparison of the crash records of 2017 and 2018 provided by the traffic police and highway management companies, no missing incidents were found. In addition, there were many crashes in which vehicles slightly scratched the road guardrails.
(2). Accuracy of crashes: The time of the crash was accurate to the minute. The location information in some crash records in 2013 and 2014 was only accurate to the kilometer level. In the crash records after 2015, the location information was accurate to the 10 m level.
In essence, the crash data used in this study had good completeness and accuracy, thereby laying a good foundation for risk analysis and risk prediction.

2.1.2. Feature Variables

Relevant studies have demonstrated that road alignment, pavement state, tunnels, and expressway facilities are key factors affecting the level of road safety [26,27]. Lee et al. [28] found that factors such as road horizontal alignment, vertical alignment, road surface type, and weather have a significant impact on crash severity. Schloegl et al. [29] used random forests and boosted trees to predict the occurrence of crashes and concluded that the application of feature variables, including road alignment, pavement conditions, traffic volume, and weather variables, can effectively predict the occurrence of traffic crashes.
Altogether, 17 road features and one weather feature were selected as independent variables, with the selected variables being used in the literature and available in our datasets (Table 2).
Among the variables, the maximum curve radius index (RIMX) and RICU (cumulative radius index) were first proposed in this study, and the calculation formulas are as follows:
R I M X = M A X ( R I n )
R I C U = R I n
R I = { R * R 1 ( R < R * ) ( R R * )
where R I is an indicator of the radius of the horizontal alignment. The value ranges from 1 to infinity, and the larger the value, the smaller the radius. R is the radius of the horizontal alignment. R * is the minimum radius without superelevation. According to Chinese standards, when the design speed is 80 km/h, R * is 2500 m.

2.1.3. Road Risk Level

In order to quantify the traffic risk of different weather and road sections, this study proposed a new traffic crash statistical index, the temporal-spatial density ratio ( P t s ), as the target variable. The basic premise of P t s is to characterize the relative safety level of the road section. Considering the spatial influence of the section length on the crash frequency, the number of crashes per kilometer was calculated. Considering the temporal influence of weather on the crash frequency, the number of crashes per day under each weather condition was calculated. Finally, the relative safety level of the road section was characterized by the temporal-spatial density ratio, as defined in Equation (4):
P t s = N x e / L e / D x / V x N / L / D / V
where P t s is the temporal-spatial density ratio, N x e is the number of crashes with x weather in the spatial unit, L e is the length of the spatial unit, D x is the number of days with x weather, N is the total number of crashes, L is the total length of the road network, and D is the total number of days.
When P t s = 1, it means that the crash rate of this road section in given weather is at the average level of the road network. When P t s =3, it means that the crash rate of this road section in given weather is 3 times the average level. Herein, the road risk can be divided into three levels: when P t s < 1, the section’ risk level is L_1; when 1 ≤ P t s < 3, the section risk level is L_2; and when P t s ≥ 3, the section risk level is L_3. P t s of 1680 units was enumerated, as shown in Figure 2.

2.2. Data Balancing

When there is a large difference in the number of samples among the classes, the problem of class imbalance occurs. A dataset with such a problem is usually called an imbalanced dataset. The imbalance ratio ( I R ) value is commonly used to measure the degree of imbalance in a dataset. The I R value is the ratio of the number of samples in the majority class to the number of samples in the minority class. In this study, the ratio of road section data for the three risk levels is about 12:2:1, and the I R value is 12, which is an extremely imbalanced dataset.
When using the classifiers, imbalanced data can cause degradation of the performance of the classifier [30]. A popular solution is to balance the number of samples of each category by sampling the original data when the sample size cannot be expanded. The data balancing process can be carried out using three methods: under-sampling, oversampling, and mixed sampling. Under-sampling is used to reduce the majority class samples to achieve equilibrium. Oversampling is used to increase the minority class samples to achieve equilibrium. Mixed sampling combines both methods and can handle both majority and minority class samples. In this study, 11 resampling algorithms were selected from 3 data balancing methods: under-sampling (4), oversampling (5), and mixed sampling (2), as described in Table 3.

2.3. Classifier Selection

The classification algorithm is a type of supervised learning. The training of a model with known classes of data can classify unknown data. Classification algorithms are widely used in traffic crash data mining. K-nearest neighbors (KNN), support vector machine (SVM), Bayesian, and Ensemble algorithms are the four most widely used methods [38,39].

2.3.1. KNN Classifier

The KNN classifier performs classification according to the calculated distance, which is the distance from an unlabeled object to all labeled objects [40]. Its advantage is simple, and the application has good adaptability in most scenarios. Since the algorithm relies on the distance for classification, it is necessary to normalize feature variables with large differences in their physical units or scales [41]. However, normalization cannot completely solve the shortcomings of the KNN algorithm’s complete dependence on distance, especially the lack of accuracy in the voting decisions.
According to the number of neighbors, KNN can be divided into fine KNN, medium KNN, and coarse KNN [42]. The number of neighbors in fine KNN is 1, which describes a slight gap between sample classes. Since the value gap of CULR, GAE, and other feature variables in the data of this study is very small, fine KNN is used for classification. Yigit [43] introduced weight on the basis of distance and proposed weighted KNN, an improved KNN algorithm. Kuang et al. [44] applied the weighted KNN algorithm to improve the accuracy of duration prediction in traffic crashes. Finally, in this study, fine KNN and weighted KNN were selected to represent the KNN method for model training.

2.3.2. SVM Classifier

SVM achieves sample classification by constructing hyperplanes in multi-dimensional space. The distance from the sample point to the hyperplane is called the interval. The interval is the distance from the sample point to the hyperplane. The goal of SVM is to find an optimal hyperplane with the largest interval from each sample [45]. SVM was originally designed to solve the two-class problem, but the present study belongs to the three-class problem. To make the SVM algorithm suitable for this study, the one-versus-rest mode was adopted to train the classifier in two steps: (1) Divide the dataset into two parts: L_1 level and non-L_1 to train the first SVM classifier; (2) using L_2 and L_3 data, train a second SVM classifier.
To train the model with the SVM classifier, a kernel function needs to be used to transform the space for nonlinear data. Commonly used kernel functions are the radial basis function, Gaussian (RBF) kernel, and cubic (polynomial degree 3) kernel, etc. [46]. Finally, in this study, cubic SVM and fine Gaussian SVM were selected to represent the SVM method for model training.

2.3.3. Ensemble Classifier

The ensemble classifier classifies samples by combining multiple models. The basic idea is to learn a set of classifiers and decide the classification result based on their votes. This method has a better predictive performance than the use of a single model. Compared with KNN and SVM, its advantage is that it does not need to normalize the data. There are two commonly used ensemble classifiers: bagging (bootstrap aggregation) and random forest. Bagging classifiers are used to reduce the variance of decision trees [47]. Random forest is an extension to bagging. Each classifier in the ensemble is a decision tree classifier and uses randomly chosen attributes at each node to determine the split. When classifying, each tree votes and returns the most popular class. Malik et al. [48] compared the performance of six classifiers in predicting the severity of road crashes. The study found that random forests, decision trees, and bagging are significantly better than other algorithms in all performance measures. In addition, XGBoost is an emerging algorithm that combines the loss function and the regularization term to build an overall loss function. Shi et al. [49,50] established a driving risk assessment method based on XGBoost, and the risk prediction accuracy rate can reach 89%. On this basis, a risk prediction method for autonomous vehicle decision-making was further established. The prediction accuracy reached 91.7%. Parsa et al. [51] used XGBoost to predict traffic crashes and reported that XGBoost has a more robust effect. Finally, in this study, bagging and random forest were selected to represent the ensemble classifier for model training.

2.3.4. Bayesian Classifier

The Bayesian classifier is a classification method based on Bayes’ theorem. The algorithm decides the classification result by minimizing the probability of misclassification. The naive Bayesian classifier is the simplest and most common classification method in the Bayes classifier [52]. The classification result can be described by an explicit network, and each feature variable is a network node. The naive Bayesian classifier requires input feature variables to be independent of each other [53]. Some feature variables in the horizontal alignment and vertical alignment in this study are not independent of each other. Therefore, the naive Bayes classifier is not suitable for this study.
In order to relax the conditional independence assumption that governs standard naive Bayes learning algorithms, scholars have proposed various solutions [54]. Chen et al. [55] used a Bayesian network to study the causes of vehicle crashes, but this method requires the construction of explicit topological relationships between random variables. Another solution is the AODE algorithm (averaged one-dependence estimators), which does not require explicit topological relationships and improves the classification accuracy at the cost of a modest increase in computation [56]. Finally, in this study, the AODE algorithm was selected to represent the Bayesian classifier training model.

2.4. Performance Evaluation Measures

The confusion matrix is an important method used to evaluate the classification effect of a model. The commonly used two-class confusion matrix divides classification results into four categories: true positives (TPs), true negatives (TNs), false positives (FPs), and false negatives (FNs). According to the four results, five measures of accuracy, sensitivity, specificity, precision, and F1 were proposed. This study is a three-classification problem, and the confusion matrix is shown in Figure 3. The equations of commonly used performance measures are as follows:
Accuracy :   A c c = i = j N i , j N i , j
Precision :   P P V n = N n , n N 1 , n + N 2 , n + N 3 , n ( n = 1 , 2 , 3 )
Sensitivity :   T P R n = N n , n N n , 1 + N n , 2 + N n , 3 ( n = 1 , 2 , 3 )
Specificity :   T N R n = i n j n N i , j N i , j N n , j ( n = 1 , 2 , 3 )
F 1 ( n ) = 2 × P P V n × T P R n P P V n + T P R n ( n = 1 , 2 , 3 )
where N i , j is the number of road sections, i is the observed risk level of the section, and j is the predicted risk potential level of the section.
When i = j, N i , j is located on the diagonal of the confusion matrix, which means that the observed risk level of the road section is consistent with the predicted risk level. When i < j, N i , j is located in the upper right triangle of the matrix, which means that the observed risk level of the road section is lower than the predicted risk level. When i > j, N i , j is located in the lower left triangle of the matrix, indicating that the observed risk level of the road section is greater than the predicted risk level.
Road crash risk is affected by many factors, which can be divided into 59 risk factors such as road alignment, pavement, environment, presence of work zones, cross-section, and traffic control [57]. This study focused on the impact of inherent properties on road risk. The feature variables considered, except for the road surface state, are all inherent properties of the road. The prediction results characterize the traffic risk potential caused by the inherent properties of the road. The positive impact of safety control measures and other factors on traffic risk was not considered. Meanwhile, traffic crashes have a certain randomness, including the crash location and time. Therefore, there is a significant difference between the number of crashes on high-risk road sections and low-risk road sections under a larger observation time and sample size. When the observed risk level of the section is lower than the predicted risk level, the prediction result is not necessarily inaccurate. For example, some road sections with L_3 traffic risk potential are subject to traffic control measures and their observed risk level is L_2 or L_1. When the observed risk level of the section is greater than the predicted risk level, the prediction result is less affected by the inherent properties of the road, and the risk source may be attributed in part to other factors. Considering the characteristics of this study, a new performance measure, Score, was proposed to evaluate the predicted results, as defined in Equation (10):
Score :   S = A c c + T P R 1 + T P R 2 3

3. Results and Discussion

In this section, we focus on the following points for the purpose of finding the best method for predicting the traffic risk potential of the road:
  • Analysis of the correlation of 18 kinds of feature variables, retaining the necessary feature variables to construct a dataset.
  • Comparison of the improvement effects of 11 different resampling algorithms on model training and identification of the best resampling algorithm.
  • Comparison of the impact of seven classification algorithms on model performance and identification of the best classification algorithm.
  • Use of the best algorithms and classifiers to analyze the contribution of various feature variables in model training and identification of the key risk factors that affect road risk.

3.1. Feature Analysis and Dimensionality Reduction

The focus of feature selection is to select a subset of variables from the input that can efficiently describe the input data while reducing the effects from noise or irrelevant variables and still provide good prediction results [58]. The Pearson correlation coefficient is one of the commonly used indicators in feature selection [59]. The removal of redundant data by correlation analysis can make each input variable relatively independent.
There were 18 features variables in this study, among which tunnel (TNN), interchange (ITC), service area (SVA), and weather (WET) were categorical variables and had no direct relationship with the other studied variables. The remaining 14 variables were continuous variables. Using the Pearson correlation coefficient as an indicator, correlation analysis of the continuous variables was carried out, and the analysis results are shown in Figure 4.
It can be seen that the correlation coefficient between the straight length ratio (STLR) and curve length ratio (CULR) in the horizontal alignment variable is −0.75. Considering that a gentle curve has a greater impact on the driver than a straight section, STLR was excluded from the feature variables. In addition, the correlation coefficients of the cumulative radius index (RICU) with the number of horizontal change points (HPNE) and maximum curve radius index (RIMX) are 0.84 and 0.82, respectively. So, HPNE and RIMX were excluded from the feature variables. The correlation coefficient between the average grade (GAE) and maximum grade (GMX), GAE and downhill length ratio (DWLR), and GMX and DWLR in the vertical line shape variable is 0.97, −0.92, and −0.9, respectively. GAE can represent GMX and DWLR to a great extent, so GMX and DWLR were excluded. Finally, 13 feature variables were retained, including 4 categorical variables and 9 continuous variables.

3.2. Imbalanced Versus Balanced Datasets

The original dataset included 1680 road sections, and 80% of the samples were selected as training data (1260 road sections in total). The sample size distribution of the training data was: 1093 L_1 risk sections, 168 L_2 risk sections, and 83 L_3 risk sections. To address the imbalanced dataset problem, under-sampling, oversampling, and mixed balanced datasets were developed using 12 different resampling techniques. The normalization algorithm was used in the establishment of the dataset. Table 4 shows the totals in all used datasets and their distribution among different risk classes. It can be seen that except for the ALLKNN and ENN algorithms, the other algorithms basically achieved the goal of data re-collection. When the ENN algorithm was used to process unbalanced data, the sample size of L_2 risk road sections in the processed dataset was abnormally reduced. However, the algorithm combining SMOTE and ENN (i.e., SMOTEENN) can effectively improve the effect of data re-collection.

3.3. Classifier Performance

Based on the 12 datasets in Table 4, we trained the different models using the 7 classifiers described in Section 2.3. We randomly selected 80% of the original dataset as training data and the remaining 20% as the test set. The random assignment process was performed 10 times, with a total of 10 sets of “training + testing” data. Each training data was balanced with 11 resampling techniques. Based on the processed dataset and the original dataset, 7 kinds of classifiers were used for model training, and a total of 84 groups of models (10 in each group) were obtained. In order to verify the validity of the model of the method, we applied the trained model to predict the test data and compare the observed and predicted results. The performance of each group of models was evaluated. The mean and standard deviation of each of the six performance indicators were calculated, as shown in Appendix A (Table A1).
Then, we calculated the scores of all models using Equation (10). All results were ranked according to Score (the average of Score was used first, and when the averages were equal, it was sorted according to the standard deviation), as shown in Appendix B (Table A2). Table 5 lists the top 10 results.
With respect to the results obtained by the testing set, the following findings were extracted:
  • Evaluating the performance of a model based solely on the accuracy, precision, sensitivity, and specificity did not result in a suitable model. For example, the accuracy of the model trained by the SVM algorithm in the original dataset is 81%, the precision of L_1 is 81%, the sensitivity is 100%, and the specificity is 0. However, when the precision and sensitivity of L_2 and L_3 are both 0, the specificity is 100%. Obviously, it cannot meet the needs of road section risk classification.
  • The F1 indicator can integrate the accuracy, precision, sensitivity, and specificity, and collectively evaluate the classification effect of a single category. According to F1(1), it is found that the original dataset and the under-sampling dataset are biased towards the majority class (L_1). According to F1(2) and F1(3), it is found that the muse ensemble classifier (XGBoost or random forest) can effectively improve the classification effect of minority categories (L_2 and L_3).
  • Not all data balancing algorithms improve the classification performance. Meanwhile, whether the data is balanced hardly affects the ranking order of the classifier performance.
  • The ensemble classifier is used in the top 10 method combinations in the Score index. Among them, the model trained by SMOTEENN and random forest has the highest score.
Using Score to evaluate the model performance, the bias of the results is unclear. Although Score can theoretically take into account both the accuracy and sensitivity, there may be two types of classification effects that are excellent while the other is poor. In order to verify the Score evaluation results, the probability distributions of the prediction results of the models trained by random forest and XGBoost were compared, as shown in Figure 5.
The expected result is that the predicted probability of the positive class (the predicted result is consistent with the observed result) is higher than that of the other two negative classes. Similar to the distribution of the observed L_3 of random forest in Figure 6, the predicted probability of L_3 is almost always very high. It shows that there is a certain regularity in the traffic crash risk level at L_3, and the prediction model can be effectively classified. It can be seen that the classification effect of the random forest model is significantly better than that of the XGBoost model. These results are consistent with the trend of the Score values, indicating that the use of Score to evaluate the classification effect is reliable, and can balance the sensitivity of the three risk levels and the overall accuracy.
According to the results of Score in Appendix A, the tree-based ensemble classifiers outperform the KNN and SVM methods in dealing with the three-class problem in this study. The reason may be that KNN and SVM are “distance”-based classification algorithms, which are not applicable to a traffic crash dataset, even if the data are normalized. Both the random forest and XGBoost showed better results in this study. This is consistent with the findings of Schloegl et al. [29] and Parsa et al. [51]. This shows that the ensemble method has a better application prospect in traffic crash-related research.
At the same time, it is found that various resampling methods can improve the classification effect of the random forest and XGBoost. It shows that the choice of classifier is more important than the data balance method. In this study, the model trained by the random forest method based on the SMOTEENN-balanced dataset is the best, which is consistent with the results of Mujalli [19]. This again shows that the mixed resampling method can deal with the imbalance problem in traffic crashes more effectively.
Schloegl [60] studied the impact of environmental factors on the occurrence of traffic crashes and found that XGBoost has a higher classification effect than random forest. This is inconsistent with the analysis results of this study. The reason may lie in the small sample size and the existence of potentially high-risk road sections, resulting in a large proportion of noise in the data. Compared with random forest, XGBoost is prone to overfitting due to noisy samples. However, the effect of XGBoost may improve with an increased number of feature variables and samples.
Finally, SMOTEENN and random forest were selected to build a traffic crash risk prediction model of expressway.

3.4. Feature Importance and Interpretation

Random forest is essentially a tree model, and its classification method is nonlinear. Feature importance measures alone do not represent an impact on classification. However, the ranking of feature importance represents its relative influence. Three feature importance indicators were used in this study: Gini importance, permutation importance, and SHAP importance. The Gini importance represents the probability of each feature in the tree node, and the sum of its values is 1. The larger the value, the greater the importance of the feature in the node. However, if there is a correlation between multiple features, the feature that participates in the classification first will have a higher importance, and the feature value that participates in the classification later will be reduced. So, this easily results in an unstable order of influence. Permutation importance is used to analyze the feature importance by disrupting the data of a feature after model training is completed and observing whether the model accuracy is reduced. When the value is less than 0, it means that the observation accuracy is improved after the data is scrambled. However, in this study, the accuracy cannot evaluate the effect of the model, so it cannot truly reflect the feature importance. SHAP (Shapley additive explanation) is proposed based on game theory. The SHAP baseline value of a feature is calculated to indicate its importance, and its value is consistent with the influence weight value of the feature, which can better describe the classification contribution of the feature [61]. Feature importance analysis was carried out based on the trained model, as shown in Figure 6.
In Figure 6, the permutation importance is relatively unstable. For example, the permutation importance of SRI is inconsistent with the other two importance indicators. The ranking results of Gini and the SHAP importance are basically the same. Two horizontal alignment feature variables, RICU and STLR, have high importance. Among the three vertical alignment feature variables of GAE, GDF, and VPNE, GAE and GDF are more important while VPNE is less important. The four road condition indicators of PCI, RDI, SRI, and RQI have high importance. The three structural feature indicators of TNN, ITC, and SVA can be ignored except for TNN. The results show that the traffic risk potential of this study section is highly correlated with horizontal alignment, vertical alignment, road surface conditions, and weather, and is less affected by expressway facilities.
This study focuses on the prediction of traffic risk based on inherent road properties rather than crash rate prediction. Inherent road properties are only part of the contributions to crash risk. It should be emphasized that the purpose of this study was to optimize road design and road maintenance programs. Therefore, the model in this study does not need a particularly high accuracy, but the influence of feature variables on the prediction results must be consistent with the theory of road risk.
The feature importance analysis results show which feature variables are more important. However, its influence law is not clear, and the influence law of some factors may be inconsistent with the theory.
The influence of inherent road attributes on the traffic risk potential was analyzed, and the reliability of the proposed traffic crash risk prediction method was verified in this study. The partial dependence was introduced to analyze the influence of feature variables on the prediction results. The larger its value, the greater the influence of the feature variables on the results. It is worth noting that RICU, GAE, and PCI are the three most important feature variables, representing the horizontal alignment, vertical alignment, and pavement status, respectively. The partial dependence plot under the interaction of RICU, GAE, and PCI is shown in Figure 7.
According to Figure 7, we can find the following points:
  • When RICU is greater than 30 and GAE is less than 0, the partial dependence value of the road section with the traffic risk level L_3 is greater than 0.32. At the same time, when RICU is less than 5 and GAE is greater than 0.02, the partial dependence value of the road section with the traffic risk level L_1 is greater than 0.65. This shows that the model established in this paper can effectively identify two types of typical high-risk road sections: downhill small radius curves or downhill continuous curved road sections.
  • When RICU is greater than 20 and PCI is less than 98, the partial dependence value of the road section with the traffic risk level L_3 is greater than 0.32. However, when RICU is less than 5 and PCI is less than 98, the partial dependence value of the road section with the traffic risk level L_1 is greater than 0.64. This shows that in the model established in this paper, the pavement condition has a greater impact on the driving safety of small-radius road sections or continuous curved road sections while it has a lesser impact on the driving safety of straight sections.
  • When GAE is less than 0 and PCI is less than 95, the partial dependence value of the road section with the traffic risk level L_3 is greater than 0.30. However, when GAE is greater than 0.02 and PCI is greater than 0.4, the partial dependence value of the road section with the traffic risk level L_1 is greater than 0.5. This is basically consistent with the above two rules. At the same time, this also shows that in the model established in this paper, the downhill section has higher requirements regarding the pavement performance, and the higher the uphill slope, the safer the road section.
The results show that the larger the RICU (the smaller the radius of the curve or the higher the number of curves), the higher the risk level of the road section. The smaller the GAE (downhill and the greater the grade), the higher the risk level of the road section. The smaller the PCI (more severe road damage), the higher the risk level of the road section. The classification principle of the model is consistent with the traffic risk mechanism and the results of Elvik [62]. This shows that the traffic crash risk prediction model is reliable.
Compared to the study by Schlögl M [29], the variables in this study are very simple in terms of weather. However, the results also show that weather is an important variable after the road alignment and pavement condition. However, the impact of tunnels and expressway facilities on traffic crashes seems to be very low, which may be due to three aspects. First, the data sample size was small and there are only nine interchanges and two service areas in this study. Second, this study excluded the crash data of ramps, and some interchange entrance and exit crashes may be recorded as ramp crashes. Third, the service area itself has a low impact on traffic crashes.

4. Conclusions

For highway management to develop precise and effective solutions, the traffic risk potential must be predicted accurately. The inherent properties of the road are the main factors affecting the traffic risk potential, which require many variables to be described. Machine learning techniques offer an efficient method for analyzing many variables at once, which is difficult to achieve with traditional data analysis techniques. Therefore, by evaluating the performance of several resampling approaches and classifiers, this research developed a new model for predicting the risk potential of highway traffic. The main conclusions are as follows:
  • In predicting the risk potential of expressway, horizontal alignment, vertical alignment, pavement, and weather play a significant role. This study provided 18 feature variables from 5 different perspectives, and 13 feature variables were retained after dimensionality reduction. Incorporating Gini importance, permutation importance, and SHAP importance, three feature importance indicators, the contribution of the feature variables to the model was ranked. The results showed that 11 of these feature variables have a high contribution to the prediction of the risk potential.
  • The ensemble classifier demonstrated a good performance in processing traffic accident data, and the addition of a resampling algorithm further enhanced the classifier’s performance. By comparing the prediction results of XGBoost and the random forest algorithm, it was demonstrated that the model performance index Score provided in this research can effectively assess the performance of the three-level traffic risk prediction model.
  • The combination of SMOTEENN and the random forest algorithm developed the best model for predicting the highway traffic risk potential. The three most representative feature variables, RICU, GAE, and PCI, were analyzed using the partial dependence plot, and the results showed that the classification principle of the established traffic crash risk prediction model was consistent with the objective risk effect law.
This research demonstrated the effectiveness of employing machine learning algorithms to assess traffic crash risk. In addition, it provided recommendations for data balancing and classification algorithm selection for machine learning applications based on the inherent features of roads. This research proposed a method for traffic crash risk prediction, which may be utilized to enhance road design and maintenance planning.
In the future, the sample size and feature variables of the dataset will be further expanded for a more comprehensive analysis of the impact on road crash risk. At the same time, an interpretable classifier will be applied in depth to study the mechanism of inherent road features and traffic risk.

Author Contributions

Methodology, C.Z. and B.W.; software, M.Z. and B.W.; supervision, C.Z. and Y.D.W.; writing—original draft preparation, B.W. and Y.X.; writing—review and editing, Y.D.W. and L.H.; Funding acquisition, C.Z.; Project administration, C.Z. and Y.D.W. All authors have read and agreed to the published version of the manuscript.

Funding

Key Technologies Research and Development Program of China (No.2020YFC1512005), Key Research and Development Program of Sichuan Province (No.2022YFG0048), Science and Technology Project of Sichuan Transportation Department (No.2019-ZL-12, No.2022-ZL-04), Key Research and Development Program of Shanxi Province (No. 202102020101014).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Acknowledgments

The authors would like to express their sincere thanks to the anonymous reviewers for their constructive comments on an earlier version of this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Comparison between the different models using performance measure.
Table A1. Comparison between the different models using performance measure.
DatasetFine KNNWeighted KNNCubic SVMFine Gaussian SVMXGBoostRandom ForestAODE
MeanS.DMeanS.DMeanS.DMeanS.DMeanS.DMeanS.DMeanS.D
ACC
Original0.700.020.760.010.810.000.810.000.780.010.770.010.500.04
PG0.410.020.340.010.270.030.330.000.150.020.190.010.350.05
RUS0.380.020.340.050.250.110.260.090.430.020.440.020.350.04
ENN0.760.010.800.000.810.000.810.000.800.010.800.010.440.07
ALLKNN0.720.010.800.000.810.000.810.000.780.000.790.010.420.05
ROS0.700.020.540.060.310.070.370.070.750.010.740.020.350.05
SMOTE0.630.030.520.070.320.080.310.080.750.010.750.020.350.05
ADASYN0.620.040.530.060.320.080.300.080.760.010.760.010.350.04
Borderline-SMOTE0.640.030.550.050.380.060.370.060.760.010.750.020.370.05
SMOTENC0.630.030.530.060.320.070.300.080.750.020.750.020.350.04
SMOTEENN0.380.030.300.040.190.040.160.040.540.040.570.030.220.04
SMOT-Tomek Links0.620.040.510.060.320.070.300.070.750.010.750.020.350.04
PPV1
Original0.820.010.830.010.810.000.810.000.850.000.850.000.830.01
PG0.810.010.840.020.810.030.860.010.860.060.900.010.850.03
RUS0.820.020.800.020.650.070.640.150.890.010.880.010.850.01
ENN0.820.000.820.000.810.000.810.000.840.000.840.000.830.01
ALLKNN0.820.000.830.000.810.000.810.000.840.010.840.010.830.01
ROS0.820.010.820.030.750.050.760.060.860.010.860.010.850.01
SMOTE0.820.010.810.030.750.040.740.050.850.000.850.010.840.01
ADASYN0.830.010.820.020.740.050.720.060.850.000.850.010.850.01
Borderline-SMOTE0.820.010.820.020.800.020.790.010.850.010.850.000.830.01
SMOTENC0.830.010.810.020.740.050.720.060.850.010.860.010.850.01
SMOTEENN0.810.010.770.030.690.080.780.080.850.010.870.010.860.03
SMOT-Tomek Links0.820.010.810.020.740.050.730.060.850.010.850.010.850.01
PPV2
Original0.140.030.110.04////0.260.090.260.050.300.17
PG0.100.010.100.02//0.450.200.120.010.110.010.220.14
RUS0.090.020.080.02//0.250.170.120.020.120.010.070.04
ENN////////0.050.040.070.06//
ALLKNN0.190.05//////0.130.060.240.08//
ROS0.140.030.150.040.040.010.060.030.250.070.240.050.320.18
SMOTE0.120.020.120.030.040.020.040.020.230.080.270.070.170.12
ADASYN0.130.030.130.030.040.010.040.010.260.090.270.070.290.18
Borderline-SMOTE0.130.020.120.030.060.020.040.020.240.080.220.060.150.12
SMOTENC0.130.030.130.030.040.020.040.010.260.100.270.080.260.17
SMOTEENN0.090.010.090.010.080.020.090.020.140.020.180.020.190.10
SMOT-Tomek Links0.140.020.130.020.040.020.040.010.220.050.220.070.150.09
PPV3
Original0.150.070.160.09////0.510.040.480.090.090.01
PG0.090.020.080.010.080.010.080.010.090.010.110.010.080.00
RUS0.070.020.070.010.070.010.070.010.160.010.170.010.080.00
ENN0.120.040.140.08////0.360.020.340.020.090.01
ALLKNN0.090.030.150.08////0.320.030.320.020.090.00
ROS0.150.070.150.050.110.010.110.010.410.050.460.090.080.00
SMOTE0.140.060.150.060.120.010.110.010.440.050.470.060.080.00
ADASYN0.160.060.150.060.120.010.110.000.430.040.480.080.080.00
Borderline-SMOTE0.140.060.160.050.120.010.100.010.480.080.450.060.080.01
SMOTENC0.160.050.140.060.120.010.090.010.450.040.450.060.080.00
SMOTEENN0.080.020.080.020.100.000.080.010.210.010.270.030.080.00
SMOT-Tomek Links0.130.050.120.050.120.010.090.010.450.060.480.070.080.00
TPR1
Original0.830.020.920.021.000.001.000.000.910.020.890.010.560.06
PG0.440.020.350.020.250.050.340.000.070.010.120.010.360.06
RUS0.400.030.360.070.250.150.260.120.430.030.450.030.380.05
ENN0.920.010.980.011.000.001.000.000.950.010.950.010.490.09
ALLKNN0.860.020.970.011.000.001.000.000.930.010.930.010.460.06
ROS0.830.020.610.070.320.090.400.090.860.020.850.030.370.06
SMOTE0.730.040.590.080.330.100.320.100.880.020.870.020.370.06
ADASYN0.720.040.590.070.330.100.310.100.880.020.880.020.370.05
Borderline-SMOTE0.740.040.620.060.410.080.400.080.880.020.870.020.400.06
SMOTENC0.730.040.590.070.320.090.310.100.880.020.860.020.370.05
SMOTEENN0.410.040.290.040.140.060.110.050.580.060.600.050.180.06
SMOT-Tomek Links0.720.040.570.070.320.090.310.090.870.020.860.020.370.05
TPR2
Original0.130.020.060.030.000.000.000.000.120.020.170.020.130.08
PG0.240.030.250.050.120.050.070.030.380.040.370.020.060.04
RUS0.240.050.210.080.170.070.130.050.330.030.340.030.010.01
ENN0.000.000.000.000.000.000.000.000.000.000.010.010.000.00
ALLKNN0.040.010.010.010.000.000.000.000.020.010.060.020.000.00
ROS0.130.020.270.060.170.060.130.060.200.020.230.030.020.01
SMOTE0.170.020.230.030.180.060.170.060.150.010.210.030.020.01
ADASYN0.190.030.250.040.180.060.160.060.160.010.200.030.020.01
Borderline-SMOTE0.180.020.220.030.160.050.140.050.150.020.170.010.010.01
SMOTENC0.190.030.260.040.180.060.170.060.140.020.220.020.020.01
SMOTEENN0.290.020.310.030.290.080.300.080.310.030.410.040.200.06
SMOT-Tomek Links0.210.030.270.040.180.060.160.060.160.020.180.020.020.01
TPR3
Original0.150.090.140.090.000.000.000.000.370.050.340.060.530.08
PG0.360.090.450.090.760.060.730.060.780.050.750.050.800.04
RUS0.350.090.370.080.470.120.480.110.630.060.630.070.760.05
ENN0.170.080.140.090.000.000.000.000.380.050.370.050.680.09
ALLKNN0.180.080.140.090.000.000.000.000.350.060.370.050.750.05
ROS0.150.090.260.070.480.070.510.080.400.070.400.060.770.06
SMOTE0.200.090.250.090.490.070.480.080.360.070.380.050.750.06
ADASYN0.230.080.260.080.490.070.480.080.360.060.380.060.770.06
Borderline-SMOTE0.190.080.230.070.470.080.460.100.390.070.340.060.720.06
SMOTENC0.220.070.240.080.510.080.470.100.360.050.380.060.760.05
SMOTEENN0.270.080.340.090.570.090.540.110.470.040.530.060.750.06
SMOT-Tomek Links0.190.080.230.080.510.080.480.100.390.060.400.060.760.05
TNR1
Original0.230.040.150.050.000.000.000.000.290.040.300.030.490.06
PG0.540.050.700.040.790.020.750.030.960.010.940.010.730.05
RUS0.620.050.640.030.720.110.760.060.760.020.730.040.700.04
ENN0.120.030.090.030.000.000.000.000.210.020.210.030.560.09
ALLKNN0.170.030.100.030.010.010.000.000.210.030.240.030.600.06
ROS0.230.040.450.050.640.050.600.050.380.050.390.050.720.04
SMOTE0.330.040.430.040.630.060.630.060.310.040.350.040.700.04
ADASYN0.350.050.450.050.620.070.630.060.310.030.340.040.720.04
Borderline-SMOTE0.310.030.410.030.570.060.570.050.310.040.320.030.650.04
SMOTENC0.340.040.440.040.620.060.610.060.310.040.370.040.710.04
SMOTEENN0.580.020.660.050.830.050.860.060.560.040.600.050.830.06
SMOT-Tomek Links0.340.030.450.040.620.060.610.060.320.040.340.040.710.04
TNR2
Original0.880.010.940.011.000.001.000.000.920.020.920.010.900.04
PG0.680.010.670.020.830.070.910.070.590.030.540.030.970.02
RUS0.680.030.640.070.630.120.670.120.640.030.640.010.940.04
ENN0.990.000.990.001.000.001.000.000.990.010.990.011.000.00
ALLKNN0.970.010.990.001.000.001.000.000.980.010.970.010.990.00
ROS0.880.010.720.060.570.130.680.140.880.030.880.030.950.04
SMOTE0.800.030.700.060.570.120.570.120.900.020.890.020.930.04
ADASYN0.800.030.710.050.570.120.560.120.900.020.900.020.940.04
Borderline-SMOTE0.810.030.710.060.620.090.630.090.900.020.890.020.930.04
SMOTENC0.810.030.720.050.570.120.570.120.900.020.890.020.940.04
SMOTEENN0.600.030.530.040.500.110.500.110.690.060.700.050.750.06
SMOT-Tomek Links0.800.030.700.050.580.120.580.120.900.020.890.020.940.04
TNR3
Original0.950.000.980.011.000.001.000.000.970.010.970.010.670.05
PG0.760.020.660.020.420.070.420.070.490.040.580.030.390.05
RUS0.710.010.710.020.610.100.570.090.780.020.800.020.430.06
ENN0.930.010.980.011.000.001.000.000.950.010.950.010.490.10
ALLKNN0.890.010.980.011.000.001.000.000.950.010.950.010.470.06
ROS0.950.000.880.020.740.050.710.050.960.010.960.010.420.06
SMOTE0.920.010.880.020.750.040.740.050.970.010.970.010.440.06
ADASYN0.910.010.870.020.750.040.740.040.970.010.970.010.420.06
Borderline-SMOTE0.930.010.910.010.780.040.760.040.970.010.970.010.470.06
SMOTENC0.920.010.880.020.750.040.730.040.970.010.970.010.430.06
SMOTEENN0.800.010.750.000.630.060.600.070.890.010.900.010.430.06
SMOT-Tomek Links0.910.010.870.020.740.040.730.040.970.010.970.010.430.06
F1(1)
Original0.830.010.870.010.900.000.900.000.880.010.870.010.660.04
PG0.570.010.490.020.370.060.490.000.130.030.220.010.500.06
RUS0.540.030.480.070.280.14//0.580.020.590.020.510.05
ENN0.870.000.890.000.900.000.900.000.890.000.890.000.590.07
ALLKNN0.840.010.890.000.900.000.900.000.880.000.880.010.580.05
ROS0.830.010.690.060.420.090.500.100.860.010.850.010.510.05
SMOTE0.770.030.670.060.430.100.420.100.860.010.860.010.510.05
ADASYN0.770.030.680.050.420.100.410.100.860.010.860.010.510.05
Borderline-SMOTE0.780.020.700.050.520.070.510.070.860.010.860.010.530.05
SMOTENC0.770.030.680.060.420.100.410.100.860.010.860.010.510.04
SMOTEENN0.540.040.410.050.220.080.180.080.680.040.710.030.270.08
SMOT-Tomek Links0.770.030.660.060.420.090.410.090.860.010.860.010.510.04
F1(2)
Original0.130.02//////0.150.030.200.02//
PG0.140.010.140.03////0.180.020.160.01//
RUS0.130.03//////0.170.020.180.01//
ENN//////////////
ALLKNN0.060.01////////////
ROS0.130.020.180.04////0.200.030.220.03//
SMOTE0.140.020.150.02////0.170.020.210.02//
ADASYN0.150.030.160.03////0.190.030.200.03//
Borderline-SMOTE0.150.020.150.030.070.02//0.170.030.180.02//
SMOTENC0.150.030.170.03////0.160.020.230.03//
SMOTEENN0.140.010.140.010.120.030.130.030.180.010.240.02//
SMOT-Tomek Links0.160.020.170.02////0.170.020.190.03//
F1(3)
Original////////0.410.030.370.040.160.01
PG0.140.030.130.020.150.010.140.010.170.010.190.010.150.00
RUS0.120.030.120.020.120.010.120.020.250.020.260.020.150.00
ENN////////0.350.010.340.010.150.01
ALLKNN////////0.310.020.330.020.160.01
ROS//0.180.060.180.010.170.010.390.050.410.050.150.01
SMOTE0.170.070.180.070.190.010.180.010.380.060.400.040.150.01
ADASYN0.190.060.180.070.180.010.170.010.380.040.410.050.150.00
Borderline-SMOTE//0.180.060.190.010.160.030.410.060.370.030.150.01
SMOTENC0.180.060.170.070.190.010.150.030.380.030.400.050.150.00
SMOTEENN0.120.030.130.030.160.010.130.020.290.010.350.040.150.00
SMOT-Tomek Links//0.160.060.180.010.160.030.400.050.420.060.150.00
Score
Original0.330.040.320.040.270.000.270.000.420.030.430.030.390.07
PG0.340.040.350.050.380.050.370.030.440.040.440.030.400.04
RUS0.320.050.310.070.300.100.290.080.460.040.470.040.370.03
ENN0.310.030.320.030.270.000.270.000.390.020.390.020.370.05
ALLKNN0.310.030.320.030.270.000.270.000.380.020.400.030.390.03
ROS0.330.040.360.060.320.070.340.070.450.040.460.040.380.04
SMOTE0.330.050.330.060.330.070.320.070.420.030.440.030.370.04
ADASYN0.350.050.350.060.330.070.310.070.430.030.440.040.380.04
Borderline-SMOTE0.330.040.330.050.340.060.320.070.430.040.420.030.370.04
SMOTENC0.350.050.340.060.330.070.310.080.420.030.450.030.380.04
SMOTEENN0.310.040.320.050.350.070.330.080.440.040.510.040.390.05
SMOT-Tomek Links0.340.050.340.060.330.070.310.080.430.030.440.030.380.04

Appendix B

Table A2. Ranking results based on Score.
Table A2. Ranking results based on Score.
DatasetFine KNNWeighted KNNCubic SVMFine Gaussian SVMXGBoostRandom ForestAODERank
Original5963848217152711
PG484430351112211
RUS64767778323710
ENN7569828123243413
ALLKNN7266798028222512
ROS5939674964293
SMOTE54525865187367
ADASYN42416170169314
Borderline SMOTE515550621419388
SMOTENC42455674205325
SMOTEENN73684053101262
SMOT-Tomek Links46475771138326
Rank5467213/

References

  1. Yang, D.; Zhang, X.; Zhang, Y.; Liu, D.; Chang, M.; Zhang, L. Comparative study on factors for injury severity between highway and roadway motor vehicle crashes in China, 2004–2015. J. Third Mil. Med. Univ. 2017, 39, 589–596. [Google Scholar]
  2. Zhang, H.; Zhang, M.; Zhang, C.; Hou, L. Formulating a Gis-Based Geometric Design Quality Assessment Model for Mountain Highways. Accid. Anal. Prev. 2021, 157, 106172. [Google Scholar] [CrossRef] [PubMed]
  3. Yeung, J.S.; Wong, Y.D. Road traffic accidents in Singapore expressway tunnels. Tunn. Undergr. Space Technol. 2013, 38, 534–541. [Google Scholar] [CrossRef]
  4. Xu, C.; Wang, W.; Liu, P. Identifying crash-prone traffic conditions under different weather on freeways. J. Saf. Res. 2013, 46, 135–144. [Google Scholar] [CrossRef]
  5. National Bureau of Statistics. China Statistical Yearbook; China Statistics Press: Beijing, China, 2019. Available online: http://www.stats.gov.cn/tjsj/ndsj/2019/indexeh.htm (accessed on 14 October 2022).
  6. Chai, H.; Xie, J.; Li, X. A Policy Review of Road Safety Infrastructure Facilities in China. In Proceedings of the Asia-Pacific Conference on Intelligent Medical (APCIM)/7th International Conference on Transportation and Traffic Engineering (ICTTE), Beijing, China, 21–23 December 2018. [Google Scholar]
  7. Shi, X.; Wong, Y.D.; Li, M.Z.F.; Chai, C. Key risk indicators for accident assessment conditioned on pre-crash vehicle trajectory. Accid. Anal. Prev. 2018, 117, 346–356. [Google Scholar] [CrossRef] [PubMed]
  8. Wang, C.; Quddus, M.A.; Ison, S.G. The effect of traffic and road characteristics on road safety: A review and future research direction. Saf. Sci. 2013, 57, 264–275. [Google Scholar] [CrossRef]
  9. Ghadi, M.; Torok, A. A comparative analysis of black spot identification methods and road accident segmentation methods. Accid. Anal. Prev. 2019, 128, 1–7. [Google Scholar] [CrossRef]
  10. Yannis, G.; Dragomanovits, A.; Laiou, A.; La Torre, F.; Domenichini, L.; Richter, T.; Ruhl, S.; Graham, D.; Karathodorou, N. Road traffic accident prediction modelling: A literature review. Proc. Inst. Civ. Eng. -Transp. 2017, 170, 245–254. [Google Scholar] [CrossRef]
  11. Gutierrez-Osorio, C.; Pedraza, C. Modern data sources and techniques for analysis and forecast of road accidents: A review. J. Traffic Transp. Eng. -Engl. Ed. 2020, 7, 432–446. [Google Scholar] [CrossRef]
  12. Santos, K.; Dias, J.P.; Amado, C. A literature review of machine learning algorithms for crash injury severity prediction. J. Saf. Res. 2022, 80, 254–269. [Google Scholar] [CrossRef]
  13. Chen, T.; Shi, X.; Wong, Y.D. A lane-changing risk profile analysis method based on time-series clustering. Phys. A: Stat. Mech. Its Appl. 2021, 565, 125567. [Google Scholar] [CrossRef]
  14. Shi, X.; Wong, Y.D.; Chai, C.; Li, M.Z.-F.; Chen, T.; Zeng, Z. Automatic clustering for unsupervised risk diagnosis of vehicle driving for smart road. IEEE Trans. Intell. Transp. Syst. 2022, 23, 17451–17465. [Google Scholar] [CrossRef]
  15. Chang, L.-Y.; Chen, W.-C. Data mining of tree-based models to analyze freeway accident frequency. J. Saf. Res. 2005, 36, 365–375. [Google Scholar] [CrossRef]
  16. Krishnaveni, S.; Hemalatha, M. A perspective analysis of traffic accident using data mining techniques. Int. J. Comput. Appl. 2011, 23, 40–48. [Google Scholar] [CrossRef]
  17. Tiwari, P.; Dao, H.; Gia Nhu, N. Performance evaluation of lazy, decision tree classifier and multilayer perceptron on traffic accident analysis. Inform. -J. Comput. Inform. 2017, 41, 39–46. [Google Scholar]
  18. Sohn, S.Y.; Lee, S.H. Data fusion, ensemble and clustering to improve the classification accuracy for the severity of road traffic accidents in Korea. Saf. Sci. 2003, 41, 1–14. [Google Scholar] [CrossRef]
  19. Mujalli, R.O.; López, G.; Garach, L. Bayes classifiers for imbalanced traffic accidents datasets. Accid. Anal. Prev. 2016, 88, 37–51. [Google Scholar] [CrossRef] [PubMed]
  20. Fiorentini, N.; Losa, M. Handling imbalanced data in road crash severity prediction by machine learning algorithms. Infrastructures 2020, 5, 61. [Google Scholar] [CrossRef]
  21. Danesh, A.; Ehsani, M.; Nejad, F.M.; Zakeri, H. Prediction model of crash severity in imbalanced dataset using data leveling methods and metaheuristic optimization algorithms. Int. J. Crashworthiness 2022, 1–14. [Google Scholar] [CrossRef]
  22. Chen, T.; Shi, X.; Wong, Y.D.; Yu, X. Predicting lane-changing risk level based on vehicles’ space-series features: A pre-emptive learning approach. Transp. Res. Part C-Emerg. Technol. 2020, 116, 102646. [Google Scholar] [CrossRef]
  23. Silva, P.B.; Andrade, M.; Ferreira, S. Machine learning applied to road safety modeling: A systematic literature review. J. Traffic Transp. Eng. -Engl. Ed. 2020, 7, 775–790. [Google Scholar] [CrossRef]
  24. Li, L.; Zhu, L.; Sui, D.Z. A GIS-based Bayesian approach for analyzing spatial–temporal patterns of intra-city motor vehicle crashes. J. Transp. Geogr. 2007, 15, 274–285. [Google Scholar] [CrossRef]
  25. Chen, T.; Shi, X.; Wong, Y.D. Key feature selection and risk prediction for lane-changing behaviors based on vehicles’ trajectory data. Accid. Anal. Prev. 2019, 129, 156–169. [Google Scholar] [CrossRef] [PubMed]
  26. Wong, Y.D.; Nicholson, A. Driver behaviour at horizontal curves: Risk compensation and the margin of safety. Accid. Anal. Prev. 1992, 24, 425–436. [Google Scholar] [CrossRef]
  27. Malyshkina, N.V.; Mannering, F.L.; Tarko, A.P. Markov switching negative binomial models: An application to vehicle accident frequencies. Accid. Anal. Prev. 2009, 41, 217–226. [Google Scholar] [CrossRef] [Green Version]
  28. Lee, J.-Y.; Chung, J.-H.; Son, B. Analysis of traffic accident size for Korean highway using structural equation models. Accid. Anal. Prev. 2008, 40, 1955–1963. [Google Scholar] [CrossRef] [PubMed]
  29. Schloegl, M.; Stuetz, R.; Laaha, G.; Melcher, M. A comparison of statistical learning methods for deriving determining factors of accident occurrence from an imbalanced high resolution dataset. Accid. Anal. Prev. 2019, 127, 134–149. [Google Scholar] [CrossRef]
  30. Mazurowski, M.A.; Habas, P.A.; Zurada, J.A.; Lo, J.Y.; Baker, J.A.; Tourassi, G.D. Training neural network classifiers for medical decision making: The effects of imbalanced datasets on classification performance. Neural Netw. 2008, 21, 427–436. [Google Scholar] [CrossRef] [Green Version]
  31. Triguero, I.; Derrac, J.; Garcia, S.; Herrera, F. A taxonomy and experimental study on prototype generation for nearest neighbor classification. IEEE Trans. Syst. Man Cybern. Part C-Appl. Rev. 2012, 42, 86–100. [Google Scholar] [CrossRef]
  32. Wilson, D.L. Asymptotic properties of nearest neighbor rules using edited data. IEEE Trans. Syst. Man Cybern. 1972, 3, 408–421. [Google Scholar] [CrossRef] [Green Version]
  33. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  34. Han, H.; Wang, W.Y.; Mao, B.H. Borderline-SMOTE: A new over-sampling method in imbalanced data sets learning. In Proceedings of the International Conference on Intelligent Computing, Hefei, China, 23–26 August 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 878–887. [Google Scholar]
  35. He, H.; Bai, Y.; Garcia, E.A.; Li, S. ADASYN: Adaptive synthetic sampling approach for imbalanced learning. In Proceedings of the International Joint Conference on Neural Networks, Hong Kong, China, 1–8 June 2008. [Google Scholar]
  36. Batista, G.E.; Bazzan, A.L.; Monard, M.C. Balancing training data for automated annotation of keywords: A case study. WOB 2003, 10–18. Available online: https://www.inf.ufrgs.br/maslab/masbio/papers/balancing-training-data-for.pdf (accessed on 23 September 2022).
  37. Batista, G.E.; Prati, R.C.; Monard, M.C. A study of the behavior of several methods for balancing machine learning training data. ACM SIGKDD Explor. Newsl. 2004, 6, 20–29. [Google Scholar] [CrossRef]
  38. Cuenca, L.G.; Puertas, E.; Aliane, N.; Andres, J.F. Traffic Accidents Classification and Injury Severity Prediction. In Proceedings of the 2018 3rd IEEE International Conference on Intelligent Transportation Engineering (ICITE), Singapore, 3–5 September 2018; pp. 52–57. [Google Scholar]
  39. Al Mamlook, R.E.; Kwayu, K.M.; Alkasisbeh, M.R.; Frefer, A.A. Comparison of Machine Learning Algorithms for Predicting Traffic Accident Severity. In Proceedings of the IEEE Jordan International Joint Conference on Electrical Engineering and Information Technology (JEEIT), Amman, Jordan, 9–11 April 2019. [Google Scholar]
  40. Guo, G.D.; Wang, H.; Bell, D.; Bi, Y.X.; Greer, K. KNN model-based approach in classification. In On the move to meaningful internet systems 2003: CoopIS, DOA, and ODBASE; Meersman, R., Tari, Z., Schmidt, D.C., Eds.; Springer: Berlin/Heidelberg, Germany, 2003; Volume 2888, pp. 986–996. [Google Scholar]
  41. Dudani, S.A. The distance-weighted K-nearest-neighbor rule. IEEE Trans. Syst. Man Cybern. 1976, 4, 325–327. [Google Scholar] [CrossRef]
  42. Chen, W.; Wang, F.; Yan, K.; Zhang, Y.; Wu, D. Micro-stereolithography of KNN-based lead-free piezoceramics. Ceram. Int. 2019, 45, 4880–4885. [Google Scholar] [CrossRef]
  43. Yigit, H. A weighting approach for KNN classifier. In Proceedings of the 10th International Conference on Electronics, Computer and Computation (ICECCO), Turgut Ozal University, Ankara, Turkey, 7–9 November 2013. [Google Scholar]
  44. Kuang, L.; Yan, H.; Zhu, Y.; Tu, S.; Fan, X. Predicting duration of traffic accidents based on cost-sensitive Bayesian network and weighted K-nearest neighbor. J. Intell. Transp. Syst. 2019, 23, 161–174. [Google Scholar] [CrossRef]
  45. Altman, N.S. An introduction to kernel and nearest-neighbor nonparametric regression. Am. Stat. 1992, 46, 175–185. [Google Scholar]
  46. Bagasta, A.R.; Rustam, Z.; Pandelaki, J.; Nugroho, W.A. Comparison of cubic SVM with Gaussian SVM: Classification of infarction for detecting ischemic stroke. In Proceedings of the 9th Annual Basic Science International Conference (BaSIC)—Recent Advances in Basic Sciences Toward 4.0 Industrial Revolution, Brawijaya University, Malang, Indonesia, 20–21 May 2019. [Google Scholar]
  47. Zareapoor, M.; Shamsolmoali, P. Application of Credit Card Fraud Detection: Based on Bagging Ensemble Classifier. In Proceedings of the 1st International Conference on Intelligent Computing, Communication and Convergence (ICCC), Bhubaneshwar, India, 27–28 December 2014. [Google Scholar]
  48. Malik, S.; El Sayed, H.; Khan, M.A.; Khan, M.J. Road Accident Severity Prediction—A Comparative Analysis of Machine Learning Algorithms. In Proceedings of the IEEE Global Conference on Artificial Intelligence and Internet of Things (GCAIoT), Dubai, United Arab Emirates, 12–16 December 2021. [Google Scholar]
  49. Shi, X.; Wong, Y.D.; Li, M.Z.-F.; Palanisamy, C.; Chai, C. A feature learning approach based on XGBoost for driving assessment and risk prediction. Accid. Anal. Prev. 2019, 129, 170–179. [Google Scholar] [CrossRef]
  50. Shi, X.; Wong, Y.D.; Chai, C.; Li, M.Z.-F. An automated machine learning (AUTOML) method of risk prediction for decision-making of autonomous vehicles. IEEE Trans. Intell. Transp. Syst. 2021, 22, 7145–7154. [Google Scholar] [CrossRef]
  51. Parsa, A.B.; Movahedi, A.; Taghipour, H.; Derrible, S.; Mohammadian, A. Toward safer highways, application of XGBoost and SHAP for real-time accident detection and feature analysis. Accid. Anal. Prev. 2020, 136, 105405. [Google Scholar] [CrossRef]
  52. Leung, K.M. Naive Bayesian classifier. Polytech. Univ. Dep. Comput. Sci./Financ. Risk Eng. 2007, 2007, 123–156. [Google Scholar]
  53. Yang, K.; Wang, X.; Yu, R. A Bayesian dynamic updating approach for urban expressway real-time crash risk evaluation. Transp. Res. Part C-Emerg. Technol. 2018, 96, 192–207. [Google Scholar] [CrossRef]
  54. Flores, M.J.; Gámez, J.A.; Martínez, A.M. Supervised classification with Bayesian networks: A review on models and applications. Intell. Data Anal. Real-Life Appl. Theory Pract. 2012, 72–102. [Google Scholar] [CrossRef]
  55. Chen, T.; Wong, Y.D.; Shi, X.; Wang, X. Optimized structure learning of Bayesian network for investigating causation of vehicles? On-road crashes. Reliab. Eng. Syst. Saf. 2022, 224, 108527. [Google Scholar] [CrossRef]
  56. Mizianty, M.; Kurgan, L.; Ogiela, M. Comparative Analysis of the Impact of Discretization on the Classification with Naive Bayes and Semi-Naive Bayes Classifiers. In Proceedings of the 7th International Conference on Machine Learning and Applications, San Diego, CA, USA, 11–13 December 2008. [Google Scholar]
  57. Papadimitriou, E.; Filtness, A.; Theofilatos, A.; Ziakopoulos, A.; Quigley, C.; Yannis, G. Review and ranking of crash risk factors related to the road infrastructure. Accid. Anal. Prev. 2019, 125, 85–97. [Google Scholar] [CrossRef]
  58. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  59. Yin, L.; Ge, Y.; Xiao, K.; Wang, X.; Quan, X. Feature selection for high-dimensional imbalanced data. Neurocomputing 2013, 105, 3–11. [Google Scholar] [CrossRef]
  60. Schloegl, M. A multivariate analysis of environmental effects on road accident occurrence using a balanced bagging approach. Accid. Anal. Prev. 2020, 136, 105398. [Google Scholar] [CrossRef]
  61. Chen, T.; Wong, Y.D.; Shi, X.; Yang, Y. A data-driven feature learning approach based on Copula-Bayesian network and its application in comparative investigation on risky lane-changing and car-following maneuvers. Accid. Anal. Prev. 2021, 154, 106061. [Google Scholar] [CrossRef]
  62. Elvik, R. The more (sharp) curves, the lower the risk. Accid. Anal. Prev. 2019, 133, 105322. [Google Scholar] [CrossRef]
Figure 1. Section division by the homogeneous method; stratified k-fold cross-validation was used to compare the developed models, which was repeated 5 times.
Figure 1. Section division by the homogeneous method; stratified k-fold cross-validation was used to compare the developed models, which was repeated 5 times.
Ijerph 19 13693 g001
Figure 2. Statistics on the number and cumulative frequency of sections at differentiated risk levels; each 1 km section is counted 4 times according to the weather conditions (sunny, cloudy, overcast, and rainy).
Figure 2. Statistics on the number and cumulative frequency of sections at differentiated risk levels; each 1 km section is counted 4 times according to the weather conditions (sunny, cloudy, overcast, and rainy).
Ijerph 19 13693 g002
Figure 3. Three-category confusion matrix in this study.
Figure 3. Three-category confusion matrix in this study.
Ijerph 19 13693 g003
Figure 4. Pearson correlation matrix of continuous feature variables. If the correlation coefficient is greater than 0, it is a positive correlation, and if it is less than 0, it is a negative correlation. The stronger the correlation, the darker the color. The stronger the positive correlation, the closer to blue; the stronger the negative correlation, the closer to red.
Figure 4. Pearson correlation matrix of continuous feature variables. If the correlation coefficient is greater than 0, it is a positive correlation, and if it is less than 0, it is a negative correlation. The stronger the correlation, the darker the color. The stronger the positive correlation, the closer to blue; the stronger the negative correlation, the closer to red.
Ijerph 19 13693 g004
Figure 5. Predicted probability distribution of models trained by random forest and XGBoost. The test data are the same, and the training data are processed by different balancing algorithms: random forest uses SMOTEENN-balanced data and XGBoost uses RUS-balanced data.
Figure 5. Predicted probability distribution of models trained by random forest and XGBoost. The test data are the same, and the training data are processed by different balancing algorithms: random forest uses SMOTEENN-balanced data and XGBoost uses RUS-balanced data.
Ijerph 19 13693 g005
Figure 6. Feature importance (Gini importance, permutation importance, SHAP importance), sorted in accordance with the combined results of the three importance indicators.
Figure 6. Feature importance (Gini importance, permutation importance, SHAP importance), sorted in accordance with the combined results of the three importance indicators.
Ijerph 19 13693 g006
Figure 7. Feature variable partial dependence plot illustrating the effects of section risk by the interaction of RICU, GAE, and PCI; the missing part of the value is filled by interpolation.
Figure 7. Feature variable partial dependence plot illustrating the effects of section risk by the interaction of RICU, GAE, and PCI; the missing part of the value is filled by interpolation.
Ijerph 19 13693 g007
Table 1. Overview of the sampled units used in this study.
Table 1. Overview of the sampled units used in this study.
NumberStake NumberDriving DirectionWeather
1K1~K2forwardsunny
2K1~K2forwardcloudy
……
1679K209~K210reverseovercast
1680K209~K210reverserainy
Table 2. Overview of the feature variables used in this study.
Table 2. Overview of the feature variables used in this study.
FactorCodeVariableDescription
Horizontal alignmentHPNENumber of horizontal change pointsThe number of change points of horizontal alignment
CULRCurve length ratioRatio of curve length per kilometer (value between 0 and 1)
STLRStraight length ratioRatio of straight length per kilometer (value between 0 and 1)
RIMXMaximum curve radius indexMaximum value of the inverse of the radius. RIMX takes 0 in the straight sections
RICUCumulative radius indexCumulative value of the inverse of the radius
Vertical
alignment
VPNENumber of vertical change pointsThe number of change points of vertical alignment
GAEAverage gradeAverage grade per kilometer
GMXMaximum gradeMaximum grade per kilometer
GDFGrade differenceThe difference in the maximum grade per kilometer
DWLRDownhill length ratioThe proportion of downhill in the section
PavementPCIPavement surface condition indexThe value is between 0 and 100; the lower the value, the more serious the road damage.
RQIRiding quality indexThe value is between 0 and 100; the lower the value, the lower the smoothness of the road surface.
RDIRutting depth indexThe value is between 0 and 100; the lower the value, the deeper the road rut.
SRISkidding resistance indexThe value is between 0 and 100; the lower the value, the lower the anti-skid performance of the pavement.
Tunnels and expressway facilitiesTNNTunnel0—No tunnel in the section; 1—There are tunnels in the section.
ITCInterchange0—There is no interchange in the section; 1—There are interchanges in the section.
SVAService area0—There is no service area in the section; 1—There are service areas in the section.
WeatherWETWeather1—Sunny; 2—Cloudy; 3—Overcast; 4—Rainy.
Table 3. Overview of the 11 resampling algorithms used in this study.
Table 3. Overview of the 11 resampling algorithms used in this study.
MethodAlgorithmDescription
Under-sampling●Prototype generation (PG)Generate new samples based on original samples to achieve sample balance. Use K-means to cluster majority class samples and then use cluster centroids as newly generated replacement samples [31].
●Random under-sampling (RUS)Some samples are randomly removed from the majority class, so that the samples of each class are balanced.
●Edited nearest neighbor (ENN)Apply the nearest-neighbors algorithm to edit the dataset to remove samples with an insufficient neighborhood [32].
●All-KNN (ALLKNN)Apply ENN several times and vary the number of nearest neighbors [32].
Oversampling●Naive random over-sampling (ROS)Using the method of extraction with replacement, random sampling from minority class samples to replace the existing sample set; can increase the weight of minority class samples.
●Synthetic minority oversampling technique (SMOTE)For each minority class sample, the nearest k minority class samples are identified, a sample point is randomly selected each time, the corresponding adjacent sample point is randomly selected, and a new sample point is obtained by interpolating the sample point and adjacent sample point, thereby increasing the minority class samples to balance the data [33].
●Borderline-SMOTEThis is an improved algorithm of SMOTE. Divide the minority class sample points into “noise points”, “dangerous points”, and “safe points”, and only use the dangerous points when calculating the nearest k minority class samples [34].
●SMOTENCThis is an improved algorithm of SMOTE. Categorical variables are not properly distanced and interpolated. SMOTENC uses the value difference metric (VDM) algorithm to calculate the distance of categorical variables, which enables the processing of categorical variables [33].
●ADASYNSimilar to SMOTE, it is also based on k adjacent and interpolation algorithms, the difference being that ADASYN considers other types of samples when calculating k adjacent samples [35].
Mixed sampling●SMOTEENNThe SMOTE method is used to generate new minority class samples. There may be some noisy samples in the new samples. Apply the ENN method to remove noisy samples and obtain cleaner data [36].
●SMOT-Tomek LinksSimilar to SMOTEENN, Tomek Links are applied to remove noisy samples to obtain cleaner data [37].
Table 4. Number of sections with different risk levels in the different datasets.
Table 4. Number of sections with different risk levels in the different datasets.
DatasetBalancing
Algorithm
TotalL_1L_2L_3
Original dataset/1344109316883
Under-sampling●PG249838383
RUS249838383
ENN112410231883
●ALLKNN10659483483
Over-samplingROS3279109310931093
SMOTE3279109310931093
ADASYN3278109310991086
Borderline-SMOTE3279109310931093
SMOTENC3279109310931093
Mixed samplingSMOTEENN2394598798998
SMOT-Tomek Links3219106810661085
Table 5. Top 10 models ranked according to Score; when the average value of Score was similar, S.D. was used.
Table 5. Top 10 models ranked according to Score; when the average value of Score was similar, S.D. was used.
RankResampling AlgorithmClassifierScore
MeanS.D.
1SMOTEENNRandom forest0.510.04
2RUSRandom forest0.470.04
3RUSXGBoost0.460.04
4ROSRandom forest0.460.04
5SMOTENCRandom forest0.450..03
6ROSXGBoost0.450.04
7SMOTERandom forest0.440.03
9SMOT-Tomek LinksRandom forest0.440.03
8ADASYNRandom Forest0.440.04
10SMOTEENNXGBoost0.440.04
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, B.; Zhang, C.; Wong, Y.D.; Hou, L.; Zhang, M.; Xiang, Y. Comparing Resampling Algorithms and Classifiers for Modeling Traffic Risk Prediction. Int. J. Environ. Res. Public Health 2022, 19, 13693. https://doi.org/10.3390/ijerph192013693

AMA Style

Wang B, Zhang C, Wong YD, Hou L, Zhang M, Xiang Y. Comparing Resampling Algorithms and Classifiers for Modeling Traffic Risk Prediction. International Journal of Environmental Research and Public Health. 2022; 19(20):13693. https://doi.org/10.3390/ijerph192013693

Chicago/Turabian Style

Wang, Bo, Chi Zhang, Yiik Diew Wong, Lei Hou, Min Zhang, and Yujie Xiang. 2022. "Comparing Resampling Algorithms and Classifiers for Modeling Traffic Risk Prediction" International Journal of Environmental Research and Public Health 19, no. 20: 13693. https://doi.org/10.3390/ijerph192013693

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop