Next Article in Journal
Automated Mapping of Land Cover Type within International Heterogenous Landscapes Using Sentinel-2 Imagery with Ancillary Geospatial Data
Previous Article in Journal
Restoration of UAV-Based Backlit Images for Geological Mapping of a High-Steep Slope
Previous Article in Special Issue
Improved YOLOv7-Based Algorithm for Detecting Foreign Objects on the Roof of a Subway Vehicle
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Vision-Based On-Road Nighttime Vehicle Detection and Tracking Using Improved HOG Features

1
School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
2
Changzhou Xingyu Automotive Lighting System Co., Ltd., 182 Qinling Road, Changzhou 213000, China
3
School of Mechanical Engineering & Rail Transit, Changzhou University, Changzhou 213164, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(5), 1590; https://doi.org/10.3390/s24051590
Submission received: 9 January 2024 / Revised: 22 February 2024 / Accepted: 25 February 2024 / Published: 29 February 2024
(This article belongs to the Special Issue Object Detection and IOU Based on Sensors: Methods and Applications)

Abstract

:
The lack of discernible vehicle contour features in low-light conditions poses a formidable challenge for nighttime vehicle detection under hardware cost constraints. Addressing this issue, an enhanced histogram of oriented gradients (HOGs) approach is introduced to extract relevant vehicle features. Initially, vehicle lights are extracted using a combination of background illumination removal and a saliency model. Subsequently, these lights are integrated with a template-based approach to delineate regions containing potential vehicles. In the next step, the fusion of superpixel and HOG (S-HOG) features within these regions is performed, and the support vector machine (SVM) is employed for classification. A non-maximum suppression (NMS) method is applied to eliminate overlapping areas, incorporating the fusion of vertical histograms of symmetrical features of oriented gradients (V-HOGs). Finally, the Kalman filter is utilized for tracking candidate vehicles over time. Experimental results demonstrate a significant improvement in the accuracy of vehicle recognition in nighttime scenarios with the proposed method.

1. Introduction

As one of the basic directions of automatic driving technology, vehicle detection provides technical support for the follow-up anti-collision early warning system and has become the current research hotspot. Numerous studies on vehicle detection using monocular vision primarily concentrate on identifying vehicles during daylight hours [1,2]. Despite the traffic flow at night being much lower than during the daytime, insufficient illumination, complex road conditions, and the inapparent car contour lead to 42% of accidents occurring at nighttime [3]. The rear lights are the most significant feature of moving vehicles at night [4]. In contrast, features such as symmetry, color, shadows, and edges, usually used in daytime, are not prominent [5,6,7].
Currently, some research on vehicle detection at nighttime uses a template-based method [8,9], which mainly depends on the shape similarity and coordinate orientation features of vehicle light. Nevertheless, establishing appropriate thresholds for similarity and coordinates poses a challenge, leading to potential false positives or missed detections and consequently compromising the system’s robustness. Some scholars also proposed machine learning-based methods, mainly using features such as Haar, histogram of oriented gradient (HOG), or deformable parts model (DPM) [10,11] for classification. Song [12] used the Gaussian mixture model of spatial relations to detect daytime and nighttime vehicles in complex urban scenes. Chien et al. [13] proposed an improved multi-scale retinal image enhancement algorithm, which utilizes a two-level classifier based on a bag of features (BoF) and a convolutional neural network (CNN) to reduce false positives. Kuang et al. [14] proposed using a bionic night image enhancement method and support vector machines (SVM) classifier based on the fractional feature fusion method to achieve vehicle detection. Satzoda and Trivedi [10] used the adaptive boosting (AdaBoost) classifier based on Haar features to generate the boundary box of the vehicle for the input gray image; they used the calibration, color, and other features to verify the vehicle area producing a dataset for night vehicle detection. The vehicle feature information is always incomplete, influenced by insufficient illumination and light reflection at night. Dai et al. [15] proposed a method for night vehicle detection using grayscale division and Hough transform strategy and obtained a high detection rate for night vehicle detection and low false alarms.
Machine learning algorithms exhibit good generalization to unseen data. However, most machine learning methods only focus on the lamp information, which causes substantial difficulties for vehicle detection when the vehicle lamp is similar to other light sources on the road. Then, the deep learning methods abounded, for example, the one-stage object detection models (you only look once, YOLO, and single shot multi-box detector, SSD) [16,17] and two-stage object detection models (region CNN (R-CNN), fast-RCNN, and faster R-CNN) [18,19,20]. Gao et al. [21] used vision and lidar data fusion to detect vehicles. Mo et al. [22] proposed a neural network iteratively trained with pseudo labels based on the highlighted area of cars, using a hierarchical training strategy to solve the problem of significant intra-class differences and small inter-class differences between vehicles. Zhang and Zhu [23] proposed an improved multi-source vehicle detection algorithm based on YOLOv3, which can detect vehicles in low-resolution and low-contrast aerial infrared images. The one-stage models performed well in detection speed, and the two-stage models showed good accuracy in vehicle detection [24]. However, studies are also needed to explore the advantages of machine learning through low-cost labeling of images, computational resources, and interpretability requirements. There are scenarios where traditional machine learning methods might have advantages over deep learning in image recognition. (1) Data efficiency: machine learning methods can perform well with smaller datasets. Deep learning often requires large amounts of labeled data to generalize effectively, and it may struggle when labeled examples are limited. (2) Interpretability: machine learning models are generally more interpretable than deep learning models. Understanding the decision-making process of a machine learning algorithm can be critical, especially in sensitive applications or industries. (3) Computational resources: deep learning models and intense neural networks typically require significant computational resources (such as powerful GPUs or TPUs) for training. In contrast, some machine learning algorithms can be trained on less powerful hardware. (4) Feature engineering: machine learning often requires manual feature engineering, where domain knowledge is used to extract relevant features from the data. In certain situations, carefully crafted features may outperform automatically learned features in deep learning. (5) Computational speed: for some tasks, machine learning models might provide faster inference times than deep learning models, especially in critical real-time processing situations. (6) Resource constraints: machine learning models might be more practical than deploying resource-intensive deep learning models in applications with limited computational resources, such as edge devices or IoT devices. Therefore, deep learning methods suffer the properties of data-driven and a great deal of data labeling. In practical vehicle identification problems, it is difficult to ensure a substantial and wide variety of objects [25].
Regarding the hardware cost problems mentioned above, a vehicle nighttime detection method based on the saliency model and improved saliency histogram of oriented gradient (S-HOG) features is proposed in this paper. The flow chart is shown in Figure 1. One can mainly be divided into two steps: region of interest (ROI) generation and vehicle verification. (1) The removal of background illumination (saliency segmentation-based and template-based methods) was applied in the object proposal method to create the ROI. (2) The region-based superpixel and HOG features of grey and red images were combined for SVM classification. Then, the non-maximum suppression method fusion of S-HOG symmetric features is used to eliminate the overlapped areas and improve vehicle detection accuracy. Finally, the Kalman filter is utilized to track the candidate vehicle.

2. Proposed Method

2.1. Object Proposal Generation Method

2.1.1. Background Light Removing Based on Saliency Segmentation

In a complex illumination environment, significantly interfered with by vehicle reflection and streetlamp illumination, the performance of global vehicle binarization is poor. The traditional Otsu method cannot separate small bright spots from an image. In the literature [26], a fixed threshold method is proposed to obtain a fixed threshold for segmentation by the brightness characteristics of lights. As shown in Figure 2c, although this method performs better than the traditional Otsu method illustrated in Figure 2b, it still lacks robustness.
To solve the problems above, we proposed a background light removal method based on saliency detection (Figure 3). A good saliency detection model should satisfy at least three criteria: (1) good detection: the probability of missing real salient regions and falsely marking the background as a salient region should be low, (2) high resolution: saliency maps should have a high or full resolution to locate salient objects and retain original image information accurately, and (3) computational efficiency: as front-ends to other complex processes, these models should detect salient regions quickly.
There are two main classifications of saliency model detection: patch-based and region-based. Patch-based detection models include fast and efficient saliency (FES), self-resemblance (SER), saliency segmentation (SEG), and spatially weighted dissimilarity (SWD). These methods suffer from two drawbacks: high-contrast edges usually stand out rather than salient objects, and the boundaries of salient objects are not correctly preserved. To overcome these problems, some methods propose computing saliency based on cues. Two main advantages: the number of regions is much less than the number of blocks, which implies the potential to develop efficient and fast algorithms; more informative features can be extracted from regions, leading to better performance. Region-based saliency models include covariances (COV), dense and sparse reconstruction (DSR), Markov chain (MC), and robust background detection (RBD). The main advantages of region-based saliency models: (1) complementary priors are employed to improve overall performance, which is the main advantage; (2) regions provide more complex cues (such as color histograms) compared to pixels and color patches to better capture the salient objects of the scene; (3) since the number of regions in the image is much smaller than the number of pixels, the computational saliency at the region level can significantly reduce the computational cost when generating full-resolution saliency maps. Additionally, there are some other state-of-the-art models based on context, estimation, and graph, such as context-aware (CA), graph-regularized (GR), and spectral residual (SR).
Compare Figure 2a with Figure 2d; it shows that the method we proposed removes the complex illumination. Figure 2e is a traditional Otsu method based on Figure 2d, and Figure 2f is a proposed method based on Figure 2d; it is evident that the proposed method performs better than other methods.

2.1.2. Segmentation Method Based on Template

Although the most recent research uses a template-based method, due to the angle and overexposure, using the similarities of light shapes and coordinate features can easily cause false detection. The two lights of one vehicle are often horizontally aligned on a captured image, so we only use light coordinate features weighted by a saliency map to generate ROI.
The ROI generate stage uses the following characteristics of lights:
The lateral distance D l between two lights on the same vehicle is shorter than the maximum of the vehicle T h l :
D l L 1 , L 2 < T h l
Two lights on the same vehicle are basically at the same horizontal position:
D h L 1 , L 2 D l L 1 , L 2 < T h h l
Although only using coordinate features of lights could cause false detection, it can avoid missing detection and facilitate subsequent verification phase operations.

2.2. Hypotheses Verification

Saliency areas generated by coordinate features may contain some non-vehicle regions requiring further verification. Due to the complex nighttime environment, the traditional method using the similarity of light shapes will lead to false detection [27]. Since the HOG feature has the advantage of robustness to lighting conditions, invariance to geometric transformations, effectiveness in cluttered backgrounds, dimensionality reduction, compatibility with machine learning algorithms, high detection accuracy, and adaptability to real-time applications [28], it is often used in vehicle detection. However, due to the time-consuming and low noise immunity of the HOG feature and the lack of vehicle contour pieces of information at night, the effect of using contour features for classification does not perform as well as in daytime [22]. We propose S-HOG features combining superpixel and HOG by grey and red images, which reflect the vehicle contour and lamp contour, respectively, and then use SVM for classification to improve the accuracy of vehicle recognition.

2.2.1. S-HOG Feature Extraction

A region-based superpixel method was fused in the HOG feature. Simple linear iterative clustering zero (SLIC0) superpixel was used in the pre-processing method to weigh each pixel in HOG feature generation without compactness factor as input and calculated automatically. The fusion approach aims to improve spatial coherence, handle variations in vehicle orientation, provide adaptive feature extraction, reduce noise, enhance discriminative power, utilize contextual information, and create a more comprehensive and robust feature representation for accurate vehicle recognition in diverse and challenging conditions (Figure 4a). The good performance of S-HOG relies on its ability to exploit the synergies between superpixels and HOG, leading to improved performance over using either technique in isolation. The sketch map is shown in Figure 4. The automatic value adapts to the content of the superpixel, leading to SLIC0 being better suited for texture and non-texture regions, the advantages of which come at the cost of slightly poor boundary adherence to regions. There is a small computational overhead and memory overhead to original simple linear iterative clustering (SLIC), which outweigh the small disadvantages for most applications of superpixels.
This paragraph adopts a method that combines S-HOG features of grey and red images. As mentioned in [29], the middle of rear lights often highlights white surrounding the red area. Firstly, change the image from RGB to HSV, the grey image is obtained to the V channel, and the red image is obtained in Table 1, and HOG features such as H G r a y and H R e d are extracted, respectively.
The grey and red channel images are normalized to 32 × 32 pixels, 8 × 8 pixels are selected as one HOG cell, and each block contains four cells. Since the edge gradients differ by 180° can be regarded as the same direction, the gradient direction of pixels is evenly divided into nine histogram channels, and the HOG feature vector of each training sample is 324. The 324-dimensional feature of the grey image is concatenated with the 324-dimensional feature of the red image to obtain a 648-dimensional vector for vehicle verification.

2.2.2. SVM Training

The HOG feature with SVM classification is widely used in vehicle and pedestrian detection [7,30]. This paper uses the radial basis function (RBF) to construct the optimal hyper-plane. Samples of 5705 positive and 3669 negative are selected for training to obtain the classifier. The positive samples include cars, vans, buses, and trucks, and the negative samples include traffic signs, road guardrails, streetlights, and billboards. As Figure 5b shows, this method performs better than the traditional HOG feature but also is required to remove overlapping.

2.2.3. Overlapped Area Removal

The principle of non-maximum suppression is to select the most confident bounding boxes while suppressing overlapping or redundant detections. NMS is chosen for its ability to improve precision, recall, and localization accuracy in object detection tasks, making it a crucial component in many state-of-the-art detection systems (such as vehicle detection) [31]. Non-maximum suppression (NMS) uses the highest score obtained by the classifier to suppress low-scoring windows and eliminate redundant detection areas. Due to the symmetry of vehicles, it is proposed to use the V-HOG feature [30] to calculate the score. Figure 6 shows that the V-HOG feature divides the image into four vertical parts; each part is a HOG cell with nine bins.
As Figure 7 shows, since the vehicle is more symmetrical than in other areas, so if there is a vehicle in the ROI area, cell1 and cell2 are symmetrical to cell3s and cell4s, and the definition of these four cells is in Equations (3)–(6):
c e l l 1 = b i n 11   b i n 12   b i n 13   b i n 14   b i n 15   b i n 16   b i n 17   b i n 18   b i n 19
c e l l 2 = b i n 21   b i n 22   b i n 23   b i n 24   b i n 25   b i n 26   b i n 27   b i n 28   b i n 29
c e l l 3 s = b i n 39   b i n 38   b i n 37   b i n 36   b i n 35   b i n 34   b i n 33   b i n 32   b i n 31
c e l l 4 s = b i n 49   b i n 48   b i n 47   b i n 46   b i n 45   b i n 44   b i n 43   b i n 42   b i n 41
The similarity value calculation based on Euclidean distance is a method with a high usage rate. As Equations (7) and (8), the similarity value s i n x , y is between [0,1]. The x i means bins in cell 1 and cell 2, y i means bins in cell3s and cell4s, d x , y means the distance between cells. The higher the value of the s i n x , y , the higher the similarity.
d x , y = Σ x i y i 2
s i n x , y = 1 1 + d x , y
The similarity value is used as the score of NMS, and the overlap degree is set to 0.3 to remove the overlap area, as shown in Figure 7c.

2.2.4. Vehicle Tracking

The vehicle movement is continuous, so the position and size changes of the same car in sequence images are also continuous. Prediction and tracking of vehicle targets can improve the accuracy of vehicle detection.
In the video, the displacement of the same vehicle between two consecutive frames is relatively minor, and the vehicle’s motion approximates uniform linear movement. In variable-speed linear motion, turning, parking, and similar conditions, the small-time interval between frames mitigates the approximation error associated with uniform rectilinear motion, treating these errors as system noise. Furthermore, a corresponding relationship exists between the position and size of the vehicle in the image. Therefore, it can be assumed that the size change of the vehicle in sequence images is approximately uniform, and the approximate error can be considered system noise. The system state equation can be expressed as follows:
X k = A X k 1 + B u k + w k
X = [ x , y , w , h , Δ x , Δ y ] T
A = 1   0   0   0   1   0 0   1   0   0   0   1 0   0   1   0   0   0 0   0   0   1   0   0 0   0   0   0   1   0 0   0   0   0   0   1
B = [ 0,0 , 0,0 , 0,0 ] T
X is the system state vector,   ( x ,   y ) is the image coordinate in the midpoint of the vehicle lower boundary, w and h are the width and height of the target in the image coordinate system. A is the system status transfer matrix. The system measurement equation is:
Z k = H X k + v k
The system measurement vector is:
H = 1   0   0   0   0   0 0   1   0   0   0   0 0   0   1   0   0   0 0   0   0   1   0   0 0   0   0   0   1   0 0   0   0   0   0   1
The Kalman filter can handle dynamic systems, robustness to noisy measurements, predictive-corrective nature, adaptability to changes in motion, continuous state estimation, efficient memory usage, integration with multiple sensors, predictive capability, and its well-established track record in tracking and estimation. In this article, the Kalman filter predicts the new position, width, and height and generates the expected vehicle area [32]. The prediction frame at time t matches the image recognition frame at time t + 1 . The maximum intersection over union (IOU) is the observed value at time t + 1 . If the matching fails, the final predicted value will be supposed to be the observed value. If the matching is a failure for 10 consecutive times, the tracking object will be discarded.

3. Results and Discussion

3.1. Comparison of Different Saliency Methods

The performance of the SWD method based on patches was better than that of others for removing background illumination at night. All the streetlights were judged to be the non-saliency section, as shown in Figure 8 (column SWD). Other methods, for example, FES, SER, CA, DSR, MC, and RBD, included the environment light, e.g., streetlights and building lights. COV method also avoided the non-relevant lights, as shown in Figure 8 (column COV); however, the integrity of prospective vehicles is lower, e.g., the road near the target vehicles was identified as foreground.
The performance of saliency detection near scenes was better than the far scenes. The targets cars and non-targets are easier to confuse together in far scenes, for example, in DSR, MC, RBD, FES, and SER of the first line in Figure 8.

3.2. Comparison of Different Feature Extraction Methods

To verify the effectiveness of the proposed method, MATLAB R2016a is used to conduct vehicle recognition experiments in the pictures collected under real-time urban and elevated roads. The processor is Intel i5-8250, 8 G memory, with 500 and 720 × 480 image resolution.
The experimental results are presented in Figure 8. Figure 9a indicates that under different road scenes, there will be misdetections when using the template method [29,33], such as color, coordinate, and shape characteristics. Figure 9b shows that some vehicles with less obvious contours will fail to be detected when using the traditional HOG feature. Figure 9c shows that the detection accuracy improved using the proposed method, especially when the vehicle contour is not apparent.
Some images of urban and highways were obtained in the night scene, and the results are shown in Table 2, using the method mentioned above. The accuracy rate and the recall rate of two key objectives are mainly based on three parameters, which are true positive (TP), false positive (FP), and false negative (FN). They are defined as Equations (15) and (16). The higher the rates, the better the results.
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
Table 2 shows the test results of the proposed and traditional methods in urban and highway scenarios. It shows that the method proposed in this paper performs better than conventional methods.

3.3. Deep Learning Algorithm Detection and Comparison

YOLO series object detection algorithms are object detection algorithms based on regression networks. These algorithms have obvious advantages in performance, speed, lightweight design, and flexibility. Moreover, the YOLO algorithms have been developed to derive various versions, and the development is relatively complete. They have become some of the most concerned algorithms in target detection [34,35].
The training dataset in network detection experiments includes 7800 vehicle images and 1260 headlight images with a pixel resolution of 720 × 480 from a real-time tachograph. The classification of test images and test results are shown in Table 3.
All the state-of-the-art deep learning algorithms have excellent performance in mAP, precision, and recall (above 96%). However, these deep learning networks have more train time, model size, and labeling image size than traditional machine learning methods. For example, the training time of deep learning is 19.82 times than feature engineering methods and 78.19 times in model size. The number of annotated images required is only 5% of machine learning methods than deep learning methods. The YOLO identification experiment exhibited robust detection capabilities primarily tailored to models within the dataset and similar series, indicating limited universality. Specifically, in vehicle-related projects, such as headlight recognition in engineering, a substantial amount of diverse data is essential for comprehensive training [3]. Consequently, our proposed method often outperforms in scenarios with limited data volumes. The algorithm presented in this study demonstrates a strong generalization ability, significantly enhancing classification accuracy and the overall adaptability of the final robust classifier. Moreover, it is noteworthy that deep learning models discussed in this chapter impose hardware requirements and demand extensive training periods. In contrast, the proposed model offers advantages in terms of low training time costs and significantly smaller model parameters than traditional deep learning algorithms. This makes our model well-suited for deployment in embedded devices and other hardware environments commonly encountered in general settings [36].

4. Conclusions

This paper proposes a nighttime vehicle detection method with hardware cost constraints, using a saliency map and superpixel weighting the grey and red channel HOG features for SVM classification and then using the NMS method to fuse S-HOG. The main advantage of the proposed method is that our method performs better in detecting vehicles with less obvious contours, especially when we use the color and S-HOG features. Additionally, the result demonstrates that this method performed better than the traditional HOG methods. Lastly, the Kalman filter was used to improve the prediction accuracy. Experimental results have shown that the proposed method performs better than conventional methods.

Author Contributions

L.Z.: conceptualization, data curation, validation, formal analysis, methodology, writing—original draft. W.X.: conceptualization, software, formal analysis, methodology, funding acquisition, writing—original draft. C.S.: software, visualization, writing—review and editing. Y.H.: conceptualization, data curation, formal analysis, supervision, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National High Technology Research and Development Program of China, grant number 2016YFD0701003, and the Postgraduate Research & Practice Innovation Program of Jiangsu Province, grant number SJCX23_1488, SJCX23_1499.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The evaluation code and dataset are available at https://github.com/weiyuexu/Rear-Lamp-Vehicle-Detection (accessed on 1 January 2024).

Conflicts of Interest

Author Li Zhang was employed by the company Changzhou Xingyu Automotive Lighting System Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Shabestari, Z.B.; Hosseininaveh, A.; Remondino, F. Motorcycle Detection and Collision Warning Using Monocular Images from a Vehicle. Remote Sens. 2023, 15, 5548. [Google Scholar] [CrossRef]
  2. Mu, K.; Hui, F.; Zhao, X.; Prehofer, C. Multiscale edge fusion for vehicle detection based on difference of Gaussian. Optik 2016, 11, 127. [Google Scholar] [CrossRef]
  3. Toney, G.; Bhargava, C. Adaptive headlamps in automobile: A review on the models, detection techniques, and mathematical models. IEEE Access 2021, 9, 87462–87474. [Google Scholar] [CrossRef]
  4. Yang, Z.; Pun-Cheng, L.S. Vehicle detection in intelligent transportation systems and its applications under varying environments: A review. Image Vision Comput. 2018, 69, 143–154. [Google Scholar] [CrossRef]
  5. Ibarra-Arenado, M.; Tjahjadi, T.; Perez-Oria, J.; Robla-Gomez, S.; Jimenez-Avello, A. Shadow-based vehicle detection in urban traffic. Sensors 2017, 17, 975. [Google Scholar] [CrossRef] [PubMed]
  6. Wang, X.; Tang, J.; Niu, J.; Zhao, X.J.N. Vision-based two-step brake detection method for vehicle collision avoidance. Neurocomputing 2016, 173, 450–461. [Google Scholar] [CrossRef]
  7. Huang, D.Y.; Chen, C.H.; Chen, T.Y.; Hu, W.C.; Feng, K.W. Vehicle detection and inter-vehicle distance estimation using single-lens video camera on urban/suburb roads. J. Vis. Commun. Image Represent. 2017, 46, 250–259. [Google Scholar] [CrossRef]
  8. Pradeep, C.S.; Ramanathan, R. An improved technique for night-time vehicle detection. In Proceedings of the 2018 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Karnataka, India, 19–22 September 2018; pp. 508–513. [Google Scholar] [CrossRef]
  9. Gao, F.; Ge, Y.; Lu, S.; Zhang, Y. On-line vehicle detection at nighttime-based tail-light pairing with saliency detection in the multi-lane intersection. IET Intell. Transp. Syst. 2019, 13, 515–522. [Google Scholar] [CrossRef]
  10. Satzoda, R.K.; Trivedi, M.M. Looking at vehicles in the night: Detection and dynamics of rear lights. IEEE Trans. Intell. Transp. Syst. 2019, 20, 4297–4307. [Google Scholar] [CrossRef]
  11. Tang, Y.; Zhang, C.; Gu, R.; Li, P.; Yang, B. Vehicle Detection and recognition for intelligent traffic surveillance system. Multimed. Tools. Appl. 2017, 76, 5817–5832. [Google Scholar] [CrossRef]
  12. Song, J.f. Vehicle detection using spatial relationship gmm for complex urban surveillance in daytime and nighttime. Int. J. Parallel Program. 2018, 46, 859–872. [Google Scholar] [CrossRef]
  13. Chien, J.C.; Chen, Y.S.; Lee, J.D. Improving night time driving safety using vision-based classification techniques. Sensors 2017, 17, 2199. [Google Scholar] [CrossRef] [PubMed]
  14. Kuang, H.; Zhang, X.; Li, Y.J.; Chan, L.L.H.; Yan, H. Nighttime vehicle detection based on bio-inspired image enhancement and weighted score-level feature fusion. IEEE Trans. Intell. Transp. Syst. 2017, 18, 927–936. [Google Scholar] [CrossRef]
  15. Dai, X.; Liu, D.; Yang, L.; Liu, Y. Research on headlight technology of night vehicle intelligent detection based on hough transform. In Proceedings of the 2019 International Conference on Intelligent Transportation, Big Data & Smart City (ICITBS), Changsha, China, 12–13 January 2019; pp. 49–52. [Google Scholar] [CrossRef]
  16. Tang, J.; Li, S.; Liu, P.J.P.R. A review of lane detection methods based on deep learning. Pattern Recognit. 2021, 111, 107623. [Google Scholar] [CrossRef]
  17. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 10–16 October 2016; pp. 21–37. [Google Scholar] [CrossRef]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  19. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  20. Gao, H.; Cheng, B.; Wang, J.; Li, K.; Zhao, J.; Li, D. Object Classification Using CNN-Based Fusion of Vision and LIDAR in Autonomous Vehicle Environment. IEEE Trans. Industr. Inform. 2018, 14, 4224–4231. [Google Scholar] [CrossRef]
  21. Mo, Y.; Han, G.; Zhang, H.; Xu, X.; Qu, W.J.N. Highlight-assisted nighttime vehicle detection using a multi-level fusion network and label hierarchy. Neurocomputing 2019, 355, 13–23. [Google Scholar] [CrossRef]
  22. Zhang, X.X.; Zhu, X. Moving vehicle detection in aerial infrared image sequences via fast image registration and improved YOLOv3 network. Int. J. Remote Sens 2020, 41, 4312–4335. [Google Scholar] [CrossRef]
  23. Wang, H.; Yu, Y.; Cai, Y.; Chen, X.; Chen, L.; Liu, Q. A Comparative Study of State-of-the-Art Deep Learning Algorithms for Vehicle Detection. IEEE Intell. Transp. Syst. Mag. 2019, 11, 82–95. [Google Scholar] [CrossRef]
  24. Jeong, K.M.; Song, B.C. Night time vehicle detection using rear-lamp intensity. In Proceedings of the 2016 IEEE International Conference on Consumer Electronics-Asia (ICCE-Asia), Seoul, Republic of Korea, 26–28 October 2016; pp. 1–3. [Google Scholar] [CrossRef]
  25. Gautam, S.; Kumar, A.J.M.T. Applications. Image-based automatic traffic lights detection system for autonomous cars: A review. Multimed. Tools. Appl. 2023, 82, 26135–26182. [Google Scholar] [CrossRef]
  26. Ding, Z.; Mo, W. Vehicle Type and Sub-type Recognition by Lamp Pairs Distance and Lamp Contour in ITS. In Proceedings of the 2020 IEEE 8th International Conference on Computer Science and Network Technology (ICCSNT), Dalian, China, 20–22 November 2020; pp. 134–138. [Google Scholar] [CrossRef]
  27. Arunmozhi, A.; Park, J. Comparison of HOG, LBP and Haar-Like Features for On-Road Vehicle Detection. In Proceedings of the 2018 IEEE International Conference on Electro/Information Technology (EIT), Rochester, NY, USA, 3–5 May 2018; pp. 0362–0367. [Google Scholar] [CrossRef]
  28. Chien, C.L.; Hang, H.M.; Tseng, D.C.; Chen, Y.S. An image based overexposed taillight detection method for frontal vehicle detection in night vision. In Proceedings of the 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), Jeju, Republic of Korea, 13–16 December 2016; pp. 1–9. [Google Scholar] [CrossRef]
  29. Laopracha, N. The Combination of Different Cell Sizes of HOG with KELM for Vehicle Detection. In Proceedings of the Recent Advances in Information and Communication Technology 2019: Proceedings of the 15th International Conference on Computing and Information Technology (IC2IT 2019), Bangkok, Thailand, 4–5 July 2019; pp. 180–190. [Google Scholar] [CrossRef]
  30. Yin, G.; Yu, M.; Wang, M.; Hu, Y.; Zhang, Y. Research on highway vehicle detection based on faster R-CNN and domain adaptation. Appl. Intell. 2022, 52, 3483–3498. [Google Scholar] [CrossRef]
  31. Ju, Z.; Zhang, H.; Li, X.; Chen, X.; Han, J.; Yang, M. A Survey on Attack Detection and Resilience for Connected and Automated Vehicles: From Vehicle Dynamics and Control Perspective. IEEE TIV. 2022, 7, 815–837. [Google Scholar] [CrossRef]
  32. Chen, X.; Wang, Z.; Hua, Q.; Shang, W.L.; Luo, Q.; Yu, K. AI-Empowered Speed Extraction via Port-Like Videos for Vehicular Trajectory Analysis. IEEE Trans. Intell. Transp. Syst. 2023, 24, 4541–4552. [Google Scholar] [CrossRef]
  33. Wu, J.; Shen, T.; Wang, Q.; Tao, Z.; Zeng, K.; Song, J. Local Adaptive Illumination-Driven Input-Level Fusion for Infrared and Visible Object Detection. Remote Sens. 2023, 15, 660. [Google Scholar] [CrossRef]
  34. Lee, D.H.; Liu, J.L. End-to-end deep learning of lane detection and path prediction for real-time autonomous driving. Signal Image Video Process. 2022, 17, 199–205. [Google Scholar] [CrossRef]
  35. Nadeem, H.; Javed, K.; Nadeem, Z.; Khan, M.J.; Rubab, S.; Yon, D.K.; Naqvi, R.A.J.S. Road feature detection for advance driver assistance system using deep learning. Sensors 2023, 23, 4466. [Google Scholar] [CrossRef] [PubMed]
  36. Li, S.; Zhao, L. A low-cost and fast vehicle detection algorithm with a monocular camera for adaptive driving beam systems. IEEE Access 2021, 9, 26147–26155. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed method for vehicle detection at night.
Figure 1. Flowchart of the proposed method for vehicle detection at night.
Sensors 24 01590 g001
Figure 2. Highlight extract: (a) gray image; (b) traditional Otsu; (c) fixed threshold method; (d) background illumination removal; (e) traditional Otsu based on saliency model; and (f) fixed threshold based on saliency model.
Figure 2. Highlight extract: (a) gray image; (b) traditional Otsu; (c) fixed threshold method; (d) background illumination removal; (e) traditional Otsu based on saliency model; and (f) fixed threshold based on saliency model.
Sensors 24 01590 g002
Figure 3. Saliency model generation.
Figure 3. Saliency model generation.
Sensors 24 01590 g003
Figure 4. Superpixel and HOG aggregation: (a) aggregation flowchart and (b) superpixel weighted HOG feature generation.
Figure 4. Superpixel and HOG aggregation: (a) aggregation flowchart and (b) superpixel weighted HOG feature generation.
Sensors 24 01590 g004
Figure 5. HOG features: (a) cells and blocks and (b) bins.
Figure 5. HOG features: (a) cells and blocks and (b) bins.
Sensors 24 01590 g005
Figure 6. V-HOG features: (a) cells and (b) bins.
Figure 6. V-HOG features: (a) cells and (b) bins.
Sensors 24 01590 g006
Figure 7. V-HOG symmetry: (a) ROI image; (b) cell1 and cell4s HOG symmetry; and (c) cell2 and cell3s HOG symmetry.
Figure 7. V-HOG symmetry: (a) ROI image; (b) cell1 and cell4s HOG symmetry; and (c) cell2 and cell3s HOG symmetry.
Sensors 24 01590 g007
Figure 8. Visual comparison of 12 ROI methods: (a) methods based on patches; (b) methods based on region; and (c) methods based on others.
Figure 8. Visual comparison of 12 ROI methods: (a) methods based on patches; (b) methods based on region; and (c) methods based on others.
Sensors 24 01590 g008
Figure 9. (a) Template method; (b) traditional HOG and SVM; (c) improved HOG and SVM. Note: the green boxes mean the detection results.
Figure 9. (a) Template method; (b) traditional HOG and SVM; (c) improved HOG and SVM. Note: the green boxes mean the detection results.
Sensors 24 01590 g009
Table 1. Red color threshold parameters.
Table 1. Red color threshold parameters.
HueSaturationValue
0–30 and 150–18010–80190–255
Table 2. The test result of improved and traditional methods in urban and highway scenarios.
Table 2. The test result of improved and traditional methods in urban and highway scenarios.
ScenarioMethodTPFPFNPrecision (%)Recall (%)Train Time (h)Model Size (MB)Image Label Size
UrbanTemplate176546921179.0189.320.500.11500
HOG167628430085.5184.820.510.15
Proposed185414712292.6593.830.530.16
HighwayTemplate472422891.8394.400.490.11500
HOG465243595.0993.000.520.13
Proposed482281894.5196.400.520.14
Table 3. YOLO algorithms detect the results of vehicles.
Table 3. YOLO algorithms detect the results of vehicles.
ModelmAP (%)Precision (%)Recall (%)Param (M)GFLOPsFPSTrain Time (h)Model Size (MB)Image Label Size
YOLO v598.097.896.36.715.884.712.58214.410,000
YOLO v699.599.499.04.011.887.710.9508.7
YOLO v799.899.099.05.713.189.321.82912.3
YOLO v899.598.998.73.08.1303.037.2026.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, L.; Xu, W.; Shen, C.; Huang, Y. Vision-Based On-Road Nighttime Vehicle Detection and Tracking Using Improved HOG Features. Sensors 2024, 24, 1590. https://doi.org/10.3390/s24051590

AMA Style

Zhang L, Xu W, Shen C, Huang Y. Vision-Based On-Road Nighttime Vehicle Detection and Tracking Using Improved HOG Features. Sensors. 2024; 24(5):1590. https://doi.org/10.3390/s24051590

Chicago/Turabian Style

Zhang, Li, Weiyue Xu, Cong Shen, and Yingping Huang. 2024. "Vision-Based On-Road Nighttime Vehicle Detection and Tracking Using Improved HOG Features" Sensors 24, no. 5: 1590. https://doi.org/10.3390/s24051590

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop