Next Article in Journal
EffResUNet: Encoder Decoder Architecture for Cloud-Type Segmentation
Next Article in Special Issue
A Gradient Boosted Decision Tree-Based Influencer Prediction in Social Network Analysis
Previous Article in Journal / Special Issue
Trust-Based Data Communication in Wireless Body Area Network for Healthcare Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Yolov5 Series Algorithm for Road Marking Sign Identification

by
Christine Dewi
1,*,
Rung-Ching Chen
2,*,
Yong-Cun Zhuang
2 and
Henoch Juli Christanto
3
1
Department of Information Technology, Satya Wacana Christian University, Salatiga 50711, Indonesia
2
Department of Information Management, Chaoyang University of Technology, Taichung 41349, Taiwan
3
Department of Information System, Atma Jaya Catholic University of Indonesia, Jakarta 12930, Indonesia
*
Authors to whom correspondence should be addressed.
Big Data Cogn. Comput. 2022, 6(4), 149; https://doi.org/10.3390/bdcc6040149
Submission received: 23 October 2022 / Revised: 3 December 2022 / Accepted: 5 December 2022 / Published: 7 December 2022
(This article belongs to the Special Issue Computational Collective Intelligence with Big Data–AI Society)

Abstract

:
Road markings and signs provide vehicles and pedestrians with essential information that assists them to follow the traffic regulations. Road surface markings include pedestrian crossings, directional arrows, zebra crossings, speed limit signs, other similar signs and text, and so on, which are usually painted directly onto the road surface. Road markings fulfill a variety of important functions, such as alerting drivers to the potentially hazardous road section, directing traffic, prohibiting certain actions, and slowing down. This research paper provides a summary of the Yolov5 algorithm series for road marking sign identification, which includes Yolov5s, Yolov5m, Yolov5n, Yolov5l, and Yolov5x. This study explores a wide range of contemporary object detectors, such as the ones that are used to determine the location of road marking signs. Performance metrics monitor important data, including the quantity of BFLOPS, the mean average precision (mAP), and the detection time (IoU). Our findings shows that Yolov5m is the most stable method compared to other methods with 76% precision, 86% recall, and 83% mAP during the training stage. Moreover, Yolov5m and Yolov5l achieve the highest score, mAP 87% on average in the testing stage. In addition, we have created a new dataset for road marking signs in Taiwan, called TRMSD.

1. Introduction

Traffic signs are used to control traffic, alert drivers to potential danger ahead providing useful information to make driving safer and easier [1]. To provide unambiguous information, traffic signs are designed using several fundamentally distinct design styles. In addition, the background with many different buildings and shop signs makes it hard for the system to identify the street signs automatically. Thus, it becomes difficult to locate the road signs in many environments [2]. Road markings alert drivers to dangerous areas on the road, indicate traffic directions, suggest they slow down, and perform other useful tasks. There are various types of landmarks that are used for vehicle localization purposes, among which one of the most common types of landmarks is road surface markings.
Automatic detection of road markings has been used for autonomous vehicle guidance [3] or for lane detection systems [4], which is a common active safety feature of commercial cars. In recent years, various strategies for identifying road signs have been developed [5,6]. Most of the strategies deal with single images with live backgrounds [7]. Ref. [8] described a method for detecting road signs using a color index system. The first step in the process of designing road signs involved depicting each model in terms of a color histogram. An investigation on four distinct varieties of road markings was carried out by Qin and colleagues [9] using a machine vision approach. The images of the marking outlines were then retrieved from the photographs after being created at random using image processing techniques.
Due to the high computational capability, various deep learning methods were recently proposed for recognizing road markers. Following the calculation of the geometric characteristics and pixel distribution of road markings by Soilan et al. [10], a two-layer feedforward network was used to extract pedestrian crossings and five different arrow classes. In the study by Wen et al. [11], a pixel-level U-net segmentation network was created to extract road markings. Chen et al. [12] introduced a deep learning model based on a dense feature pyramid network, which, to extract the path markers, was trained from start to finish. Cheng et al. [13] proposed using road markings as an intensity threshold technique based on normalization of unsupervised intensity and a deep learning strategy [14].
Current state-of-the-art detection frameworks, such as single shot detector (SSD) [15] and You Only Look Once (Yolo) [16], can be inferred in real time while maintaining robustness for the task of road marking detection. Yolo has demonstrated an impressively quick detection speed, even though sacrificing its accuracy. Since then, Joseph et al. made a few minor adjustments and came up with Yolov2, Yolov3, Yolov4 [17], and Yolov5 [18] as a result. Yolo V4, based on cross stage partial network (CSPNet) [19], was proposed for object detection. In that research, a network scaling approach was used to modify the depth, width, and resolution of the network, as well as the network’s topology, resulting in the construction of Scaled-Yolo V4. Yolo V4 is designed for real-time object detection on general GPU. To obtain the best speed and accuracy trade-off, C.Y. Wang et al. [19] re-designed Yolo V4 to Yolo V4-CSP.
The Yolov5 model that was made available to the public in May of 2020 has several advantages. It is lightweight, quick, and able to run on the mobile platform. Additionally, it can support a greater variety of applications. As a result, the purpose of this experiment is to evaluate Yolov5, which is supposed to increase both the accuracy and the efficiency. The model’s architecture is entirely determined by the model-configurations file. The ultralytics are compatible with a variety of Yolov5 designs and these models are referred to collectively as P5. Yolov5 models are distinguished from one another principally by the size of the parameters that they make use of, which are as follows: Yolov5n (nano), Yolov5s (small), Yolov5m (medium), Yolov5l (large), and Yolov5x (extra-large). When using this design for teaching purposes, it is advised to use images with a resolution of 640 pixels by 640 pixels.
The main contribution of this research is as follows: (1) The Yolov5 series of object identification algorithms, including Yolov5s, Yolov5m, Yolov5n, Yolov5l, and Yolov5x, are summarized in this research work. (2) This research investigates a variety of modern object detectors, including those used to identify road marking signs. Performance metrics track crucial data including the mean average precision (mAP), detection time (IoU), and the quantity of BFLOPS. (3) We create a new dataset called the Taiwan road marking sign dataset (TRMSD).
The organizational structure of the paper is as follows. In Section 2, we review the most recent publications in the field and describe our approach. Section 3 summarizes the experimental results. Section 4 focuses on discussing our findings, and Section 5 describes our conclusions and future research goals.

2. Materials and Methods

2.1. Road Marking Sign Identification

The identification of road markings is an essential task of intelligent transportation systems [20]. Road markings are as vital as road signs; not only do they help to provide a more secure environment for drivers but also provide drivers and other motorists with information that would not be able to be transmitted via traditional signs. Road markings can either be advisory, such as the reminder to keep the lines clear, or enforceable, such as yellow lines, box junctions, and stop lines. A large body of works has been created to tackle the challenge of automatically recognizing road markings. Previous researchers used various image processing approaches to identify road markings and signs. For instance, Foucher et al. [21,22] described a method for detecting and recognizing various painted road markings, including lanes, crosswalks, arrows, and other related road markings. These markings were all applied to the road. They presented a method for the recognition of road markings that was comprised of two stages: the first stage was the extraction of marking elements, and the second stage was the identification of related components based on a single pattern or recurring rectangular patterns [23]. Another method for the detection and identification of road markings was presented by Ding et al. [24]. In order to recognize and categorize five different road markings, HOG characteristics and a support vector machine (SVM) were applied. The technique for recognizing symbol-based road markings that was published by Greenhalgh et al. [25] made use of HOG features in addition to a support vector machine.

2.2. Yolov5 Architecture

There are five distinct architectures for the Yolov5, including Yolov5s, Yolov5m, Yolov5n, and Yolov5l. The primary distinction is based on the quantity of feature extraction modules and convolution kernels that are dispersed over the network at various predetermined locations. Figure 1 presents a diagrammatic representation of the internal network that Yolov5 possesses. The Yolov5 design incorporates a number of different technologies, including automatic learning bounding box anchoring, mosaic data improvement, and cross-stage partial networking. This design makes use of the most effective algorithm optimization methods that were developed recently for convolutional neural networks. It is built on the Yolo detection architecture.
Yolov5 has four main components: input, backbone, neck, and output [26]. The backbone model’s primary responsibility is to single out significant elements within the input image for analysis. When it comes to extracting rich, important characteristics from input photos, Yolov5 relies on cross stage partial networks (CSP) and spatial pyramid pooling (SPP) as its primary building blocks. SPP is beneficial for identifying the same item in multiple sizes and scales, which is important when it comes to the correct generalization of a model concerning object scaling. The feature pyramid architectures of the feature pyramid network (FPN) [27] and path aggregation network (PANet) [28,29] are utilized in the construction of the neck network. Powerful semantic features are distributed throughout the FPN structure, beginning at the top feature maps and working their way down to the lower feature maps. At the same time, it is the responsibility of the PAN structure to ensure the transmission of reliable localization features from lower feature maps to higher feature maps. PANet is utilized as a neck in Yolo v5, which allows for the generation of a feature pyramid [30,31].
Yolov5 is quite similar to Yolov4, although there are a few key differences between them: (1) Yolov4 is distributed using the Darknet framework, which is written in the programming language C. PyTorch serves as the underlying infrastructure for Yolov5. (2) In Yolov4, the configuration file is a a.cfg file, while in Yolov5, the configuration file is a a.yaml file. The cross stage partial network is the full name of CSP Net, which is a solution that primarily addresses the issue of a significant quantity of calculation in reasoning from the point of view of the design of network structures. According to the author of CSP Net, the issue of unnecessary inference calculations can be traced back to the network optimization gradient information repetition [32].
Using the structure of Yolov5 as an example, an initial image with dimensions of 608 × 608 × 3 is input into the focus structure, as shown in Figure 1. Next, the slicing operation is used to change the image into a feature map with dimensions of 304 × 304 × 12, and this is followed by 32 convolution operation kernels, which result in a final feature map with dimensions of 304 × 304 × 32. The information presented in Table 1 provides an overview of all of Yolov5’s models, including the inference speed on CPU and GPU, as well as the number of parameters with an image size of 640 [33,34].
The vast majority of annotation tools produce their results in the Yolo labeling format, which generates a single text file containing annotations for each image. Each text file has one bounding box, abbreviated as “BBox”, with annotation for each of the objects that are displayed in the image. The annotations are scaled appropriately to the image, and their values range from 0 to 1 inclusively [35]. The following equations are the basis for the adjustment procedure for the Yolo format calculation.
d w = 1 / W
x = x 1 + x 2 2 × d w
d h = 1 / H
y = y 1 + y 2 2 × d h
w = x 2 x 1 × d w
h = y 2 y 1 × d h
H indicates the height of the image, dh refers to the absolute height of the image, W serves as the width of the image, and dw represents the absolute width of the picture.
Based on Table 1, Yolov5n is a newly announced nano model that is the smallest in the family. It is designed to be used for edge computing, Internet of Things devices, and has support for OpenCV deep neural networks. When saved in INT8 format, it is less than 2.5 MB, and when saved in FP32 format, it is roughly 4 MB. It works wonderfully for applications on mobile devices. Next, Yolov5s is the smallest model in the family and has approximately 7.2 million parameters. It is an excellent choice for performing inference on the CPU because of its tiny size. This is the medium-sized model, Yolov5m, which has a total of 21.2 million different parameters [36]. Due to the fact that it strikes a healthy mix between speed and accuracy, it is possibly the model that is most suited for the majority of datasets and training [37]. Moreover, the Yolov5l model has 46.5 million different parameters and is the largest member of the Yolov5 family. It works wonderfully for datasets in which we need to identify more discrete items. Furthermore, the Yolov5x is the largest of the five models, and it also has the highest mAP value of the five. It is considerably slower than the others and contains 86.7 million parameters [6].

3. Results

3.1. Taiwan Road Marking Sign Dataset (TRMSD)

In addition, we carried out this experiment utilizing images obtained from various video and image sources in order to simulate the road marking signs that are used in Taiwan. Eighty percent of images in the dataset are used for training, while the remaining twenty percent are used for validation. Figure 2 shows the Taiwan road marking sign dataset (TRMSD) sample image and Table 2 displays the TRMSD statistics. Furthermore, to prevent an imbalance in the data, we use images with a range of 391–409 for each class. Thus, our dataset consists of 6009 images in total, each with a size of 512 × 288.
Our experiment consisted of 15 classes (P1–P15), including Go Straight Turn Left, Turn Right, Turn Right or Go Straight, Turn Left or Go Straight, Speed Limit 40, Speed Limit 50, Speed Limit 60, Speed Limit 70, Zebra Crossing, Slow Sign, Overtaking Prohibited, Barrier Line, Cross Hatch, and Stop Line.

3.2. Training Result

The training process and result are explained in this section. We divided our dataset into a training subset and a testing subset: 70 percent for training and 30 percent for testing. Figure 3 describes the training process for batch 0 and batch 1. The anchor boxes are generated by Yolov5 with the help of a genetic algorithm. This procedure is referred to as the auto anchor process, which recalculates the anchor boxes in order to better match the data in the event that the default ones are inadequate. To generate k-Means evolved anchor boxes, this is combined with the k-Means algorithm. This is one of the reasons that Yolov5 performs so admirably even on a wide variety of datasets. Figure 4 describes the Yolov5 validation process for batch 0 and batch 1. During the training phase of Yolov5, four individual images are spliced to form a larger image. Each of the four individual images is subject to a random processing step during the splicing phase, which results in varying dimensions and configurations for each individual image. To make an analysis of our model, we will use the validation script. Using the “task” option, one can decide whether the performance is evaluated across the training dataset, the validation dataset, or the test dataset.
By default, all the results are logged into the runs/train directory, and a new experiment directory is created for each subsequent training, with the names runs/train/exp2, runs/train/exp, etc. We can view the mosaics, labels, predictions, and augmentation effects by looking at the train and val jpgs. It is important to note that an Ultralytics mosaic data loader is needed for training, and this device merges four photos into one mosaic while training. The weights are saved after 50 epochs during the training stage.
The last stage in training is called fine-tuning, which is completely optional. It involves unfreezing the entire model that we obtained in the previous step and then retraining it on our data using a very small learning rate. This has the potential to achieve considerable improvements by gradually adjusting the previously trained features on the fresh data. The hyperparameters-configurations file is where the learning rate can be modified to suit our needs. For the sake of demonstrating the tutorial, we will use the hyperparameters that are defined in the built-in hyp.finetune.yaml file. These hyperparameters have a significantly lower learning rate than the default. The weights will initially be set to the values that were saved on the stage before this one. Once finished training, it is stored using the standard PyTorch protocol, which entails using the.pt file extension.
Transfer learning is an efficient method for retraining a model on new data without retraining the entire network. Instead, a portion of the initial weight is assigned, while the remainder is used to calculate losses and is modified by the optimizer. It uses fewer resources than standard training and allows for shorter training periods, although it can result in a decrease in the accuracy of the units being trained. In our model, we use transfer learning while training with the Yolov5 series model.
During the training phase, we watch the mAP@0.5 to assess how well our detector performs on the validation set; a greater number indicates better performance. The dataset yet another markup language (YAML) file is possibly one of the most essential aspects of the Yolov5 training. This file provides the class names with the path to the data that is used for training and validation. For the training script to be able to detect the image paths, the label paths, and the class names, we need to supply this file path as an argument when we are running the training script. The information is already included in the dataset. Table 3 describes the training process of the Yolov5 series. Yolov5m, Yolov5l, and Yolov5x achieve the same average of mAP at approximately 83%, followed by Yolov5s with an average of mAP 81%. Based on the training results, it can be concluded that Yolov5m is the most stable method compared to other methods with 76% precision, 86% recall, and 83% mAP. Furthermore, Class P6 and Class P9 achieve the highest accuracy for all models with mAP in the range of 99% to 100%.
Additionally, we can acquire the precision–recall curve, which is automatically saved for every validation. Figure 5 depicts the precision and recall for Yolov5m and Yolov5l. The performance metrics we use to evaluate the performance of our dataset TRMSD in Yolov5 model experiments include precision, recall, accuracy score, and F1. Among them, precision and recall are defined in Equations (7) and (8) [38], then accuracy and F1 are defined in Equations (9) and (10) [39]. A compound loss is computed for the Yolo family of algorithms, with the score for objectiveness, the score for class probability, and the score for bounding box regression serving as the inputs. For the purpose of calculating the loss of class probability and object score, Ultralytics uses the binary cross-entropy with logits loss function that is available in PyTorch [40]. True positive (TP) is the number of “yes” in the real situation when the model evaluation is also “yes”, and true negative (TN) is the number of “no” in the real situation when the model evaluation is also “no”. False positive (FP) is the number of “no” in the real situation when the model evaluation is also “yes”; false negative (FN) is the number of “yes” in the real situation when the model evaluation is also “no” [41]. The mean average precision (mAP) is a popular indicator for evaluating the performance of object identification models and defined in Equation (11).
P r e c i s i o n   P = T P T P + F P
R e c a l l   R = T P T P + F N
A c c u r a c y   A c c = T P + T N T P + F N + F P + F N
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
mAP = 1 N i = 1 N A c c
The Yolo loss function is calculated using the Equation (12) [16].
Yolo Loss Function = λ c o o r d i = 0 s 2 j = 0 B 𝕝 i j o b j x i x ^ i 2 + y y ^ i 2   + λ c o o r d i = 0 s 2 j = 0 B 𝕝 i j o b j w i w ^ i 2 + h i h ^ i 2   + i = 0 s 2 j = 0 B 𝕝 i j o b j C i C ^ i 2 + λ n o o b j i = 0 s 2 j = 0 B 𝕝 i j n o o b j C i C ^ i 2   + i = 0 s 2 𝕝 i o b j c ϵ c l a s s e s p i c p ^ i ( c ) 2
where S is the number of cells in the image, B is the number of bounding boxes predicted in each grid cell, and c represents the class prediction for each grid cell. Furthermore, p i c   represents the confidence probability. For any box j of cell i, x i j and y i j represent the coordinates of the center of the anchor box, hij gives height, w i j gives width of the box and C i j gives the confidence score. Finally, λcoord and λnoobj are the weights to decide the importance of localization and recognition in the training.

4. Discussion

Table 4 describes the testing performance of the Yolov5 series. After achieving satisfactory results during the training phase, our model is now prepared for the inference phase. Each image is augmented (with a horizontal flip and three different resolutions), and the final prediction is an ensemble of all these augmentations. After inference, we can further improve the accuracy of the predictions by adding test-time augmentations (TTA). Because the TTA results in an inference that is two to three times longer, we have to abandon it to maintain a high frames-per-second (FPS) rate. An inference can be made based on an image, a video, a directory, a stream, a camera, or even a link to YouTube. These can all serve as the input. In our experiment, we test the Yolov5 series with some groups of images for each class. Based on our experiment results, Yolov5m and Yolov5l achieve the highest average mAP 87%, followed by Yolov5s with the mAP 84%, Yolov5x with mAP 83%, and Yolov5 with mAP 80%. Class P6 achieves the highest mAP for all Yolov5 series ranging from 99% to 100%.
When conducting deep learning, parameters known as hyperparameters are set in advance of formal training. The use of appropriate hyperparameters has the potential to increase the performance of the model. The Yolov5 method has a total of 23 hyperparameters, most of which are employed in the process of configuring the learning rate, loss function, and data improvement parameters and so on. It is discovered unequivocally that the more complex the network structure model, the lower the training loss convergence and the higher the validation loss, which indicates that the overfitting of the model is more severe. The complexity of the model can lead to an increase in the validation loss in proportional measure. The model’s capacity to identify abnormalities shows just a slight improvement overall. The weight volume and the number of parameters are two measures that can be used to characterize the complexity of the model. The more complicated the model is, the higher these indexes are, and the more RAM the GPU requires to store it during training.
Figure 6 shows the recognition results of the Yolov5 series. The recognition results for Yolov5m are shown in Figure 6c. Class P1 achieved 73% accuracy, Class P14 obtained 68%, Class P3 showed accuracy of 66%, 83%, and 38%. All models can detect road markings sign in the image very well. The Yolov5 is a lightweight and relatively simple device to operate. Moreover, there are a total of 25 blocks in the Yolov5m medium model (from 0 to 24). Each block is composed of a stacked arrangement of a variety of layers. Yolov5 is quick to train, quick to draw conclusions, and does well in performance.
Figure 7 contains specific information regarding the detection performance and the level of complexity of a variety of models including layers, parameters, GFLOPS, and FPS. In our experiment, Yolov5x contains the most layers at approximately 444 layers, then Yolov5l has 367 layers, and Yolov5m, Yolov5s, Yolov5n have the same 213 layers. Yolov5x loads the most parameters at 86,267,620, and Yolov5m and Yolov5n contain the same parameters, 1,779,460. In our experiment, Yolov5s loads 24.6 FPS; Yolov5n, 40.5 FPS; Yolov5m, 11.4 FPS; Yolov5l, 2.8 FPS; and Yolov5x, 1.7 FPS.
The following is a list of benefits in using Yolov5: First, Yolo is the first model in the Yolo family that was originally written on PyTorch rather than on PJ Reddie’s Darknet. Darknet is a research framework that offers a great degree of adaptability but was not developed with production situations in mind. It caters to a more limited population of end consumers. When all these factors are considered together, the result is that Darknets are more difficult to configure and less production ready. Because Yolov5 was originally implemented in PyTorch, Yolov5 reaps the benefits of the well-established ecosystem that PyTorch has to offer, including simplified support and simplified deployment. Moreover, as Yolov5 becomes more popular, the research community may find that iterations on Yolov5 become simpler. Because of this, deployment for mobile devices is also becoming easier, as the models can be easily compiled to ONNX or CoreML. Second, Yolov5 is incredibly fast, even by the current standards. Yolov5 is now capable of producing 140 frames per second in batches, and the Yolov5 implementation testing uses this setting by default. Yolov4 is capable of performing 30 FPS while Yolov5 only produces 10 FPS when the batch size is set to 1. Third, the Yolov5 formula is reliable. Although EfficientDet and Yolov4 yield similar results to our expectations, it is very unusual to observe such a broad increase in performance that is not accompanied by a reduction in accuracy. Fourth, Yolov5 has a low memory capacity. To be more specific, the weight file size used by Yolov5 is 27 megabytes. The weight file for Yolov4 (with Darknet architecture) takes up 244 MB of our storage space. When compared to Yolov4, Yolov5 is approximately 90 percent smaller. Because of this, Yolov5 can be applied to embedded devices in a much simpler way.

5. Conclusions

The objective of this study is to provide a concise overview of CNN-based object identification methods, with a particular focus on the Yolov5 algorithm series. During our experimental research, we tested and analyzed a variety of modern object detectors. Among the detectors, we investigated those that were designed to identify road marking signs. Important characteristics such as the mean average precisions (mAP), the detection time (IoU), and the number of BFLOPS were measured by the assessment criteria.
We have reached the following conclusion. First, Yolov5m is the most stable method compared to other methods with 76% precision, 86% recall, and 83% mAP during the training stage. Second, Yolov5m and Yolov5l achieve the highest average mAP 87% in the testing stage. Next, we created our own dataset for road marking signs in Taiwan (TRMSD). In future research, we plan to integrate the detection of road markings with explainable artificial intelligence (XAI). In addition, we are planning to upgrade our Taiwan road marking sign dataset (TRMSD) with an emphasis on the recognition of pothole signs, which can add different illuminations and textures conditions into our dataset.

Author Contributions

Conceptualization, C.D.; data curation, R.-C.C. and Y.-C.Z.; formal analysis, C.D.; investigation, C.D. and H.J.C.; methodology, C.D.; project administration, R.-C.C.; resources, Y.-C.Z.; software, Y.-C.Z. and H.J.C.; supervision, R.-C.C.; validation, C.D. and H.J.C.; visualization, H.J.C.; writing—original draft, C.D.; writing—review and editing, C.D. and R.-C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by the Ministry of Science and Technology, Taiwan. The Nos are MOST-110-2927-I-324-50, MOST-110-2221-E-324-010, MOST-109-2622-E-324-004, and MOST-111-2622-E-324-002-Taiwan.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The author would like to thank all colleagues from Chaoyang Technology University, Satya Wacana Christian University, Indonesia, Atma Jaya Catholic University of Indonesia, Jakarta, and all those involved in this research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fang, C.Y.; Chen, S.W.; Fuh, C.S. Road-Sign Detection and Tracking. IEEE Trans. Veh. Technol. 2003, 52, 1329–1341. [Google Scholar] [CrossRef]
  2. Wontorczyk, A.; Gaca, S. Study on the Relationship between Drivers’ Personal Characters and Non-Standard Traffic Signs Comprehensibility. Int. J. Environ. Res. Public Health 2021, 18, 2678. [Google Scholar] [CrossRef] [PubMed]
  3. Vacek, S.; Schimmel, C.; Dillmann, R. Road-Marking Analysis for Autonomous Vehicle Guidance. Available online: http://ecmr07.informatik.uni-freiburg.de/proceedings/ECMR07_0034.pdf (accessed on 4 December 2022).
  4. Muhammad, K.; Ullah, A.; Lloret, J.; Del Ser, J.; de Albuquerque, V.H.C. Deep Learning for Safe Autonomous Driving: Current Challenges and Future Directions. IEEE Trans. Intell. Transp. Syst. 2021, 22, 4316–4336. [Google Scholar] [CrossRef]
  5. Tai, S.K.; Dewi, C.; Chen, R.C.; Liu, Y.T.; Jiang, X.; Yu, H. Deep Learning for Traffic Sign Recognition Based on Spatial Pyramid Pooling with Scale Analysis. Appl. Sci. 2020, 10, 6997. [Google Scholar] [CrossRef]
  6. Mathew, M.P.; Mahesh, T.Y. Leaf-Based Disease Detection in Bell Pepper Plant Using YOLO V5. Signal Image Video Process. 2022, 16, 841–847. [Google Scholar] [CrossRef]
  7. Bachute, M.R.; Subhedar, J.M. Autonomous Driving Architectures: Insights of Machine Learning and Deep Learning Algorithms. Mach. Learn. Appl. 2021, 6, 100164. [Google Scholar] [CrossRef]
  8. Javanmardi, M.; Song, Z.; Qi, X. A Fusion Approach to Detect Traffic Signs Using Registered Color Images and Noisy Airborne LiDAR Data. Appl. Sci. 2021, 11, 309. [Google Scholar] [CrossRef]
  9. Qin, B.; Liu, W.; Shen, X.; Chong, Z.J.; Bandyopadhyay, T.; Ang, M.H.; Frazzoli, E.; Rus, D. A General Framework for Road Marking Detection and Analysis. In Proceedings of the IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC, The Hague, The Netherlands, 6–9 October 2013; pp. 619–625. [Google Scholar]
  10. Soilán, M.; Riveiro, B.; Martínez-Sánchez, J.; Arias, P. Segmentation and Classification of Road Markings Using MLS Data. ISPRS J. Photogramm. Remote Sens. 2017, 123, 94–103. [Google Scholar] [CrossRef]
  11. Wen, C.; Sun, X.; Li, J.; Wang, C.; Guo, Y.; Habib, A. A Deep Learning Framework for Road Marking Extraction, Classification and Completion from Mobile Laser Scanning Point Clouds. ISPRS J. Photogramm. Remote Sens. 2019, 147, 178–192. [Google Scholar] [CrossRef]
  12. Chen, S.; Zhang, Z.; Zhong, R.; Zhang, L.; Ma, H.; Liu, L. A Dense Feature Pyramid Network-Based Deep Learning Model for Road Marking Instance Segmentation Using MLS Point Clouds. IEEE Trans. Geosci. Remote Sens. 2021, 59, 784–800. [Google Scholar] [CrossRef]
  13. Cheng, Y.T.; Patel, A.; Wen, C.; Bullock, D.; Habib, A. Intensity Thresholding and Deep Learning Based Lane Marking Extraction and Lanewidth Estimation from Mobile Light Detection and Ranging (LiDAR) Point Clouds. Remote Sens. 2020, 12, 1379. [Google Scholar] [CrossRef]
  14. Chen, J.; Jia, K.; Chen, W.; Lv, Z.; Zhang, R. A Real-Time and High-Precision Method for Small Traffic-Signs Recognition. Neural Comput. Appl. 2022, 34, 2233–2245. [Google Scholar] [CrossRef]
  15. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot Multibox Detector. In Lecture Notes in Computer Science; including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics; Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  16. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  17. Adibhatla, V.A.; Chih, H.C.; Hsu, C.C.; Cheng, J.; Abbod, M.F.; Shieh, J.S. Applying Deep Learning to Defect Detection in Printed Circuit Boards via a Newest Model of You-Only-Look-Once. Math. Biosci. Eng. 2021, 18. [Google Scholar] [CrossRef] [PubMed]
  18. Farooq, M.A.; Shariff, W.; Corcoran, P. Evaluation of Thermal Imaging on Embedded GPU Platforms for Application in Vehicular Assistance Systems. IEEE Trans. Intell. Veh. 2022. [Google Scholar] [CrossRef]
  19. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. Scaled-Yolov4: Scaling Cross Stage Partial Network. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13024–13033. [Google Scholar]
  20. Park, Y.K.; Park, H.; Woo, Y.S.; Choi, I.G.; Han, S.S. Traffic Landmark Matching Framework for HD-Map Update: Dataset Training Case Study. Electronics 2022, 11, 863. [Google Scholar] [CrossRef]
  21. Foucher, P.; Sebsadji, Y.; Tarel, J.P.; Charbonnier, P.; Nicolle, P. Detection and Recognition of Urban Road Markings Using Images. In Proceedings of the IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC, Washington, DC, USA, 5–7 October 2011. [Google Scholar]
  22. Dewi, C.; Chen, R.-C.; Jiang, X.; Yu, H. Adjusting Eye Aspect Ratio for Strong Eye Blink Detection Based on Facial Landmarks. PeerJ Comput. Sci. 2022, 8, e943. [Google Scholar] [CrossRef]
  23. Hussain, Q.; Alhajyaseen, W.K.M.; Pirdavani, A.; Brijs, K.; Shaaban, K.; Brijs, T. Do Detection-Based Warning Strategies Improve Vehicle Yielding Behavior at Uncontrolled Midblock Crosswalks? Accid. Anal. Prev. 2021, 157, 106166. [Google Scholar] [CrossRef]
  24. Ding, D.; Yoo, J.; Jung, J.; Jin, S.; Kwon, S. Efficient Road-Sign Detection Based on Machine Learning. Bull. Netw. Comput. Syst. Softw. 2015, 4, 15–17. [Google Scholar]
  25. Greenhalgh, J.; Mirmehdi, M. Detection and Recognition of Painted Road Surface Markings. In Proceedings of the ICPRAM 2015—4th International Conference on Pattern Recognition Applications and Methods, Proceedings, Lisbon, Portugal, 10–12 January 2015; Volume 1. [Google Scholar]
  26. Li, Z.; Tian, X.; Liu, X.; Liu, Y.; Shi, X. A Two-Stage Industrial Defect Detection Framework Based on Improved-YOLOv5 and Optimized-Inception-ResnetV2 Models. Appl. Sci. 2022, 12, 834. [Google Scholar] [CrossRef]
  27. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 January 2017; pp. 2117–2124. [Google Scholar]
  28. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  29. Dewi, C.; Chen, R.C.; Yu, H.; Jiang, X. Robust Detection Method for Improving Small Traffic Sign Recognition Based on Spatial Pyramid Pooling. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 1–18. [Google Scholar] [CrossRef]
  30. Yuan, D.L.; Xu, Y. Lightweight Vehicle Detection Algorithm Based on Improved YOLOv4. Eng. Lett. 2021, 29, 1544–1551. [Google Scholar]
  31. Vinitha, V.; Velantina, V. COVID-19 Facemask Detection With Deep Learning and Computer Vision. Int. Res. J. Eng. Technol. 2020, 7, 3127–3132. [Google Scholar]
  32. Yao, J.; Qi, J.; Zhang, J.; Shao, H.; Yang, J.; Li, X. A Real-Time Detection Algorithm for Kiwifruit Defects Based on Yolov5. Electronics 2021, 10, 1711. [Google Scholar] [CrossRef]
  33. Mihalj, T.; Li, H.; Babić, D.; Lex, C.; Jeudy, M.; Zovak, G.; Babić, D.; Eichberger, A. Road Infrastructure Challenges Faced by Automated Driving: A Review. Appl. Sci. 2022, 12, 3477. [Google Scholar] [CrossRef]
  34. Sun, X.M.; Zhang, Y.J.; Wang, H.; Du, Y.X. Research on Ship Detection of Optical Remote Sensing Image Based on Yolov5. J. Phys. Conf. Ser. 2022, 2215, 012027. [Google Scholar] [CrossRef]
  35. Long, J.W.; Yan, Z.R.; Peng, L.; Li, T. The Geometric Attention-Aware Network for Lane Detection in Complex Road Scenes. PLoS ONE 2021, 16, e0254521. [Google Scholar] [CrossRef] [PubMed]
  36. Ultralytics Yolov5. Available online: https://github.com/ultralytics/yolov5 (accessed on 13 January 2021).
  37. Otgonbold, M.E.; Gochoo, M.; Alnajjar, F.; Ali, L.; Tan, T.H.; Hsieh, J.W.; Chen, P.Y. SHEL5K: An Extended Dataset and Benchmarking for Safety Helmet Detection. Sensors 2022, 22, 2315. [Google Scholar] [CrossRef]
  38. Han, K.; Zeng, X. Deep Learning-Based Workers Safety Helmet Wearing Detection on Construction Sites Using Multi-Scale Features. IEEE Access 2022, 10, 718–729. [Google Scholar] [CrossRef]
  39. Jiang, L.; Liu, H.; Zhu, H.; Zhang, G. Improved YOLO v5 with Balanced Feature Pyramid and Attention Module for Traffic Sign Detection. MATEC Web Conf. 2022, 355, 03023. [Google Scholar] [CrossRef]
  40. Zhao, S.; Zhang, K. Online Predictive Connected and Automated Eco-Driving on Signalized Arterials Considering Traffic Control Devices and Road Geometry Constraints under Uncertain Traffic Conditions. Transp. Res. Part B Methodol. 2021, 145, 80–117. [Google Scholar] [CrossRef]
  41. Dewi, C.; Chen, R.-C. Combination of Resnet and Spatial Pyramid Pooling for Musical Instrument Identification. Cybern. Inf. Technol. 2022, 22, 104. [Google Scholar] [CrossRef]
Figure 1. Yolov5 architecture.
Figure 1. Yolov5 architecture.
Bdcc 06 00149 g001
Figure 2. TRMSD sample image.
Figure 2. TRMSD sample image.
Bdcc 06 00149 g002
Figure 3. Training Process, (a) Batch 0 and (b) Batch 1.
Figure 3. Training Process, (a) Batch 0 and (b) Batch 1.
Bdcc 06 00149 g003
Figure 4. Validation Process batch 0. (a) Labels and (b) Predictions.
Figure 4. Validation Process batch 0. (a) Labels and (b) Predictions.
Bdcc 06 00149 g004
Figure 5. Precision and recall. (a) Yolov5m, and (b) Yolov5l.
Figure 5. Precision and recall. (a) Yolov5m, and (b) Yolov5l.
Bdcc 06 00149 g005aBdcc 06 00149 g005b
Figure 6. Recognition results. (a) Yolov5n, (b) Yolov5s, (c) Yolov5m, (d) Yolov5l, and (e) Yolov5x.
Figure 6. Recognition results. (a) Yolov5n, (b) Yolov5s, (c) Yolov5m, (d) Yolov5l, and (e) Yolov5x.
Bdcc 06 00149 g006aBdcc 06 00149 g006b
Figure 7. Comparison of the (a) Layers, (b) parameters, (c) GFLOPS, and (d) FPS.
Figure 7. Comparison of the (a) Layers, (b) parameters, (c) GFLOPS, and (d) FPS.
Bdcc 06 00149 g007
Table 1. An overview of YoloV5 models.
Table 1. An overview of YoloV5 models.
Model NameParams
(Million)
Accuracy (mAP 0.5)CPU
Time (ms)
GPU
Time (ms)
Yolov5n1.945.7456.3
Yolov5s7.256.8986.4
Yolov5m21.264.12248.2
Yolov5l46.567.343010.1
Yolov5x86.768.976612.1
Table 2. TRMSD Dataset.
Table 2. TRMSD Dataset.
Class IDClass NameTotal Image
P1Turn Right405
P2Turn Left401
P3Go Straight407
P4Turn Right or Go Straight409
P5Turn Left or Go Straight403
P6Speed Limit (40)391
P7Speed Limit (50)401
P8Speed Limit (60)400
P9Speed Limit (70)398
P10Zebra Crossing401
P11Slow Sign399
P12Overtaking Prohibited404
P13Barrier Line409
P14Cross Hatch398
P15Stop Line403
Table 3. Training Performance of Yolov5 series.
Table 3. Training Performance of Yolov5 series.
ClassYolov5nYolov5sYolov5mYolov5lYolov5x
PRmAP@.5PRmAP@.5PRmAP@.5PRmAP@.5PRmAP@.5
P10.630.890.790.660.870.770.630.870.800.610.870.800.610.870.78
P20.620.770.740.660.690.740.650.780.720.640.740.730.910.790.70
P30.530.750.640.550.750.620.600.830.720.540.820.720.600.780.74
P40.450.630.610.410.650.600.430.770.650.400.620.610.400.730.62
P50.420.620.550.370.520.460.420.710.570.410.840.550.370.720.48
P60.991.001.000.991.001.000.991.001.000.991.001.000.991.001.00
P70.810.890.910.820.870.890.890.880.900.780.860.900.790.890.90
P80.840.980.970.820.980.970.871.000.970.900.990.980.891.000.98
P90.991.001.000.991.001.000.991.001.000.981.001.000.991.001.00
P100.950.640.890.830.590.750.930.800.890.950.760.910.920.830.91
P110.840.990.990.890.950.980.860.990.990.880.990.990.920.990.99
P120.780.740.730.660.780.740.690.750.710.690.740.700.750.780.76
P130.640.750.730.580.680.630.760.810.780.720.820.790.700.840.78
P140.810.850.860.7920.840.840.870.910.890.880.880.880.860.850.87
P150.790.660.790.860.710.810.840.820.900.910.860.930.830.840.91
All0.740.810.810.720.790.790.760.860.830.750.820.830.750.860.83
Table 4. Testing Performance of Yolov5 series.
Table 4. Testing Performance of Yolov5 series.
ClassYolov5nYolov5sYolov5mYolov5lYolov5x
PRmAP@.5PRmAP@.5PRmAP@.5PRmAP@.5PRmAP@.5
P10.620.920.810.640.920.840.650.930.860.650.940.850.610.870.78
P20.690.670.750.720.750.790.770.730.820.790.730.810.910.790.70
P30.540.740.620.550.800.670.600.890.730.600.790.720.600.780.74
P40.510.590.640.500.690.670.510.750.710.540.660.700.400.730.62
P50.480.650.530.490.690.550.510.650.570.510.640.550.370.720.48
P60.991.001.000.991.001.001.001.000.990.991.001.000.991.001.00
P70.840.910.930.870.910.960.930.900.950.880.910.950.790.890.90
P80.820.970.960.830.970.970.860.990.980.820.990.980.891.000.98
P90.981.000.990.981.000.990.981.000.990.981.000.990.991.001.00
P100.900.610.820.870.660.870.930.800.940.880.800.930.920.830.91
P110.800.990.970.890.990.980.911.000.980.891.000.980.920.990.99
P120.780.690.760.710.860.810.860.890.910.880.920.930.750.780.76
P130.710.660.700.680.830.800.770.880.890.800.850.870.700.840.78
P140.730.830.820.800.860.850.860.920.900.880.880.900.860.850.87
P150.840.560.720.880.670.800.920.700.880.900.710.870.830.840.91
All0.750.790.800.760.840.840.800.870.870.80.850.870.750.860.83
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dewi, C.; Chen, R.-C.; Zhuang, Y.-C.; Christanto, H.J. Yolov5 Series Algorithm for Road Marking Sign Identification. Big Data Cogn. Comput. 2022, 6, 149. https://doi.org/10.3390/bdcc6040149

AMA Style

Dewi C, Chen R-C, Zhuang Y-C, Christanto HJ. Yolov5 Series Algorithm for Road Marking Sign Identification. Big Data and Cognitive Computing. 2022; 6(4):149. https://doi.org/10.3390/bdcc6040149

Chicago/Turabian Style

Dewi, Christine, Rung-Ching Chen, Yong-Cun Zhuang, and Henoch Juli Christanto. 2022. "Yolov5 Series Algorithm for Road Marking Sign Identification" Big Data and Cognitive Computing 6, no. 4: 149. https://doi.org/10.3390/bdcc6040149

Article Metrics

Back to TopTop